0% found this document useful (0 votes)
25 views2 pages

Syllabus

The document outlines the syllabus for the Machine Learning course offered by the Department of Information Technology at Jawaharlal Nehru Technological University, Kakinada, starting from the academic year 2019-20. It includes course objectives, outcomes, and detailed unit-wise topics covering various aspects of machine learning, including decision trees, neural networks, support vector machines, and Bayesian learning. The syllabus also lists recommended textbooks and e-resources for further study.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
25 views2 pages

Syllabus

The document outlines the syllabus for the Machine Learning course offered by the Department of Information Technology at Jawaharlal Nehru Technological University, Kakinada, starting from the academic year 2019-20. It includes course objectives, outcomes, and detailed unit-wise topics covering various aspects of machine learning, including decision trees, neural networks, support vector machines, and Bayesian learning. The syllabus also lists recommended textbooks and e-resources for further study.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 2

R-19 Syllabus for IT. JNTUK w. e. f.

2019-20

JAWAHARLAL NEHRU TECHNOLOGICAL UNIVERSITY: KAKINADA


KAKINADA 533 003, Andhra Pradesh, India

DEPARTMENT OF INFORMATION TECHNOLOGY


L T P C
IV Year I Semester
3 1 0 4

MACHINE LEARNING
Course Objectives:
The course is introduced for students to
Gain knowledge about basic concepts of Machine Learning
Study about different learning algorithms
Learn about of evaluation of learning algorithms
Learn about artificial neural networks
Course Outcomes:
Identify machine learning techniques suitable for a given problem
Solve the problems using various machine learning techniques
Apply Dimensionality reduction techniques
Design application using machine learning techniques
UNIT I
Introduction: Definition of learning systems, Goals and applications of machine learning, Aspects of
developing a learning system: training data, concept representation, function approximation.
Inductive Classification: The concept learning task, Concept learning as search through a hypothesis
space, General-to-specific ordering of hypotheses, Finding maximally specific hypotheses, Version
spaces and the candidate elimination algorithm, Learning conjunctive concepts, The importance of
inductive bias.
UNIT II
Decision Tree Learning: Representing concepts as decision trees, Recursive induction of decision trees,
Picking the best splitting attribute: entropy and information gain, Searching for simple trees and
computational complexity, Occam's razor, Overfitting, noisy data, and pruning.
Experimental Evaluation of Learning Algorithms: Measuring the accuracy of learned hypotheses.
Comparing learning algorithms: cross-validation, learning curves, and statistical hypothesis testing.
UNIT III
Computational Learning Theory: Models of learnability: learning in the limit; probably approximately
correct (PAC) learning. Sample complexity for infinite hypothesis spaces, Vapnik-Chervonenkis
dimension.
Rule Learning: Propositional and First-Order, Translating decision trees into rules, Heuristic rule
induction using separate and conquer and information gain, First-order Horn-clause induction (Inductive
Logic Programming) and Foil, Learning recursive rules, Inverse resolution, Golem, and Progol.
UNIT IV
Artificial Neural Networks: Neurons and biological motivation, Linear threshold units. Perceptrons:
representational limitation and gradient descent training, Multilayer networks and backpropagation,
Hidden layers and constructing intermediate, distributed representations. Overfitting, learning network
structure, recurrent networks.
Support Vector Machines: Maximum margin linear separators. Quadractic programming solution to
R-19 Syllabus for IT. JNTUK w. e. f. 2019-20

JAWAHARLAL NEHRU TECHNOLOGICAL UNIVERSITY: KAKINADA


KAKINADA 533 003, Andhra Pradesh, India

DEPARTMENT OF INFORMATION TECHNOLOGY


finding maximum margin separators. Kernels for learning non-linear functions.
UNIT V
Bayesian Learning: Probability theory and Bayes rule. Naive Bayes learning algorithm. Parameter
smoothing. Generative vs. discriminative training. Logisitic regression. Bayes nets and Markov nets for
representing dependencies.
Instance-Based Learning: Constructing explicit generalizations versus comparing to past specific
examples. k-Nearest-neighbor algorithm. Case-based learning.
Text Books:
1) -Hill, 1997.
2) Machine Learning, Saikat Dutt, Subramanian Chandramouli, Amit Kumar Das, Pearson, 2019.
Reference Books:
1) 2004.
2) Stephen Marsland, Learning -An Algorithmic Second Edition,
Chapman and Hall/CRC Machine Learning and Pattern Recognition Series, 2014.
3)

e-Resources:
1) https://www.deeplearning.ai/machine-learning-yearning/
2) Shai Shalev-Shwartz , Shai Ben-
Cambridge University Press
https://www.cse.huji.ac.il/~shais/UnderstandingMachineLearning/index.html

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy