Machine Learning (R17A0534) Lecture Notes: B.Tech Iv Year - I Sem (R17) (2020-21)
Machine Learning (R17A0534) Lecture Notes: B.Tech Iv Year - I Sem (R17) (2020-21)
[R17A0534]
LECTURE NOTES
DEPARTMENT OF
COMPUTER SCIENCE AND ENGINEERING
UNIT I:
Introduction to Machine Learning
Introduction ,Components of Learning , Learning Models , Geometric Models, Probabilistic
Models, Logic Models, Grouping and Grading, Designing a Learning System, Types of
Learning, Supervised, Unsupervised, Reinforcement, Perspectives and Issues, Version Spaces,
PAC Learning, VC Dimension.
UNIT II:
Supervised and Unsupervised Learning
Decision Trees: ID3, Classification and Regression Trees, Regression: Linear Regression, Multiple Linear
Regression, Logistic Regression, Neural Networks: Introduction, Perception, Multilayer Perception,
Support Vector Machines: Linear and Non-Linear, Kernel Functions, K Nearest Neighbors.
Introduction to clustering, K-means clustering, K-Mode Clustering.
UNIT III:
Ensemble and Probabilistic Learning
Model Combination Schemes, Voting, Error-Correcting Output Codes, Bagging: Random Forest Trees,
Boosting: Adaboost, Stacking.
Gaussian mixture models - The Expectation-Maximization (EM) Algorithm, Information Criteria, Nearest
neighbour methods - Nearest Neighbour Smoothing, Efficient Distance Computations: the KD-Tree,
Distance Measures.
UNIT IV:
Reinforcement Learning and Evaluating Hypotheses
Introduction, Learning Task, Q Learning, Non deterministic Rewards and actions, temporal-difference
learning, Relationship to Dynamic Programming, Active reinforcement learning, Generalization in
reinforcement learning.
Motivation, Basics of Sampling Theory: Error Estimation and Estimating Binomial Proportions, The
Binomial Distribution, Estimators, Bias, and Variance
UNIT V:
Genetic Algorithms: Motivation, Genetic Algorithms: Representing Hypotheses, Genetic Operator,
Fitness Function and Selection, An Illustrative Example, Hypothesis Space Search, Genetic
Programming, Models of Evolution and Learning: Lamarkian Evolution, Baldwin Effect, Parallelizing
Genetic Algorithms.
TEXT BOOKS:
1. Ethem Alpaydin, ”Introduction to Machine Learning”, MIT Press, Prentice Hall of India, 3 rd
Edition2014.
2. Mehryar Mohri, Afshin Rostamizadeh, Ameet Talwalkar ” Foundations of Machine Learning”, MIT
Press,2012.
3. Tom Mitchell, “Machine Learning”, McGraw Hill, 3rdEdition, 1997.
4. MACHINE LEARNING - An Algorithmic Perspective, Second Edition, Stephen Marsland, 2015.
REFERENCE BOOKS:
1. CharuC.Aggarwal,“DataClassificationAlgorithmsandApplications”,CRCPress,2014.
2. Charu C. Aggarwal, “DATA CLUSTERING Algorithms and Applications”, CRC Press,
2014.
3. Kevin P. Murphy ”Machine Learning: A Probabilistic Perspective”, The MIT Press, 2012
4. Jiawei Han and Micheline Kambers and JianPei, “Data Mining Concepts
andTechniques”,3rd edition, Morgan Kaufman Publications, 2012.
OUTCOMES:
1. Recognize the characteristics of Machine Learning techniques that enable to solve real world
problems
2. Recognize the characteristics of machine learning strategies
3. Apply various supervised learning methods to appropriate problems
4. Identify and integrate more than one techniques to enhance the performance of learning
5. Create probabilistic and unsupervised learning models for handling unknown pattern
6. Analyze the co-occurrence of data to find interesting frequent patterns
INDEX
UNIT NO TOPIC PAGE NO
Introduction 1
Learning Models 3
Types of Learning 12
I
Perspectives and Issues 13
Version Spaces 14
PAC Learning 19
VC Dimension 21
Decision Trees 23
Neural Networks 37
II
Support Vector Machines 45
Introduction to clustering 49
K-means clustering 52
EM Algorithms 69
Learning Task 79
IV Q Learning 82
Evaluating Hypotheses 86
Genetic Algorithms 92
An Illustrative Example 96
105
Parallelizing Genetic Algorithms.