SlideShare a Scribd company logo
Machine Learning
Dr. P. Kuppusamy
Prof / CSE
2
Machine Learning
Machine learning is an application of artificial intelligence (AI)
that provides systems the ability to automatically learn, think
and improve from experience without being explicitly
programmed.
3
Difference Between Traditional Programming and Machine Learning
Build a ML Model
4
ML Software Development
5
automatically learn and improve from experience
Features (Variables/Attributes) in ML
• Feature is an individual measurable attribute or characteristic of a
phenomenon being observed.
• Choosing informative, discriminating and independent features is
a crucial for effective algorithms in pattern
recognition, classification and regression.
• Features are usually numeric, but structural features such
as strings and graphs are used in syntactic pattern recognition.
• Eg. Table – Length, Breadth, Height, Weight, Color, Location,
no_of_draws, no_of _doors, Price
Features (Variables/Attributes) in ML
• Vector - collection / array of numbers in similar data type
• Feature Vector is an n-dimensional vector of
numerical features that represent some object.
• Eg. Length of 3 tables in feet
𝐿[1]
𝐿[2]
𝐿[3]
=
5
7
3
Feature extraction - definition
• Given a set of features 𝐹 = {𝑥1, … , 𝑥𝑁}
the Feature Extraction(“Construction”) problem is to map 𝐹 to
some feature set 𝐹′′ that maximizes the learner’s ability to
classify patterns
Feature Extraction
• Find a projection matrix w from n-dimensional to m-dimensional
vectors that keeps error low
𝒛 = 𝑤𝑇𝑿
w – Parameters
X – Set of features
Types of Learning
• Supervised (inductive) learning
• Training data includes desired outputs
• Unsupervised learning
• Training data does not include desired outputs
• Semi-supervised learning
• Training data includes a few desired outputs
• Reinforcement learning
• Rewards from sequence of actions
Supervised (Inductive) Learning
• Training data includes desired outputs
• Given examples of a function (X, F(X))
• Predict function F(X) for new examples X
• Discrete data - F(X): Classification
• Continuous data - F(X): Regression
• F(X) = Probability(X): Probability estimation
12
Supervised learning:
Learning a model from labeled data.
13
Supervised learning
Algorithms: Regression, Support Vector Machines, neural
networks, decision trees, K-nearest neighbors, naive Bayes, etc.
14
Unsupervised learning
Algorithms: K-means, gaussian mixtures, hierarchical clustering,
spectral clustering, etc.
Learning a model from unlabeled data.
15
Semi supervised learning:
Learning a model from unlabeled and labeled data.
Linear Regression
• Linear Regression analysis is a statistical tool
• Predictive modeling method to investigate the mathematical
relationship between a dependent variable (outcome – y) and an
independent variable (predictor – x).
• Predictor shows the changes in Dependent variable (y axis) to the
changes in explanatory variables in X axis.
Linear Regression
• It is quantitative analysis tool
• It uses current information about a phenomenon to predict its future behavior.
• Involves the graphical lines over a set of data points that most closely towards
all shape of the data.
• when the data form a set of pairs of numbers, it is interpreted as the observed
values of an independent (or predictor ) variable X and a dependent ( or
response) variable Y.
y x
  
  
0 1
Data model in Linear Regression
• Data is modelled using a straight line with continuous variable
• Relationship between variables is a linear function
Dependent
(Response)
Variable
Independent
(Explanatory)
Variable
Population
Slope
Population
y-intercept
Random
Error
y
β0 = y-intercept
x
Change
in y
Change in x
β1 = Slope
Data model in Linear Regression
Data is modelled using a straight line
Types of Relationships
Y
X
Y
X
Y
Y
X
X
Strong relationships Weak relationships
Types of Relationships
Y
X
Y
X
No relationship
(continued)
Plot for x and actual y values
 Plot the graph using x and y values
Random Error Identification
 Random Error  = Estimated Value (yi) – Actual Value (yi)
Minimize the Random Error
 Reduce the distance between estimated and actual value
 Find the best fit of the line using least square method
Least Squares Method to Minimize the Error
• ‘Best fit’ means difference between actual y values and
predicted y values are a minimum
• But positive differences off-set negative
 
2 2
1 1
ˆ ˆ
n n
i
i i
i i
y y 
 
 
 
• Least Squares minimizes the Sum of the Squared Differences
(SSE)
Least Squares Graphically
2
y
x
1 3
4
^
^
^
^
2 0 1 2 2
ˆ ˆ ˆ
y x
  
  
0 1
ˆ ˆ
ˆi i
y x
 
 
2 2 2 2 2
1 2 3 4
1
ˆ ˆ ˆ ˆ ˆ
LS minimizes
n
i
i
    

   

Case Study
 Let consider x and y values and mark in scatter plot
 Find mean of x, and mean of y
 Find the coefficients m and c in the straight line y = mx+c
 Find x-x1, and y-y1
 Find m
Plot the x, y values in the graph
x = {1, 2, 3, 4, 5} y = {3, 4, 2, 4, 5}
Plot the regression line using estimated y values
x = {1, 2, 3, 4, 5} y = {2.8, 3.2, 3.6, 4, 4.4}
Estimated y values
Estimated y values
Find the Error 
Mean Square Error minimizes the
error in the linear regression.
Regression Line with least error is
the ‘best fit’ line
 
2 2
1 1
ˆ ˆ
n n
i
i i
i i
y y 
 
 
 
How would you draw a line through the points in real time?
 Initial values (iteration 0) for slope m = 0 and y-intercept b = 0
How would you draw a line through the points?
iteration 1, slope m = 0.04 and y-
intercept b = 0
iteration 20, slope m = 0.59 and y-
intercept b = 0.01
Determine which line ‘fits best’ in 100 iterations
iteration 47, slope m = 1.03 and y-
intercept b = 0.02
iteration 99, slope m = 1.36 and y-
intercept b = 0.03
3 major Uses of Regression
•Determining the strength of predictors
•Forecasting an effect
•Trend forecasting
Where Linear Regression used?
• Evaluating trends and sales estimates
• Analyze the impact of price changes
• Insurance domain
Squared Error Cost Function
• Cost Function - J(ɵ) =
1
2𝑚
σ𝑖=1
𝑚
(𝑌(𝑖)
− 𝑦′ 𝑖
)2
𝑌(𝑖) - Ground truths or Actual output or label
𝑦′ 𝑖
- Prediction output
m - No. of data points or samples
Gradient Descent
• The objective of training a machine learning model is to minimize the loss or error
between ground truths and predictions by changing the trainable parameters.
• Gradient is the extension of derivative in multi-dimensional space, tells the direction
along which the loss or error is optimally minimized.
• Gradient is defined as the maximum rate of change.
𝜃𝑗 = 𝜃𝑗 − 𝛼
𝜕
𝜕𝜃𝑗
𝐽(𝜃)
• 𝜃𝑗-Training parameter 𝛼 – Learning rate 𝐽(𝜃) – Error / Cost function
Gradient Descent
• Gradient Descent:
𝜃𝑗 = 𝜃𝑗 − 𝛼
1
𝑚
σ𝑖=1
𝑚
(𝑦′(𝑖)
−𝑌 𝑖 )𝑥𝑗
𝑖
for All j
j=0; 𝜃0 = 𝜃0 − 𝛼
1
𝑚
σ𝑖=1
𝑚
(𝑦′(𝑖)
−𝑌 𝑖 )𝑥0
𝑖
for All j
j=1; 𝜃1 = 𝜃1 − 𝛼
1
𝑚
σ𝑖=1
𝑚
(𝑦′(𝑖)
−𝑌 𝑖 )𝑥1
𝑖
for All j
………..
j=n; 𝜃𝑛 = 𝜃𝑛 − 𝛼
1
𝑚
σ𝑖=1
𝑚
(𝑦′(𝑖)
−𝑌 𝑖 )𝑥𝑛
𝑖 for All j
References
• Tom Markiewicz& Josh Zheng,Getting started with Artificial
Intelligence, Published by O’Reilly Media,2017
• Stuart J. Russell and Peter Norvig,Artificial Intelligence A Modern
Approach
• Richard Szeliski, Computer Vision: Algorithms and Applications,
Springer 2010
• Artificial Intelligence and Machine Learning, Chandra S.S. & H.S.
Anand, PHI Publications
• Machine Learning, Rajiv Chopra, Khanna Publishing House
•

More Related Content

PDF
Feature Engineering in Machine Learning
PDF
Logistic regression in Machine Learning
PPTX
Hyperparameter Tuning
ODP
Machine Learning With Logistic Regression
PDF
Unsupervised Learning in Machine Learning
PPSX
ADABoost classifier
PPTX
Machine Learning
PPTX
Data preprocessing in Machine learning
Feature Engineering in Machine Learning
Logistic regression in Machine Learning
Hyperparameter Tuning
Machine Learning With Logistic Regression
Unsupervised Learning in Machine Learning
ADABoost classifier
Machine Learning
Data preprocessing in Machine learning

What's hot (20)

PPTX
Feature Selection in Machine Learning
PDF
Loss Functions for Deep Learning - Javier Ruiz Hidalgo - UPC Barcelona 2018
PPT
Graph colouring
PPTX
Logistic Regression | Logistic Regression In Python | Machine Learning Algori...
PDF
Deep Dive into Hyperparameter Tuning
PDF
Dimensionality Reduction
PPTX
Machine Learning: Bias and Variance Trade-off
PPTX
Introduction to Machine Learning
PDF
Logistic Regression in Python | Logistic Regression Example | Machine Learnin...
PDF
Decision tree
PPTX
k medoid clustering.pptx
PPTX
Introduction to Machine Learning
PPTX
Analysis of Feature Selection Algorithms (Branch & Bound and Beam search)
PDF
Understanding Bagging and Boosting
PPT
Apriori and Eclat algorithm in Association Rule Mining
PPTX
Supervised and Unsupervised Learning In Machine Learning | Machine Learning T...
PDF
Loss functions (DLAI D4L2 2017 UPC Deep Learning for Artificial Intelligence)
PPTX
Linear Regression and Logistic Regression in ML
PDF
Variational Autoencoders VAE - Santiago Pascual - UPC Barcelona 2018
PDF
Discrete Mathematics in Real Life ppt.pdf
Feature Selection in Machine Learning
Loss Functions for Deep Learning - Javier Ruiz Hidalgo - UPC Barcelona 2018
Graph colouring
Logistic Regression | Logistic Regression In Python | Machine Learning Algori...
Deep Dive into Hyperparameter Tuning
Dimensionality Reduction
Machine Learning: Bias and Variance Trade-off
Introduction to Machine Learning
Logistic Regression in Python | Logistic Regression Example | Machine Learnin...
Decision tree
k medoid clustering.pptx
Introduction to Machine Learning
Analysis of Feature Selection Algorithms (Branch & Bound and Beam search)
Understanding Bagging and Boosting
Apriori and Eclat algorithm in Association Rule Mining
Supervised and Unsupervised Learning In Machine Learning | Machine Learning T...
Loss functions (DLAI D4L2 2017 UPC Deep Learning for Artificial Intelligence)
Linear Regression and Logistic Regression in ML
Variational Autoencoders VAE - Santiago Pascual - UPC Barcelona 2018
Discrete Mathematics in Real Life ppt.pdf
Ad

Similar to Machine learning Introduction (20)

PPT
Machine Learning Unit 2_Supervised Learning
PPTX
regression analysis presentation slides.
PDF
Module 5.pdf Machine Learning Types and examples
PDF
[Paper reading] L-SHAPLEY AND C-SHAPLEY: EFFICIENT MODEL INTERPRETATION FOR S...
PPTX
Unit-1 Introduction and Mathematical Preliminaries.pptx
PDF
Linear Regression
PPTX
Application of Machine Learning in Agriculture
PPTX
Unit 3 – AIML.pptx
PPTX
AI & ML(Unit III).pptx.It contains also syllabus
PPTX
cs 601 - lecture 1.pptx
PPTX
Different Types of Machine Learning Algorithms
PPTX
Data Science and Machine Learning with Tensorflow
PPTX
Curve Fitting
PDF
Lecture 5 - Linear Regression Linear Regression
PPTX
Regression
PDF
Machine Learning Notes for beginners ,Step by step
PDF
Lecture 4 - Linear Regression, a lecture in subject module Statistical & Mach...
PDF
Simple linear regression
PPTX
Machine Learning techniques used in AI.
PPTX
Linear Regression.pptx
Machine Learning Unit 2_Supervised Learning
regression analysis presentation slides.
Module 5.pdf Machine Learning Types and examples
[Paper reading] L-SHAPLEY AND C-SHAPLEY: EFFICIENT MODEL INTERPRETATION FOR S...
Unit-1 Introduction and Mathematical Preliminaries.pptx
Linear Regression
Application of Machine Learning in Agriculture
Unit 3 – AIML.pptx
AI & ML(Unit III).pptx.It contains also syllabus
cs 601 - lecture 1.pptx
Different Types of Machine Learning Algorithms
Data Science and Machine Learning with Tensorflow
Curve Fitting
Lecture 5 - Linear Regression Linear Regression
Regression
Machine Learning Notes for beginners ,Step by step
Lecture 4 - Linear Regression, a lecture in subject module Statistical & Mach...
Simple linear regression
Machine Learning techniques used in AI.
Linear Regression.pptx
Ad

More from Kuppusamy P (20)

PDF
Recurrent neural networks rnn
PDF
Deep learning
PDF
Image segmentation
PDF
Image enhancement
PDF
Feature detection and matching
PDF
Image processing, Noise, Noise Removal filters
PDF
Flowchart design for algorithms
PDF
Algorithm basics
PDF
Problem solving using Programming
PDF
Parts of Computer, Hardware and Software
PDF
Strings in java
PDF
Java methods or Subroutines or Functions
PDF
Java arrays
PDF
Java iterative statements
PDF
Java conditional statements
PDF
Java data types
PDF
Java introduction
PDF
Anomaly detection (Unsupervised Learning) in Machine Learning
PDF
Machine Learning Performance metrics for classification
PDF
Reinforcement learning, Q-Learning
Recurrent neural networks rnn
Deep learning
Image segmentation
Image enhancement
Feature detection and matching
Image processing, Noise, Noise Removal filters
Flowchart design for algorithms
Algorithm basics
Problem solving using Programming
Parts of Computer, Hardware and Software
Strings in java
Java methods or Subroutines or Functions
Java arrays
Java iterative statements
Java conditional statements
Java data types
Java introduction
Anomaly detection (Unsupervised Learning) in Machine Learning
Machine Learning Performance metrics for classification
Reinforcement learning, Q-Learning

Recently uploaded (20)

PPTX
Week 4 Term 3 Study Techniques revisited.pptx
PPTX
Pharmacology of Heart Failure /Pharmacotherapy of CHF
PDF
TR - Agricultural Crops Production NC III.pdf
PPTX
Cell Structure & Organelles in detailed.
PPTX
Renaissance Architecture: A Journey from Faith to Humanism
PDF
The Lost Whites of Pakistan by Jahanzaib Mughal.pdf
PDF
Abdominal Access Techniques with Prof. Dr. R K Mishra
PDF
Anesthesia in Laparoscopic Surgery in India
PDF
Basic Mud Logging Guide for educational purpose
PPTX
human mycosis Human fungal infections are called human mycosis..pptx
PPTX
Final Presentation General Medicine 03-08-2024.pptx
PDF
Complications of Minimal Access Surgery at WLH
PPTX
master seminar digital applications in india
PDF
Business Ethics Teaching Materials for college
PDF
ANTIBIOTICS.pptx.pdf………………… xxxxxxxxxxxxx
PPTX
Microbial diseases, their pathogenesis and prophylaxis
PDF
Pre independence Education in Inndia.pdf
PDF
Saundersa Comprehensive Review for the NCLEX-RN Examination.pdf
PPTX
Introduction to Child Health Nursing – Unit I | Child Health Nursing I | B.Sc...
PDF
Origin of periodic table-Mendeleev’s Periodic-Modern Periodic table
Week 4 Term 3 Study Techniques revisited.pptx
Pharmacology of Heart Failure /Pharmacotherapy of CHF
TR - Agricultural Crops Production NC III.pdf
Cell Structure & Organelles in detailed.
Renaissance Architecture: A Journey from Faith to Humanism
The Lost Whites of Pakistan by Jahanzaib Mughal.pdf
Abdominal Access Techniques with Prof. Dr. R K Mishra
Anesthesia in Laparoscopic Surgery in India
Basic Mud Logging Guide for educational purpose
human mycosis Human fungal infections are called human mycosis..pptx
Final Presentation General Medicine 03-08-2024.pptx
Complications of Minimal Access Surgery at WLH
master seminar digital applications in india
Business Ethics Teaching Materials for college
ANTIBIOTICS.pptx.pdf………………… xxxxxxxxxxxxx
Microbial diseases, their pathogenesis and prophylaxis
Pre independence Education in Inndia.pdf
Saundersa Comprehensive Review for the NCLEX-RN Examination.pdf
Introduction to Child Health Nursing – Unit I | Child Health Nursing I | B.Sc...
Origin of periodic table-Mendeleev’s Periodic-Modern Periodic table

Machine learning Introduction

  • 1. Machine Learning Dr. P. Kuppusamy Prof / CSE
  • 2. 2 Machine Learning Machine learning is an application of artificial intelligence (AI) that provides systems the ability to automatically learn, think and improve from experience without being explicitly programmed.
  • 3. 3 Difference Between Traditional Programming and Machine Learning
  • 4. Build a ML Model 4
  • 5. ML Software Development 5 automatically learn and improve from experience
  • 6. Features (Variables/Attributes) in ML • Feature is an individual measurable attribute or characteristic of a phenomenon being observed. • Choosing informative, discriminating and independent features is a crucial for effective algorithms in pattern recognition, classification and regression. • Features are usually numeric, but structural features such as strings and graphs are used in syntactic pattern recognition. • Eg. Table – Length, Breadth, Height, Weight, Color, Location, no_of_draws, no_of _doors, Price
  • 7. Features (Variables/Attributes) in ML • Vector - collection / array of numbers in similar data type • Feature Vector is an n-dimensional vector of numerical features that represent some object. • Eg. Length of 3 tables in feet 𝐿[1] 𝐿[2] 𝐿[3] = 5 7 3
  • 8. Feature extraction - definition • Given a set of features 𝐹 = {𝑥1, … , 𝑥𝑁} the Feature Extraction(“Construction”) problem is to map 𝐹 to some feature set 𝐹′′ that maximizes the learner’s ability to classify patterns
  • 9. Feature Extraction • Find a projection matrix w from n-dimensional to m-dimensional vectors that keeps error low 𝒛 = 𝑤𝑇𝑿 w – Parameters X – Set of features
  • 10. Types of Learning • Supervised (inductive) learning • Training data includes desired outputs • Unsupervised learning • Training data does not include desired outputs • Semi-supervised learning • Training data includes a few desired outputs • Reinforcement learning • Rewards from sequence of actions
  • 11. Supervised (Inductive) Learning • Training data includes desired outputs • Given examples of a function (X, F(X)) • Predict function F(X) for new examples X • Discrete data - F(X): Classification • Continuous data - F(X): Regression • F(X) = Probability(X): Probability estimation
  • 12. 12 Supervised learning: Learning a model from labeled data.
  • 13. 13 Supervised learning Algorithms: Regression, Support Vector Machines, neural networks, decision trees, K-nearest neighbors, naive Bayes, etc.
  • 14. 14 Unsupervised learning Algorithms: K-means, gaussian mixtures, hierarchical clustering, spectral clustering, etc. Learning a model from unlabeled data.
  • 15. 15 Semi supervised learning: Learning a model from unlabeled and labeled data.
  • 16. Linear Regression • Linear Regression analysis is a statistical tool • Predictive modeling method to investigate the mathematical relationship between a dependent variable (outcome – y) and an independent variable (predictor – x). • Predictor shows the changes in Dependent variable (y axis) to the changes in explanatory variables in X axis.
  • 17. Linear Regression • It is quantitative analysis tool • It uses current information about a phenomenon to predict its future behavior. • Involves the graphical lines over a set of data points that most closely towards all shape of the data. • when the data form a set of pairs of numbers, it is interpreted as the observed values of an independent (or predictor ) variable X and a dependent ( or response) variable Y.
  • 18. y x       0 1 Data model in Linear Regression • Data is modelled using a straight line with continuous variable • Relationship between variables is a linear function Dependent (Response) Variable Independent (Explanatory) Variable Population Slope Population y-intercept Random Error
  • 19. y β0 = y-intercept x Change in y Change in x β1 = Slope Data model in Linear Regression Data is modelled using a straight line
  • 20. Types of Relationships Y X Y X Y Y X X Strong relationships Weak relationships
  • 21. Types of Relationships Y X Y X No relationship (continued)
  • 22. Plot for x and actual y values  Plot the graph using x and y values
  • 23. Random Error Identification  Random Error  = Estimated Value (yi) – Actual Value (yi)
  • 24. Minimize the Random Error  Reduce the distance between estimated and actual value  Find the best fit of the line using least square method
  • 25. Least Squares Method to Minimize the Error • ‘Best fit’ means difference between actual y values and predicted y values are a minimum • But positive differences off-set negative   2 2 1 1 ˆ ˆ n n i i i i i y y        • Least Squares minimizes the Sum of the Squared Differences (SSE)
  • 26. Least Squares Graphically 2 y x 1 3 4 ^ ^ ^ ^ 2 0 1 2 2 ˆ ˆ ˆ y x       0 1 ˆ ˆ ˆi i y x     2 2 2 2 2 1 2 3 4 1 ˆ ˆ ˆ ˆ ˆ LS minimizes n i i           
  • 27. Case Study  Let consider x and y values and mark in scatter plot
  • 28.  Find mean of x, and mean of y
  • 29.  Find the coefficients m and c in the straight line y = mx+c
  • 30.  Find x-x1, and y-y1
  • 32. Plot the x, y values in the graph x = {1, 2, 3, 4, 5} y = {3, 4, 2, 4, 5}
  • 33. Plot the regression line using estimated y values x = {1, 2, 3, 4, 5} y = {2.8, 3.2, 3.6, 4, 4.4} Estimated y values Estimated y values
  • 34. Find the Error  Mean Square Error minimizes the error in the linear regression. Regression Line with least error is the ‘best fit’ line   2 2 1 1 ˆ ˆ n n i i i i i y y       
  • 35. How would you draw a line through the points in real time?  Initial values (iteration 0) for slope m = 0 and y-intercept b = 0
  • 36. How would you draw a line through the points? iteration 1, slope m = 0.04 and y- intercept b = 0 iteration 20, slope m = 0.59 and y- intercept b = 0.01
  • 37. Determine which line ‘fits best’ in 100 iterations iteration 47, slope m = 1.03 and y- intercept b = 0.02 iteration 99, slope m = 1.36 and y- intercept b = 0.03
  • 38. 3 major Uses of Regression •Determining the strength of predictors •Forecasting an effect •Trend forecasting
  • 39. Where Linear Regression used? • Evaluating trends and sales estimates • Analyze the impact of price changes • Insurance domain
  • 40. Squared Error Cost Function • Cost Function - J(ɵ) = 1 2𝑚 σ𝑖=1 𝑚 (𝑌(𝑖) − 𝑦′ 𝑖 )2 𝑌(𝑖) - Ground truths or Actual output or label 𝑦′ 𝑖 - Prediction output m - No. of data points or samples
  • 41. Gradient Descent • The objective of training a machine learning model is to minimize the loss or error between ground truths and predictions by changing the trainable parameters. • Gradient is the extension of derivative in multi-dimensional space, tells the direction along which the loss or error is optimally minimized. • Gradient is defined as the maximum rate of change. 𝜃𝑗 = 𝜃𝑗 − 𝛼 𝜕 𝜕𝜃𝑗 𝐽(𝜃) • 𝜃𝑗-Training parameter 𝛼 – Learning rate 𝐽(𝜃) – Error / Cost function
  • 42. Gradient Descent • Gradient Descent: 𝜃𝑗 = 𝜃𝑗 − 𝛼 1 𝑚 σ𝑖=1 𝑚 (𝑦′(𝑖) −𝑌 𝑖 )𝑥𝑗 𝑖 for All j j=0; 𝜃0 = 𝜃0 − 𝛼 1 𝑚 σ𝑖=1 𝑚 (𝑦′(𝑖) −𝑌 𝑖 )𝑥0 𝑖 for All j j=1; 𝜃1 = 𝜃1 − 𝛼 1 𝑚 σ𝑖=1 𝑚 (𝑦′(𝑖) −𝑌 𝑖 )𝑥1 𝑖 for All j ……….. j=n; 𝜃𝑛 = 𝜃𝑛 − 𝛼 1 𝑚 σ𝑖=1 𝑚 (𝑦′(𝑖) −𝑌 𝑖 )𝑥𝑛 𝑖 for All j
  • 43. References • Tom Markiewicz& Josh Zheng,Getting started with Artificial Intelligence, Published by O’Reilly Media,2017 • Stuart J. Russell and Peter Norvig,Artificial Intelligence A Modern Approach • Richard Szeliski, Computer Vision: Algorithms and Applications, Springer 2010 • Artificial Intelligence and Machine Learning, Chandra S.S. & H.S. Anand, PHI Publications • Machine Learning, Rajiv Chopra, Khanna Publishing House •