SlideShare a Scribd company logo
Linear Regression
What is Regression?
Regression is a supervised learning technique which helps in finding the
correlation between variables and enables us to predict the continuous
output variable based on the one or more predictor variables.
It is mainly used for prediction, forecasting, time series modeling, and
determining the causal-effect relationship between variables. Regression
analysis helps in the prediction of a continuous variable.
Types of Regression
What is Linear Regression?
❏ Linear regression analysis is used to predict the value of a
variable based on the value of another variable
❏ Linear regression is a type of supervised machine learning
algorithm that computes the linear relationship between a
dependent variable and one or more independent
features.
Types of linear regression
❏ Simple linear regression : The goal of a simple linear regression is to
predict the value of a dependent variable based on an independent
variable. Simple linear regression is used to model the relationship
between two continuous variables. Often, the objective is to predict
the value of an output variable (or response) based on the value of
an input (or predictor) variable.
❏ Multiple linear regression: Multiple linear regression is used to
estimate the relationship between two or more independent
variables and one dependent variable.
Types of linear regression
Equation of Simple linear regression
Y= mX+b
❏ Y represents the dependent variable
❏ X represents the independent variable
❏ m is the slope of the line( how much Y changes
for a unit change in X)
❏ b is the intercept ( the value of Y when X is 0)
Example : Predicting Pizza Prices
Diameter (X) in
Inches
Price(Y) in Dollars
8 10
10 13
12 16
What will be the 20-inch pizza price?
Example : Predicting Pizza Prices
m= sum of product of deviations / sum of square of deviation for x
= 12/8
= 1.5
b= Mean of Y - (m* Mean of X)
= 13 - (1.5*10)
= -2
Y= mX+b
= 1.5X- 2
= 1.5*20-2
= 28 Dollar
Multiple linear regression formula
Example of Multiple Linear Regression
Here, the Matrices for Y and X are
given as follows
Example of Multiple Linear Regression
The Coefficient of the Multiple Regression Equation is given as
Example of Multiple Linear Regression
Example of Multiple Linear Regression
Cost Function in Linear Regression
For the Linear regression model, the cost function will be the minimum
of the Root Mean Squared Error of the model, obtained by subtracting
the predicted values from actual values. It is a numeric value that
indicates the success or failure of a particular model without needing to
understand the inner workings of a model.
Why is the cost function needed?
Basically, our machine learning model predicts the new value. We aim to have
the predicted value as close as possible to the predicted value. If the model
predicts the value close to the actual value, then we will say it’s a good model.
So here is where cost function plays an important role. This cost function is
needed to calculate the difference between actual and predicted values. So
here it is nothing, just the difference between the actual values-predicted
values.
Types of Cost Function in Linear
Regression
Mean error: These errors can be negative or positive. Therefore, they can cancel each
other out during the summation, and the average error of the model will be zero.
Mean Squared Error : MSE represents the average squared difference between the
predictions and expected results.
I = index of sample
Ŷ = predicted value
Y = expected value
M = number of samples in the data set
Types of Cost Function in Linear Regression
Mean Absolute Error: MSE is a variation of MAE that squares the
difference instead of taking the absolute value of the difference.
There is no possibility of negative errors.
I = index of sample
Ŷ = predicted value
Y = expected value
M = number of
samples in the data
set
Types of Cost Function in Linear Regression
Root mean squared error (RMSE): The Root Mean Squared Error (RMSE)
is one of the two main performance indicators for a regression model. It
measures the average difference between values predicted by a model
and the actual values.
Exploring the Cost Function
We have a training set with three points (1, 1), (2, 2), and (3, 3). We plot the
function f(x) = w * x for different values of w and calculate the corresponding cost
function J(w). When w = 1: f(x) = x, the line passes through the origin, perfectly
fitting the data. The cost function J(w) is 0 since f(x) equals y for all training
examples. Setting w = 0.5: f(x) = 0.5 * x, the line has a smaller slope. The cost
function J(w) now measures the squared errors between f(x) and y for each
example. It provides a measure of how well the line fits the data.
Function f(x) vs. Cost Function J(w)
Real-life example of cost function
Here, y denotes the target variable(Price of House), X denotes the size of the
House and a0 is the intercept. The linear regression model will fit a straight line
to our data, as shown in fig, and it will tell us what is the price of the House when
we know the size of the House. Suppose this housing company sells houses
having areas under 600 to 1500 square feet. So the prices will vary according to
the size of the plot. For example
House(1200 to 1500 square feet)= 65 lakhs.
Now suddenly, there’s a rise in demand of Houses which means the price of this
House will increase now. Suppose the price has now increased to 75 lakhs. But
what if the model predicted its price as 30 lakhs? So this difference between
these actual and predicted values is calculated by the cost function.
Applications of Linear Regression
❏ Market analysis
❏ Financial analysis
❏ Sports analysis
❏ Environmental health
❏ Medicine
Advantages of Linear Regression
❏ Linear Regression is simple to implement and easier to interpret the output
coefficients
❏ When you know the relationship between the independent and dependent
variable have a linear relationship, this algorithm is the best to use because of
it’s less complexity compared to other algorithms.
❏ Linear Regression is susceptible to over-fitting but it can be avoided using some
dimensionality reduction techniques, regularization (L1 and L2) techniques and
cross-validation.
Drawbacks of Linear Regression
❏ Linear Assumption
❏ Sensitivity to Exceptions
❏ Independence Presumption
❏ Limited to Linear Relationships
Conclusion
Linear regression is a foundational and widely used
technique in machine learning, offering a straightforward
yet effective way to model relationships between variables.
It is a valuable tool for making predictions and
understanding data patterns, though its suitability depends
on the specific problem and the linearity assumption. While
it may not always be the best choice, it remains an essential
component of the machine learning toolkit, providing
valuable insights in various domains.
References
❖ https://www.geeksforgeeks.org/ml-advantages-and-disadvantages-of-linear-
regression/
❖ https://iq.opengenus.org/advantages-and-disadvantages-of-linear-regression/
❖ https://www.intellspot.com/linear-regression-examples/
❖ https://www.tutorialspoint.com/advantages-and-disadvantages-of-linear-
regression#:~:text=The%20strategy%20is%20simple%20to,the%20subordinate%20and
%20free%20factors.
❖ https://sphweb.bumc.bu.edu/otlt/MPH-Modules/BS/BS704-
EP713_MultivariableMethods/
❖ https://www.javatpoint.com/simple-linear-regression-in-machine-learning
❖ https://youtu.be/zUQr6HAAKp4?si=V-ojf6pXnMO2p7DR
❖ https://youtu.be/lzGKRSvs5HM?si=RduTOfWWjb05n5tF
❖ https://www.shiksha.com/online-courses/articles/cost-function-in-machine-
learning/#Why-is-the-cost-function-needed?
❖ https://medium.com/@yennhi95zz/3-understanding-the-cost-function-in-linear-regression-for-
machine-learning-beginners-ec9edeecbdde
❖ https://datatab.net/tutorial/linear-regression
❖ https://www.youtube.com/watch?v=QcPycBZomac
Thank You

More Related Content

PPTX
Machine Learning-Linear regression
PPT
Linear regression
PDF
Introduction to HTML-CSS-Javascript.pdf
PPTX
Costs functions
PPTX
Boxplot
PDF
Linear regression
PDF
Introduction to Real Analysis 4th Edition Bartle Solutions Manual
PPTX
Statistical distributions
Machine Learning-Linear regression
Linear regression
Introduction to HTML-CSS-Javascript.pdf
Costs functions
Boxplot
Linear regression
Introduction to Real Analysis 4th Edition Bartle Solutions Manual
Statistical distributions

What's hot (20)

PDF
Statistical Distributions
PPTX
ML - Multiple Linear Regression
PPTX
Polynomial regression
PPTX
Logistical Regression.pptx
PDF
Regression Analysis
PPTX
Poisson distribution
PPTX
Statistical Estimation
PPTX
Regression analysis
PPTX
Regression
PPTX
Regression ppt
PDF
Regression analysis and its type
PPTX
Principal Component Analysis PCA
PPT
Correlation
PPTX
Linear regression
PDF
Correlation Analysis
PPTX
Multivariate analysis
PPT
Simple linear regression
PPTX
Regression analysis
PPTX
Probability
Statistical Distributions
ML - Multiple Linear Regression
Polynomial regression
Logistical Regression.pptx
Regression Analysis
Poisson distribution
Statistical Estimation
Regression analysis
Regression
Regression ppt
Regression analysis and its type
Principal Component Analysis PCA
Correlation
Linear regression
Correlation Analysis
Multivariate analysis
Simple linear regression
Regression analysis
Probability
Ad

Similar to Linear Regression (20)

PPTX
Machine learning session4(linear regression)
DOCX
NPTEL Machine Learning Week 2.docx
PDF
The normal presentation about linear regression in machine learning
PPTX
Unit 03 - Consolidated.pptx
PDF
Supervised Learning.pdf
PPTX
Ca-1 assignment Machine learning.ygygygpptx
PDF
ML_Lec3 introduction to regression problems.pdf
PPTX
linearregression-1909240jhgg53948.pptx
PPTX
AI & ML(Unit III).pptx.It contains also syllabus
PDF
working with python
PPT
Simple Linear Regression.pptSimple Linear Regression.ppt
PPTX
Bootcamp of new world to taken seriously
PPT
Simple lin regress_inference
DOCX
Correlation and regression in r
PPTX
regression analysis presentation slides.
PPSX
Lasso and ridge regression
PDF
Unit---5.pdf of ba in srcc du gst before exam
PPTX
Chapter III.pptx
PPTX
Linear_Regression
PPTX
Detail Study of the concept of Regression model.pptx
Machine learning session4(linear regression)
NPTEL Machine Learning Week 2.docx
The normal presentation about linear regression in machine learning
Unit 03 - Consolidated.pptx
Supervised Learning.pdf
Ca-1 assignment Machine learning.ygygygpptx
ML_Lec3 introduction to regression problems.pdf
linearregression-1909240jhgg53948.pptx
AI & ML(Unit III).pptx.It contains also syllabus
working with python
Simple Linear Regression.pptSimple Linear Regression.ppt
Bootcamp of new world to taken seriously
Simple lin regress_inference
Correlation and regression in r
regression analysis presentation slides.
Lasso and ridge regression
Unit---5.pdf of ba in srcc du gst before exam
Chapter III.pptx
Linear_Regression
Detail Study of the concept of Regression model.pptx
Ad

More from Abdullah al Mamun (20)

PPTX
Underfitting and Overfitting in Machine Learning
PPTX
Recurrent Neural Networks (RNNs)
PPTX
Random Forest
PPTX
Natural Language Processing (NLP)
PPTX
Naive Bayes
PPTX
Multilayer Perceptron Neural Network MLP
PPTX
Long Short Term Memory LSTM
PPTX
K-Nearest Neighbor(KNN)
PPTX
Hidden Markov Model (HMM)
PPTX
Ensemble Method (Bagging Boosting)
PPTX
Convolutional Neural Networks CNN
PPTX
Artificial Neural Network ANN
PPTX
Reinforcement Learning, Application and Q-Learning
PPTX
Session on evaluation of DevSecOps
PPTX
Artificial Intelligence: Classification, Applications, Opportunities, and Cha...
PPTX
DevOps Presentation.pptx
PPTX
Python Virtual Environment.pptx
PPTX
Artificial intelligence Presentation.pptx
PPT
An approach to empirical Optical Character recognition paradigm using Multi-L...
PPT
Automatic Speaker Recognition system using MFCC and VQ approach
Underfitting and Overfitting in Machine Learning
Recurrent Neural Networks (RNNs)
Random Forest
Natural Language Processing (NLP)
Naive Bayes
Multilayer Perceptron Neural Network MLP
Long Short Term Memory LSTM
K-Nearest Neighbor(KNN)
Hidden Markov Model (HMM)
Ensemble Method (Bagging Boosting)
Convolutional Neural Networks CNN
Artificial Neural Network ANN
Reinforcement Learning, Application and Q-Learning
Session on evaluation of DevSecOps
Artificial Intelligence: Classification, Applications, Opportunities, and Cha...
DevOps Presentation.pptx
Python Virtual Environment.pptx
Artificial intelligence Presentation.pptx
An approach to empirical Optical Character recognition paradigm using Multi-L...
Automatic Speaker Recognition system using MFCC and VQ approach

Recently uploaded (20)

PDF
22.Patil - Early prediction of Alzheimer’s disease using convolutional neural...
PPTX
Database Infoormation System (DBIS).pptx
PPTX
IB Computer Science - Internal Assessment.pptx
PDF
Galatica Smart Energy Infrastructure Startup Pitch Deck
PPTX
Data_Analytics_and_PowerBI_Presentation.pptx
PPTX
climate analysis of Dhaka ,Banglades.pptx
PPTX
Introduction-to-Cloud-ComputingFinal.pptx
PPTX
MODULE 8 - DISASTER risk PREPAREDNESS.pptx
PPTX
Acceptance and paychological effects of mandatory extra coach I classes.pptx
PDF
Launch Your Data Science Career in Kochi – 2025
PPTX
ALIMENTARY AND BILIARY CONDITIONS 3-1.pptx
PDF
Clinical guidelines as a resource for EBP(1).pdf
PPTX
Introduction to Knowledge Engineering Part 1
PPTX
Moving the Public Sector (Government) to a Digital Adoption
PDF
Foundation of Data Science unit number two notes
PPTX
iec ppt-1 pptx icmr ppt on rehabilitation.pptx
PPTX
oil_refinery_comprehensive_20250804084928 (1).pptx
PPTX
Supervised vs unsupervised machine learning algorithms
PPTX
Business Ppt On Nestle.pptx huunnnhhgfvu
22.Patil - Early prediction of Alzheimer’s disease using convolutional neural...
Database Infoormation System (DBIS).pptx
IB Computer Science - Internal Assessment.pptx
Galatica Smart Energy Infrastructure Startup Pitch Deck
Data_Analytics_and_PowerBI_Presentation.pptx
climate analysis of Dhaka ,Banglades.pptx
Introduction-to-Cloud-ComputingFinal.pptx
MODULE 8 - DISASTER risk PREPAREDNESS.pptx
Acceptance and paychological effects of mandatory extra coach I classes.pptx
Launch Your Data Science Career in Kochi – 2025
ALIMENTARY AND BILIARY CONDITIONS 3-1.pptx
Clinical guidelines as a resource for EBP(1).pdf
Introduction to Knowledge Engineering Part 1
Moving the Public Sector (Government) to a Digital Adoption
Foundation of Data Science unit number two notes
iec ppt-1 pptx icmr ppt on rehabilitation.pptx
oil_refinery_comprehensive_20250804084928 (1).pptx
Supervised vs unsupervised machine learning algorithms
Business Ppt On Nestle.pptx huunnnhhgfvu

Linear Regression

  • 2. What is Regression? Regression is a supervised learning technique which helps in finding the correlation between variables and enables us to predict the continuous output variable based on the one or more predictor variables. It is mainly used for prediction, forecasting, time series modeling, and determining the causal-effect relationship between variables. Regression analysis helps in the prediction of a continuous variable.
  • 4. What is Linear Regression? ❏ Linear regression analysis is used to predict the value of a variable based on the value of another variable ❏ Linear regression is a type of supervised machine learning algorithm that computes the linear relationship between a dependent variable and one or more independent features.
  • 5. Types of linear regression ❏ Simple linear regression : The goal of a simple linear regression is to predict the value of a dependent variable based on an independent variable. Simple linear regression is used to model the relationship between two continuous variables. Often, the objective is to predict the value of an output variable (or response) based on the value of an input (or predictor) variable. ❏ Multiple linear regression: Multiple linear regression is used to estimate the relationship between two or more independent variables and one dependent variable.
  • 6. Types of linear regression
  • 7. Equation of Simple linear regression Y= mX+b ❏ Y represents the dependent variable ❏ X represents the independent variable ❏ m is the slope of the line( how much Y changes for a unit change in X) ❏ b is the intercept ( the value of Y when X is 0)
  • 8. Example : Predicting Pizza Prices Diameter (X) in Inches Price(Y) in Dollars 8 10 10 13 12 16 What will be the 20-inch pizza price?
  • 9. Example : Predicting Pizza Prices m= sum of product of deviations / sum of square of deviation for x = 12/8 = 1.5 b= Mean of Y - (m* Mean of X) = 13 - (1.5*10) = -2 Y= mX+b = 1.5X- 2 = 1.5*20-2 = 28 Dollar
  • 11. Example of Multiple Linear Regression Here, the Matrices for Y and X are given as follows
  • 12. Example of Multiple Linear Regression The Coefficient of the Multiple Regression Equation is given as
  • 13. Example of Multiple Linear Regression
  • 14. Example of Multiple Linear Regression
  • 15. Cost Function in Linear Regression For the Linear regression model, the cost function will be the minimum of the Root Mean Squared Error of the model, obtained by subtracting the predicted values from actual values. It is a numeric value that indicates the success or failure of a particular model without needing to understand the inner workings of a model.
  • 16. Why is the cost function needed? Basically, our machine learning model predicts the new value. We aim to have the predicted value as close as possible to the predicted value. If the model predicts the value close to the actual value, then we will say it’s a good model. So here is where cost function plays an important role. This cost function is needed to calculate the difference between actual and predicted values. So here it is nothing, just the difference between the actual values-predicted values.
  • 17. Types of Cost Function in Linear Regression Mean error: These errors can be negative or positive. Therefore, they can cancel each other out during the summation, and the average error of the model will be zero. Mean Squared Error : MSE represents the average squared difference between the predictions and expected results. I = index of sample Ŷ = predicted value Y = expected value M = number of samples in the data set
  • 18. Types of Cost Function in Linear Regression Mean Absolute Error: MSE is a variation of MAE that squares the difference instead of taking the absolute value of the difference. There is no possibility of negative errors. I = index of sample Ŷ = predicted value Y = expected value M = number of samples in the data set
  • 19. Types of Cost Function in Linear Regression Root mean squared error (RMSE): The Root Mean Squared Error (RMSE) is one of the two main performance indicators for a regression model. It measures the average difference between values predicted by a model and the actual values.
  • 20. Exploring the Cost Function We have a training set with three points (1, 1), (2, 2), and (3, 3). We plot the function f(x) = w * x for different values of w and calculate the corresponding cost function J(w). When w = 1: f(x) = x, the line passes through the origin, perfectly fitting the data. The cost function J(w) is 0 since f(x) equals y for all training examples. Setting w = 0.5: f(x) = 0.5 * x, the line has a smaller slope. The cost function J(w) now measures the squared errors between f(x) and y for each example. It provides a measure of how well the line fits the data. Function f(x) vs. Cost Function J(w)
  • 21. Real-life example of cost function Here, y denotes the target variable(Price of House), X denotes the size of the House and a0 is the intercept. The linear regression model will fit a straight line to our data, as shown in fig, and it will tell us what is the price of the House when we know the size of the House. Suppose this housing company sells houses having areas under 600 to 1500 square feet. So the prices will vary according to the size of the plot. For example House(1200 to 1500 square feet)= 65 lakhs. Now suddenly, there’s a rise in demand of Houses which means the price of this House will increase now. Suppose the price has now increased to 75 lakhs. But what if the model predicted its price as 30 lakhs? So this difference between these actual and predicted values is calculated by the cost function.
  • 22. Applications of Linear Regression ❏ Market analysis ❏ Financial analysis ❏ Sports analysis ❏ Environmental health ❏ Medicine
  • 23. Advantages of Linear Regression ❏ Linear Regression is simple to implement and easier to interpret the output coefficients ❏ When you know the relationship between the independent and dependent variable have a linear relationship, this algorithm is the best to use because of it’s less complexity compared to other algorithms. ❏ Linear Regression is susceptible to over-fitting but it can be avoided using some dimensionality reduction techniques, regularization (L1 and L2) techniques and cross-validation.
  • 24. Drawbacks of Linear Regression ❏ Linear Assumption ❏ Sensitivity to Exceptions ❏ Independence Presumption ❏ Limited to Linear Relationships
  • 25. Conclusion Linear regression is a foundational and widely used technique in machine learning, offering a straightforward yet effective way to model relationships between variables. It is a valuable tool for making predictions and understanding data patterns, though its suitability depends on the specific problem and the linearity assumption. While it may not always be the best choice, it remains an essential component of the machine learning toolkit, providing valuable insights in various domains.
  • 26. References ❖ https://www.geeksforgeeks.org/ml-advantages-and-disadvantages-of-linear- regression/ ❖ https://iq.opengenus.org/advantages-and-disadvantages-of-linear-regression/ ❖ https://www.intellspot.com/linear-regression-examples/ ❖ https://www.tutorialspoint.com/advantages-and-disadvantages-of-linear- regression#:~:text=The%20strategy%20is%20simple%20to,the%20subordinate%20and %20free%20factors. ❖ https://sphweb.bumc.bu.edu/otlt/MPH-Modules/BS/BS704- EP713_MultivariableMethods/ ❖ https://www.javatpoint.com/simple-linear-regression-in-machine-learning ❖ https://youtu.be/zUQr6HAAKp4?si=V-ojf6pXnMO2p7DR ❖ https://youtu.be/lzGKRSvs5HM?si=RduTOfWWjb05n5tF ❖ https://www.shiksha.com/online-courses/articles/cost-function-in-machine- learning/#Why-is-the-cost-function-needed? ❖ https://medium.com/@yennhi95zz/3-understanding-the-cost-function-in-linear-regression-for- machine-learning-beginners-ec9edeecbdde ❖ https://datatab.net/tutorial/linear-regression ❖ https://www.youtube.com/watch?v=QcPycBZomac