SlideShare a Scribd company logo
3
Most read
4
Most read
5
Most read
Application of Chebyshev and Markov Inequality in
Supervised Machine Learning
Domain: Application of Supervised Machine Learning
Dr. Varun Kumar
Domain: Application of Supervised Machine Learning Dr. Varun Kumar (IIIT Surat)Lecture 9 1 / 9
Outlines
1 Introduction to Chebyshev Inequality
2 Introduction to Markov Inequality
3 Introduction to Supervised Learning
4 Application of these Inequalities in Supervised Machine Learning
5 References
Domain: Application of Supervised Machine Learning Dr. Varun Kumar (IIIT Surat)Lecture 9 2 / 9
Introduction to Chebyshev Inequality
Mathematical Description:
General mathematics for continuous random variable:
⇒ Mean
E(x) = µ =
∞
−∞
xfX (x)dx (1)
⇒ Variance
E((x − µ)2
) = σ2
=
∞
−∞
(x − µ)2
fX (x)dx (2)
Domain: Application of Supervised Machine Learning Dr. Varun Kumar (IIIT Surat)Lecture 9 3 / 9
Chebyshev inequality
∞
−∞
(x − µ)2
fX (x)dx ≥
|x−µ|≥
(x − µ)2
fX (x)dx (3)
Taking the minimum value, i.e |x − µ| = → Finite deviation
|x−µ|≥
(x − µ)2
fX (x)dx =
|x−µ|≥
2
fX (x)dx = 2
P(X − µ ≥ ) (4)
From (2) and (4)
2
P(X − µ ≥ ) ≤ σ2
⇒ P(X − µ ≥ ) ≤
σ2
2
(5)
Case 1: when = nσ
P(X − µ ≥ ) = P(X − µ ≥ nσ) ≤
1
n2
(6)
Domain: Application of Supervised Machine Learning Dr. Varun Kumar (IIIT Surat)Lecture 9 4 / 9
Continued–
As per the properties of probability, P(X ≤ µ) + P(X ≥ µ) = 1. Hence,
P(X − µ ≤ ) ≤ 1 −
σ2
2
⇒ P(X − µ ≤ nσ) ≤ 1 −
1
n2
(7)
For discrete random variable:
Mean
E(x) = µ =
∞
i=−∞
xi PX (xi ) (8)
Variance
Var(x) = σ2
= E (x − µ)2
=
∞
i=−∞
(xi − E(x))2
PX (xi ) (9)
PX (.) → Probability mass function.
Domain: Application of Supervised Machine Learning Dr. Varun Kumar (IIIT Surat)Lecture 9 5 / 9
Markov inequality
P(X − µ ≤ ) ≤ 1 −
σ2
2
⇒ P(X − µ ≤ nσ) ≤ 1 −
1
n2
(10)
Markov inequality
Statement: If X is a positive random variable, i.e X > 0, having probability
density function fX (x). Let a is an positive arbitrary constant, then
P(X ≥ a) ≤
E(x)
a
(11)
Proof: As per the properties of random variable,
E(x) =
∞
0
xfX (x)dx ≥
∞
a
xfX (x)dx
Let x = a, then ⇒ E(x) ≥
∞
a
xfX (x)dx ≥
∞
a
afX (x)dx = aP(X > a)
Domain: Application of Supervised Machine Learning Dr. Varun Kumar (IIIT Surat)Lecture 9 6 / 9
Introduction to supervised learning
Supervised learning
1 It is a method of learning, where some set of predefined training data
is available.
2 Based on these training data or sequence, a mathematical or logical
model is developed.
3 This training data sequence or developed model through these data
acts as a supervisor.
4 When new data comes then it is expected that the data will follow
the developed model.
5 For developing a model through these training data, we may utilize
some well defined statistical, mathematical or logical model.
6 Those model gives a minimum mean square error value, that may be
selected as a most suitable model.
Domain: Application of Supervised Machine Learning Dr. Varun Kumar (IIIT Surat)Lecture 9 7 / 9
Relation between supervised learning and inequality
1. Decision action plays an important role in machine learning.
2. Inequality relation helps for making a decision favorable or
non-favorable region.
3. Statistical framework helps for modeling the synthetic data that is
nothing but the theoretical bound.
4. Applying Chebyshev inequality, there is requirement of variance of the
data sequence. It is independent from the type of distribution.
5. From relation (7) and (10), we can predict or find the probability of
any real world new data that is above or below some threshold value.
6. Applying Markov inequality, only mean value is required for finding
probability. It also independent from density function.
Domain: Application of Supervised Machine Learning Dr. Varun Kumar (IIIT Surat)Lecture 9 8 / 9
References
J. Navarro, “A very simple proof of the multivariate chebyshev’s inequality,”
Communications in Statistics-Theory and Methods, vol. 45, no. 12, pp. 3458–3463,
2016.
M. I. Jordan and T. M. Mitchell, “Machine learning: Trends, perspectives, and
prospects,” Science, vol. 349, no. 6245, pp. 255–260, 2015.
Domain: Application of Supervised Machine Learning Dr. Varun Kumar (IIIT Surat)Lecture 9 9 / 9

More Related Content

PDF
Chebyshev's inequality
PDF
Logistic regression
PPTX
Binomial probability distribution
PPTX
Normal as Approximation to Binomial
PPTX
Poisson Probability Distributions
PPTX
Discrete Distribution.pptx
PDF
Gmm chapter6
PPT
08 interpolation lagrange
Chebyshev's inequality
Logistic regression
Binomial probability distribution
Normal as Approximation to Binomial
Poisson Probability Distributions
Discrete Distribution.pptx
Gmm chapter6
08 interpolation lagrange

What's hot (20)

PPTX
Probability Distribution
PPTX
Sampling Distributions and Estimators
PDF
Blogging بلاگنگ by Rehmat Alam
PPTX
Geometric Distribution
PPT
Probability 4.2
PPTX
Complements and Conditional Probability, and Bayes' Theorem
PPTX
Support vector machine-SVM's
PPTX
Central limit theorem
PDF
Langrange Interpolation Polynomials
DOCX
PROBABILITY DISTRIBUTION
PDF
Chebyshev Inequality
PPTX
Hypergeometric probability distribution
PPTX
Random variables
PDF
Probability and Distribution
PPTX
Assessing Normality
PDF
Practice Test 2 Solutions
PDF
Intro to Classification: Logistic Regression & SVM
PDF
Spanos lecture 7: An Introduction to Bayesian Inference
PPTX
Negative binomial distribution
Probability Distribution
Sampling Distributions and Estimators
Blogging بلاگنگ by Rehmat Alam
Geometric Distribution
Probability 4.2
Complements and Conditional Probability, and Bayes' Theorem
Support vector machine-SVM's
Central limit theorem
Langrange Interpolation Polynomials
PROBABILITY DISTRIBUTION
Chebyshev Inequality
Hypergeometric probability distribution
Random variables
Probability and Distribution
Assessing Normality
Practice Test 2 Solutions
Intro to Classification: Logistic Regression & SVM
Spanos lecture 7: An Introduction to Bayesian Inference
Negative binomial distribution
Ad

Similar to Application of Chebyshev and Markov Inequality in Machine Learning (20)

PDF
Concentration inequality in Machine Learning
PDF
Bias and variance trade off
PPTX
Machine Learning basics
PPTX
Bayesian Neural Networks
PPT
Lecture 1
PPTX
Statistical foundations of ml
PDF
Lecture notes
PDF
Lecture 2 - Introduction to Machine Learning, a lecture in subject module Sta...
PPT
AML_030607.ppt
PDF
Lecture 3 (Supervised learning)
PDF
Pattern Recognition
PDF
Bayes Theorem.pdf
PDF
Some Take-Home Message about Machine Learning
PPT
Statistical Machine________ Learning.ppt
PPTX
PRML Chapter 1
PDF
MLHEP Lectures - day 2, basic track
PDF
lec3_annotated.pdf ml csci 567 vatsal sharan
PPTX
Bayesian Learning by Dr.C.R.Dhivyaa Kongu Engineering College
PPTX
UNIT II (7).pptx
PDF
Machine Learning Foundations
Concentration inequality in Machine Learning
Bias and variance trade off
Machine Learning basics
Bayesian Neural Networks
Lecture 1
Statistical foundations of ml
Lecture notes
Lecture 2 - Introduction to Machine Learning, a lecture in subject module Sta...
AML_030607.ppt
Lecture 3 (Supervised learning)
Pattern Recognition
Bayes Theorem.pdf
Some Take-Home Message about Machine Learning
Statistical Machine________ Learning.ppt
PRML Chapter 1
MLHEP Lectures - day 2, basic track
lec3_annotated.pdf ml csci 567 vatsal sharan
Bayesian Learning by Dr.C.R.Dhivyaa Kongu Engineering College
UNIT II (7).pptx
Machine Learning Foundations
Ad

More from VARUN KUMAR (20)

PDF
Distributed rc Model
PDF
Electrical Wire Model
PDF
Interconnect Parameter in Digital VLSI Design
PDF
Introduction to Digital VLSI Design
PDF
Challenges of Massive MIMO System
PDF
E-democracy or Digital Democracy
PDF
Ethics of Parasitic Computing
PDF
Action Lines of Geneva Plan of Action
PDF
Geneva Plan of Action
PDF
Fair Use in the Electronic Age
PDF
Software as a Property
PDF
Orthogonal Polynomial
PDF
Patent Protection
PDF
Copyright Vs Patent and Trade Secrecy Law
PDF
Property Right and Software
PDF
Investigating Data Trials
PDF
Gaussian Numerical Integration
PDF
Censorship and Controversy
PDF
Romberg's Integration
PDF
Introduction to Censorship
Distributed rc Model
Electrical Wire Model
Interconnect Parameter in Digital VLSI Design
Introduction to Digital VLSI Design
Challenges of Massive MIMO System
E-democracy or Digital Democracy
Ethics of Parasitic Computing
Action Lines of Geneva Plan of Action
Geneva Plan of Action
Fair Use in the Electronic Age
Software as a Property
Orthogonal Polynomial
Patent Protection
Copyright Vs Patent and Trade Secrecy Law
Property Right and Software
Investigating Data Trials
Gaussian Numerical Integration
Censorship and Controversy
Romberg's Integration
Introduction to Censorship

Recently uploaded (20)

PDF
BMEC211 - INTRODUCTION TO MECHATRONICS-1.pdf
PPT
Mechanical Engineering MATERIALS Selection
PDF
Mitigating Risks through Effective Management for Enhancing Organizational Pe...
PPTX
UNIT 4 Total Quality Management .pptx
PDF
Evaluating the Democratization of the Turkish Armed Forces from a Normative P...
PPTX
CH1 Production IntroductoryConcepts.pptx
PPTX
IOT PPTs Week 10 Lecture Material.pptx of NPTEL Smart Cities contd
PPT
Project quality management in manufacturing
PDF
July 2025 - Top 10 Read Articles in International Journal of Software Enginee...
PPTX
FINAL REVIEW FOR COPD DIANOSIS FOR PULMONARY DISEASE.pptx
PDF
Automation-in-Manufacturing-Chapter-Introduction.pdf
PDF
Operating System & Kernel Study Guide-1 - converted.pdf
PPTX
OOP with Java - Java Introduction (Basics)
PPTX
bas. eng. economics group 4 presentation 1.pptx
PPTX
CYBER-CRIMES AND SECURITY A guide to understanding
PDF
PPT on Performance Review to get promotions
PPT
CRASH COURSE IN ALTERNATIVE PLUMBING CLASS
PDF
The CXO Playbook 2025 – Future-Ready Strategies for C-Suite Leaders Cerebrai...
PPTX
MCN 401 KTU-2019-PPE KITS-MODULE 2.pptx
PDF
Well-logging-methods_new................
BMEC211 - INTRODUCTION TO MECHATRONICS-1.pdf
Mechanical Engineering MATERIALS Selection
Mitigating Risks through Effective Management for Enhancing Organizational Pe...
UNIT 4 Total Quality Management .pptx
Evaluating the Democratization of the Turkish Armed Forces from a Normative P...
CH1 Production IntroductoryConcepts.pptx
IOT PPTs Week 10 Lecture Material.pptx of NPTEL Smart Cities contd
Project quality management in manufacturing
July 2025 - Top 10 Read Articles in International Journal of Software Enginee...
FINAL REVIEW FOR COPD DIANOSIS FOR PULMONARY DISEASE.pptx
Automation-in-Manufacturing-Chapter-Introduction.pdf
Operating System & Kernel Study Guide-1 - converted.pdf
OOP with Java - Java Introduction (Basics)
bas. eng. economics group 4 presentation 1.pptx
CYBER-CRIMES AND SECURITY A guide to understanding
PPT on Performance Review to get promotions
CRASH COURSE IN ALTERNATIVE PLUMBING CLASS
The CXO Playbook 2025 – Future-Ready Strategies for C-Suite Leaders Cerebrai...
MCN 401 KTU-2019-PPE KITS-MODULE 2.pptx
Well-logging-methods_new................

Application of Chebyshev and Markov Inequality in Machine Learning

  • 1. Application of Chebyshev and Markov Inequality in Supervised Machine Learning Domain: Application of Supervised Machine Learning Dr. Varun Kumar Domain: Application of Supervised Machine Learning Dr. Varun Kumar (IIIT Surat)Lecture 9 1 / 9
  • 2. Outlines 1 Introduction to Chebyshev Inequality 2 Introduction to Markov Inequality 3 Introduction to Supervised Learning 4 Application of these Inequalities in Supervised Machine Learning 5 References Domain: Application of Supervised Machine Learning Dr. Varun Kumar (IIIT Surat)Lecture 9 2 / 9
  • 3. Introduction to Chebyshev Inequality Mathematical Description: General mathematics for continuous random variable: ⇒ Mean E(x) = µ = ∞ −∞ xfX (x)dx (1) ⇒ Variance E((x − µ)2 ) = σ2 = ∞ −∞ (x − µ)2 fX (x)dx (2) Domain: Application of Supervised Machine Learning Dr. Varun Kumar (IIIT Surat)Lecture 9 3 / 9
  • 4. Chebyshev inequality ∞ −∞ (x − µ)2 fX (x)dx ≥ |x−µ|≥ (x − µ)2 fX (x)dx (3) Taking the minimum value, i.e |x − µ| = → Finite deviation |x−µ|≥ (x − µ)2 fX (x)dx = |x−µ|≥ 2 fX (x)dx = 2 P(X − µ ≥ ) (4) From (2) and (4) 2 P(X − µ ≥ ) ≤ σ2 ⇒ P(X − µ ≥ ) ≤ σ2 2 (5) Case 1: when = nσ P(X − µ ≥ ) = P(X − µ ≥ nσ) ≤ 1 n2 (6) Domain: Application of Supervised Machine Learning Dr. Varun Kumar (IIIT Surat)Lecture 9 4 / 9
  • 5. Continued– As per the properties of probability, P(X ≤ µ) + P(X ≥ µ) = 1. Hence, P(X − µ ≤ ) ≤ 1 − σ2 2 ⇒ P(X − µ ≤ nσ) ≤ 1 − 1 n2 (7) For discrete random variable: Mean E(x) = µ = ∞ i=−∞ xi PX (xi ) (8) Variance Var(x) = σ2 = E (x − µ)2 = ∞ i=−∞ (xi − E(x))2 PX (xi ) (9) PX (.) → Probability mass function. Domain: Application of Supervised Machine Learning Dr. Varun Kumar (IIIT Surat)Lecture 9 5 / 9
  • 6. Markov inequality P(X − µ ≤ ) ≤ 1 − σ2 2 ⇒ P(X − µ ≤ nσ) ≤ 1 − 1 n2 (10) Markov inequality Statement: If X is a positive random variable, i.e X > 0, having probability density function fX (x). Let a is an positive arbitrary constant, then P(X ≥ a) ≤ E(x) a (11) Proof: As per the properties of random variable, E(x) = ∞ 0 xfX (x)dx ≥ ∞ a xfX (x)dx Let x = a, then ⇒ E(x) ≥ ∞ a xfX (x)dx ≥ ∞ a afX (x)dx = aP(X > a) Domain: Application of Supervised Machine Learning Dr. Varun Kumar (IIIT Surat)Lecture 9 6 / 9
  • 7. Introduction to supervised learning Supervised learning 1 It is a method of learning, where some set of predefined training data is available. 2 Based on these training data or sequence, a mathematical or logical model is developed. 3 This training data sequence or developed model through these data acts as a supervisor. 4 When new data comes then it is expected that the data will follow the developed model. 5 For developing a model through these training data, we may utilize some well defined statistical, mathematical or logical model. 6 Those model gives a minimum mean square error value, that may be selected as a most suitable model. Domain: Application of Supervised Machine Learning Dr. Varun Kumar (IIIT Surat)Lecture 9 7 / 9
  • 8. Relation between supervised learning and inequality 1. Decision action plays an important role in machine learning. 2. Inequality relation helps for making a decision favorable or non-favorable region. 3. Statistical framework helps for modeling the synthetic data that is nothing but the theoretical bound. 4. Applying Chebyshev inequality, there is requirement of variance of the data sequence. It is independent from the type of distribution. 5. From relation (7) and (10), we can predict or find the probability of any real world new data that is above or below some threshold value. 6. Applying Markov inequality, only mean value is required for finding probability. It also independent from density function. Domain: Application of Supervised Machine Learning Dr. Varun Kumar (IIIT Surat)Lecture 9 8 / 9
  • 9. References J. Navarro, “A very simple proof of the multivariate chebyshev’s inequality,” Communications in Statistics-Theory and Methods, vol. 45, no. 12, pp. 3458–3463, 2016. M. I. Jordan and T. M. Mitchell, “Machine learning: Trends, perspectives, and prospects,” Science, vol. 349, no. 6245, pp. 255–260, 2015. Domain: Application of Supervised Machine Learning Dr. Varun Kumar (IIIT Surat)Lecture 9 9 / 9