SlideShare a Scribd company logo
Nadar saraswathi college of arts &
science, theni.
Department of cs & it
ARTIFICIAL INTELLIGENCE
PRESENTED BY
G.KAVIYA
M.SC(IT)
TOPIC:LEARNING FROM
OBSERVATION
LEARNING
LEARNING FROM OBSERVATION
 Forms of learning
 Ensemble learning
 Computational learning theory
LEARNING:
Learning is Agent’s percepts should be used for
acting.
It also used for improving the agents ability to act
in the future.
Learning takes places as the agents observes, its
interactions with the world and its own decision making
processes.
FORMS OF LEARNING:
Learning Agent can be thought of as containing
a Performance Element, that decides, what actions to take,
and a Learning Elements that modifies the performance
elements to take better decisions.
Three major issues in learning element design
 Which components the performance element are to be
learned.
 What feedback is available to learn these components.
 What representation is used for the components.
Components of Agents are;
 A direct mapping from conditions on
the current state to actions.
 A means to infer relevant properties of
the world from the percept sequence.
 Information about the way the world
evolves and about the results of the
possible action the agent can take.
 Utility information indicating the
desirability of world states.
 Action value information indicating the
desirability of action.
 Goals the describe classes of the state
whose achievement maximizes the
agent’s utility.
Classified into three categories:
Supervised Learning.
Unsupervised Learning.
Reinforcement Learning.
Supervised Learning:
The Learning here is performed with the
help of teacher. Let us take the example of the learning
process of the small child.
The child doesn’t know how to read/write.
He/she is being taught by the parents at home and by the
teacher in school.
The children are recognize the alphabet,
numerals, etc. Their and every action is supervised by a
teacher.
Continue;
Actually, a child works on
the basis of the output that
he/she has to produce. All
these real-time events
involve supervised learning
methodology.
Similarly, in ANNs
following the supervised
learning, each input vector
requires a corresponding
target vector, which
represents the desired
outputs.
The input vector
along with the target vector
is called training pair.
Continue;
 In this type of training, a
supervisor or teacher is
required for error
minimization. Hence, the
network trained by this
method is said to be using
supervised training
methodology.
 In supervised learning, It is
assumed that the correct
“target” output values are
known for each input pattern.
 The input vector is
presented to the network,
Which result is an output
vector. The output vector is the
actual output vector. Then the
actual output vector is
compared with the desired
output vector.
 If there exists a difference
between the two output
vectors then an error signal is
generated by the network. This
error signal is used for
adjustment of weights until the
actual output matches the
desired output.
Un Supervised Learning:
 The learning here is
performed without the help of
teacher. Consider, learning
process of a tadpole, it learns
by itself, that is, a child fish
learns to swim by itself, it is not
taught by its mother.
 Thus, Its learning process
is independent and is not
supervised by a teacher.
 In ANNs following
unsupervised learning, the
input vectors of similar type
are grouped without the use of.
Reinforcement Learning:
 The learning process is
similar to supervised learning.
In the case of supervised
learning the correct target
output values are known for
each input pattern.
 But, In some cases, Less
information might be available.
For example, the network
might be told that its actual
output is only “50% correct “or
so. Thus, Here only critic
information is available, not the
exact information.
 The learning information
based on the critic information
is called reinforcement learning
and the feedback sent is called
reinforcement signal.
ENSEMBLE OF LEARNING:
 Learn multiple
alternative definitions of
a concept using different
training data or different
learning algorithms.
 Combine decisions
of multiple definitions,
eg. Using weighted
voting.
VALUE OF ENSEMBLES
When combing multiple independent and diverse
decisions each of which is at least more accurate than random
guessing, random errors cancel each other out, correct decisions
are reinforcement.
Generate a group of base-learners which when
combined has higher accuracy.
Different learners use different;
Algorithm.
Hyperparameters.
Representations/Modalities/Views.
Training sets.
Subproblems.
Ensembles:
BOOSTING:
Also uses voting/averaging but models are
weighted according to their performance.
Iterative procedure: new models are influenced
by performance of previously built ones.
* New model is encouraged to become
expert for instances classified incorrectly by earlier
models.
* Intuitive justification: models should be
experts that complement each other.
There are several variants of this algorithm.
Continue;
 STRONG LEARNER:
Objective of machine learning.
o Take labeled data for training.
o Produce a classifier which can be
arbitrarily accurate.
o Strong learners are very difficult to
construct.
 WEAKER LEARNER:
o Take labeled data for training.
o Produce a classifier which is more
accurate than random guessing.
o Constructing weaker learners is
relatively easy.
ADAPTIVE BOOSTING:
Each rectangle corresponds to an example, with
weight proportional to its height.
Crosses corresponds to misclassified examples.
Size of decision tree indicates the weight of that
classifier in the final ensemble.
Using Different Data Distribution
* Start with uniform weighting.
* During each step of learning.
Increase weights of the examples which are not
correctly learned by the weak learner.
Decrease weights of the examples which are
correctly learned by the weak learner.
Continue;
IDEA:
focus on difficult example which
are not correctly classified in the
previous steps
WEIGHTED VOTING:
construct strong classifier by
weighted voting of the weak classifier.
IDEA:
Better weak classifier gets a
larger weight.
Iteratively add weak classifiers
Increase accuracy of the
combined classifier through
minimization of a cost function.
COMPUTATIONAL LEARNING
THEORY
 Computational learning theory characterize. The difficulty of
several types of machine learning problem.
 Capabilities of several types of ML algorithm.
 CLT seeks answer, question such as;
a) “under what conditions is successful learning possible and
impossible?”
b) “under what conditions is a particular learning algorithm
assured of learning successfully?” it means that, what kind of
task are learnable, what kind of data is required for
learnability.
Various issues are:
Sample complexity:
How many training examples are needed for a
learner to converge (with high probability) to a successful
hypothesis?
Computational complexity:
How much computational effort is needed for a
learner to converge to a successful hypothesis?
Mistake bound:
How many training examples will the learner
misclassify before converging to a successful hypothesis?
(PAC) Probably Learning an Approximately
Correct hypothesis:-
A particular setting for the learning problem,
called the probably approximately correct(PAC) learning
model.
This model of learning is based on following
points:
1. Specifying problem setting that defines PAC
model.
2. How many training examples are required.
3. How much computational are required in order
to learn various classes of target functions within PAC
Problem setting:-
X : Set of all the instance (eg: set of people)
each described by attributes <age, height>.
C : Target concept the leaner need to learn.
C: X{0,1}
L : Learner have to learn “people who are
skiers”
C(x) =1 : positive training example
C(x) =0 : negative training example
Error of a hypothesis: True error, denoted by errorD
(h), of hypothesis h w.r.t target concept c and
distribution D is probability that h will misclassify an
instance drawn at random according to D.
errorD (h) = pr [c(x) not equal to h(x)]
XED
PAC Learnability:
No of training examples needed to learn a
hypothesis h, for which
errorD (h) = 0.
For these two difficulties, following measures can be
taken:-
1. No requirement of zero error hypothesis for learner L. So
a bound to error can be set by constant E, that can be made small.
2.Not necessary that learner succeed for every sequence of
randomly drawn training example. So learner probably learn a
hypothesis that is approximately correct.
Bounded by same constant S which is,
Definition:
Consider a concept class define over a set of instance
X of kngtn n and a learner L using hypothesis space H.
C is PAC learnable by L using H if for all CEC.
distribution D over X
E such that 0<E<1/2 and
S such that 0<S<1/2
Learner L will with Probably at least (1-8)
output a hypothesis nEH such that
errorD (n) <= E.
Artificial Intelligence.pptx

More Related Content

PPT
learning.ppt
PPTX
Learning
PPT
ML_Overview.ppt
PPTX
ML_Overview.pptx
PPT
ML_Overview.ppt
PPT
ML overview
PDF
Buku panduan untuk Machine Learning.pdf
PPT
Different learning Techniques in Artificial Intelligence
learning.ppt
Learning
ML_Overview.ppt
ML_Overview.pptx
ML_Overview.ppt
ML overview
Buku panduan untuk Machine Learning.pdf
Different learning Techniques in Artificial Intelligence

Similar to Artificial Intelligence.pptx (20)

PPTX
Rahul_Kirtoniya_11800121032_CSE_Machine_Learning.pptx
PPTX
Statistical foundations of ml
PPTX
ML_ Unit_1_PART_A
PPTX
Chapter 6 - Learning data and analytics course
PDF
Endsem AI merged.pdf
PPT
nnml.ppt
PPTX
machine leraning : main principles and techniques
PPTX
Learning in AI
PPT
Machine Learning and Artificial Neural Networks.ppt
PPT
Machine learning
PPTX
UNIT 1 Machine Learning [KCS-055] (1).pptx
PPT
CS8082_MachineLearnigTechniques _Unit-1.ppt
PPT
Introduction to Machine Learning.
PPTX
Learning occam razor
PPTX
3171617_introduction_applied machine learning.pptx
PPTX
supervised_learning_PRESENTATION___.pptx
PPTX
Introduction to Machine Learning
PDF
Machine Learning
PDF
Unit-1.pdf
Rahul_Kirtoniya_11800121032_CSE_Machine_Learning.pptx
Statistical foundations of ml
ML_ Unit_1_PART_A
Chapter 6 - Learning data and analytics course
Endsem AI merged.pdf
nnml.ppt
machine leraning : main principles and techniques
Learning in AI
Machine Learning and Artificial Neural Networks.ppt
Machine learning
UNIT 1 Machine Learning [KCS-055] (1).pptx
CS8082_MachineLearnigTechniques _Unit-1.ppt
Introduction to Machine Learning.
Learning occam razor
3171617_introduction_applied machine learning.pptx
supervised_learning_PRESENTATION___.pptx
Introduction to Machine Learning
Machine Learning
Unit-1.pdf
Ad

More from Kaviya452563 (14)

PPTX
softcomputing.pptx
PPTX
OOAD.pptx
PPTX
DIP.pptx
PPTX
Big Data Analytics.pptx
PPTX
client server computing.pptx
PPTX
WE.pptx
PPTX
Internet of Things.pptx
PPTX
data mining.pptx
PPTX
python programming.pptx
PPTX
Distributing computing.pptx
PPTX
Advanced java programming
PPTX
Network and internet security
PPTX
Advanced computer architecture
PPTX
Data structures and algorithms
softcomputing.pptx
OOAD.pptx
DIP.pptx
Big Data Analytics.pptx
client server computing.pptx
WE.pptx
Internet of Things.pptx
data mining.pptx
python programming.pptx
Distributing computing.pptx
Advanced java programming
Network and internet security
Advanced computer architecture
Data structures and algorithms
Ad

Recently uploaded (20)

PDF
FourierSeries-QuestionsWithAnswers(Part-A).pdf
PPTX
Microbial diseases, their pathogenesis and prophylaxis
PPTX
master seminar digital applications in india
PDF
The Lost Whites of Pakistan by Jahanzaib Mughal.pdf
PDF
BÀI TẬP BỔ TRỢ 4 KỸ NĂNG TIẾNG ANH 9 GLOBAL SUCCESS - CẢ NĂM - BÁM SÁT FORM Đ...
PDF
VCE English Exam - Section C Student Revision Booklet
PDF
Classroom Observation Tools for Teachers
PDF
TR - Agricultural Crops Production NC III.pdf
PDF
Mark Klimek Lecture Notes_240423 revision books _173037.pdf
PDF
Complications of Minimal Access Surgery at WLH
PPTX
Cell Types and Its function , kingdom of life
PPTX
PPH.pptx obstetrics and gynecology in nursing
PPTX
Renaissance Architecture: A Journey from Faith to Humanism
PDF
Pre independence Education in Inndia.pdf
PDF
O7-L3 Supply Chain Operations - ICLT Program
PPTX
Pharma ospi slides which help in ospi learning
PDF
Business Ethics Teaching Materials for college
PPTX
school management -TNTEU- B.Ed., Semester II Unit 1.pptx
PPTX
PPT- ENG7_QUARTER1_LESSON1_WEEK1. IMAGERY -DESCRIPTIONS pptx.pptx
PPTX
Pharmacology of Heart Failure /Pharmacotherapy of CHF
FourierSeries-QuestionsWithAnswers(Part-A).pdf
Microbial diseases, their pathogenesis and prophylaxis
master seminar digital applications in india
The Lost Whites of Pakistan by Jahanzaib Mughal.pdf
BÀI TẬP BỔ TRỢ 4 KỸ NĂNG TIẾNG ANH 9 GLOBAL SUCCESS - CẢ NĂM - BÁM SÁT FORM Đ...
VCE English Exam - Section C Student Revision Booklet
Classroom Observation Tools for Teachers
TR - Agricultural Crops Production NC III.pdf
Mark Klimek Lecture Notes_240423 revision books _173037.pdf
Complications of Minimal Access Surgery at WLH
Cell Types and Its function , kingdom of life
PPH.pptx obstetrics and gynecology in nursing
Renaissance Architecture: A Journey from Faith to Humanism
Pre independence Education in Inndia.pdf
O7-L3 Supply Chain Operations - ICLT Program
Pharma ospi slides which help in ospi learning
Business Ethics Teaching Materials for college
school management -TNTEU- B.Ed., Semester II Unit 1.pptx
PPT- ENG7_QUARTER1_LESSON1_WEEK1. IMAGERY -DESCRIPTIONS pptx.pptx
Pharmacology of Heart Failure /Pharmacotherapy of CHF

Artificial Intelligence.pptx

  • 1. Nadar saraswathi college of arts & science, theni. Department of cs & it ARTIFICIAL INTELLIGENCE PRESENTED BY G.KAVIYA M.SC(IT) TOPIC:LEARNING FROM OBSERVATION
  • 2. LEARNING LEARNING FROM OBSERVATION  Forms of learning  Ensemble learning  Computational learning theory
  • 3. LEARNING: Learning is Agent’s percepts should be used for acting. It also used for improving the agents ability to act in the future. Learning takes places as the agents observes, its interactions with the world and its own decision making processes.
  • 4. FORMS OF LEARNING: Learning Agent can be thought of as containing a Performance Element, that decides, what actions to take, and a Learning Elements that modifies the performance elements to take better decisions. Three major issues in learning element design  Which components the performance element are to be learned.  What feedback is available to learn these components.  What representation is used for the components.
  • 5. Components of Agents are;  A direct mapping from conditions on the current state to actions.  A means to infer relevant properties of the world from the percept sequence.  Information about the way the world evolves and about the results of the possible action the agent can take.  Utility information indicating the desirability of world states.  Action value information indicating the desirability of action.  Goals the describe classes of the state whose achievement maximizes the agent’s utility.
  • 6. Classified into three categories: Supervised Learning. Unsupervised Learning. Reinforcement Learning.
  • 7. Supervised Learning: The Learning here is performed with the help of teacher. Let us take the example of the learning process of the small child. The child doesn’t know how to read/write. He/she is being taught by the parents at home and by the teacher in school. The children are recognize the alphabet, numerals, etc. Their and every action is supervised by a teacher.
  • 8. Continue; Actually, a child works on the basis of the output that he/she has to produce. All these real-time events involve supervised learning methodology. Similarly, in ANNs following the supervised learning, each input vector requires a corresponding target vector, which represents the desired outputs. The input vector along with the target vector is called training pair.
  • 9. Continue;  In this type of training, a supervisor or teacher is required for error minimization. Hence, the network trained by this method is said to be using supervised training methodology.  In supervised learning, It is assumed that the correct “target” output values are known for each input pattern.  The input vector is presented to the network, Which result is an output vector. The output vector is the actual output vector. Then the actual output vector is compared with the desired output vector.  If there exists a difference between the two output vectors then an error signal is generated by the network. This error signal is used for adjustment of weights until the actual output matches the desired output.
  • 10. Un Supervised Learning:  The learning here is performed without the help of teacher. Consider, learning process of a tadpole, it learns by itself, that is, a child fish learns to swim by itself, it is not taught by its mother.  Thus, Its learning process is independent and is not supervised by a teacher.  In ANNs following unsupervised learning, the input vectors of similar type are grouped without the use of.
  • 11. Reinforcement Learning:  The learning process is similar to supervised learning. In the case of supervised learning the correct target output values are known for each input pattern.  But, In some cases, Less information might be available. For example, the network might be told that its actual output is only “50% correct “or so. Thus, Here only critic information is available, not the exact information.  The learning information based on the critic information is called reinforcement learning and the feedback sent is called reinforcement signal.
  • 12. ENSEMBLE OF LEARNING:  Learn multiple alternative definitions of a concept using different training data or different learning algorithms.  Combine decisions of multiple definitions, eg. Using weighted voting.
  • 13. VALUE OF ENSEMBLES When combing multiple independent and diverse decisions each of which is at least more accurate than random guessing, random errors cancel each other out, correct decisions are reinforcement. Generate a group of base-learners which when combined has higher accuracy. Different learners use different; Algorithm. Hyperparameters. Representations/Modalities/Views. Training sets. Subproblems.
  • 15. BOOSTING: Also uses voting/averaging but models are weighted according to their performance. Iterative procedure: new models are influenced by performance of previously built ones. * New model is encouraged to become expert for instances classified incorrectly by earlier models. * Intuitive justification: models should be experts that complement each other. There are several variants of this algorithm.
  • 16. Continue;  STRONG LEARNER: Objective of machine learning. o Take labeled data for training. o Produce a classifier which can be arbitrarily accurate. o Strong learners are very difficult to construct.  WEAKER LEARNER: o Take labeled data for training. o Produce a classifier which is more accurate than random guessing. o Constructing weaker learners is relatively easy.
  • 17. ADAPTIVE BOOSTING: Each rectangle corresponds to an example, with weight proportional to its height. Crosses corresponds to misclassified examples. Size of decision tree indicates the weight of that classifier in the final ensemble. Using Different Data Distribution * Start with uniform weighting. * During each step of learning. Increase weights of the examples which are not correctly learned by the weak learner. Decrease weights of the examples which are correctly learned by the weak learner.
  • 18. Continue; IDEA: focus on difficult example which are not correctly classified in the previous steps WEIGHTED VOTING: construct strong classifier by weighted voting of the weak classifier. IDEA: Better weak classifier gets a larger weight. Iteratively add weak classifiers Increase accuracy of the combined classifier through minimization of a cost function.
  • 19. COMPUTATIONAL LEARNING THEORY  Computational learning theory characterize. The difficulty of several types of machine learning problem.  Capabilities of several types of ML algorithm.  CLT seeks answer, question such as; a) “under what conditions is successful learning possible and impossible?” b) “under what conditions is a particular learning algorithm assured of learning successfully?” it means that, what kind of task are learnable, what kind of data is required for learnability.
  • 20. Various issues are: Sample complexity: How many training examples are needed for a learner to converge (with high probability) to a successful hypothesis? Computational complexity: How much computational effort is needed for a learner to converge to a successful hypothesis? Mistake bound: How many training examples will the learner misclassify before converging to a successful hypothesis?
  • 21. (PAC) Probably Learning an Approximately Correct hypothesis:- A particular setting for the learning problem, called the probably approximately correct(PAC) learning model. This model of learning is based on following points: 1. Specifying problem setting that defines PAC model. 2. How many training examples are required. 3. How much computational are required in order to learn various classes of target functions within PAC
  • 22. Problem setting:- X : Set of all the instance (eg: set of people) each described by attributes <age, height>. C : Target concept the leaner need to learn. C: X{0,1} L : Learner have to learn “people who are skiers” C(x) =1 : positive training example C(x) =0 : negative training example Error of a hypothesis: True error, denoted by errorD (h), of hypothesis h w.r.t target concept c and distribution D is probability that h will misclassify an instance drawn at random according to D. errorD (h) = pr [c(x) not equal to h(x)] XED
  • 23. PAC Learnability: No of training examples needed to learn a hypothesis h, for which errorD (h) = 0. For these two difficulties, following measures can be taken:- 1. No requirement of zero error hypothesis for learner L. So a bound to error can be set by constant E, that can be made small. 2.Not necessary that learner succeed for every sequence of randomly drawn training example. So learner probably learn a hypothesis that is approximately correct. Bounded by same constant S which is,
  • 24. Definition: Consider a concept class define over a set of instance X of kngtn n and a learner L using hypothesis space H. C is PAC learnable by L using H if for all CEC. distribution D over X E such that 0<E<1/2 and S such that 0<S<1/2 Learner L will with Probably at least (1-8) output a hypothesis nEH such that errorD (n) <= E.