SlideShare a Scribd company logo
UNIT 3
AI
BY:SURBHI SAROHA
1SURBHI SAROHA
SYLLABUS
 Propositional logic
 Theory of first order logic
 Inference in First order logic
 Forward & Backward chaining
 Resolution
 Probabilistic reasoning
 Utility theory
 Hidden Markov Models(HMM)
 Bayesian Networks.
2SURBHI SAROHA
Propositional logic
 This logic is synonymous with sentential logic,
provides ways and means of combining or changing
the propositions to create a complicated structure and
to build a new logical relationship and properties.
 It combines logical connections of all the constituent
statements and the true value of the complex
statement is derived.
 While doing, it considers other factors like reasoning,
relationship, and interconnection between the
constituent statements.
3SURBHI SAROHA
Cont….
 This logic provides better clarity on data and
information in an incomplete environment by deeper
analysis and inference of the limited information
presented to it.
 Due to this unique feature and powerful algorithms
deployed in logical reasoning, it is widely adopted in
advanced search in AI in solving complex problems.
4SURBHI SAROHA
What is propositional logic?
 The natural language words may have slightly different
meanings.
 Example: A ∧ B and B ∧ A should always have the same
meaning.
 But the sentences
 She became sick and she went to the doctor.
 and
 She went to the doctor and she became sick. have
different meanings.
5SURBHI SAROHA
Theory of first order logic
 In the topic of Propositional logic, we have seen that how to
represent statements using propositional logic.
 But unfortunately, in propositional logic, we can only represent
the facts, which are either true or false.
 PL is not sufficient to represent the complex sentences or natural
language statements.
 The propositional logic has very limited expressive power.
Consider the following sentence, which we cannot represent
using PL logic.
 "Some humans are intelligent", or
 "Sachin likes cricket."
 To represent the above statements, PL logic is not sufficient, so
we required some more powerful logic, such as first-order logic.
6SURBHI SAROHA
First-Order logic:
 First-order logic is another way of knowledge
representation in artificial intelligence. It is an extension to
propositional logic.
 FOL is sufficiently expressive to represent the natural
language statements in a concise way.
 First-order logic is also known as Predicate logic or First-
order predicate logic. First-order logic is a powerful
language that develops information about the objects in a
more easy way and can also express the relationship
between those objects.
 First-order logic (like natural language) does not only
assume that the world contains facts like propositional
logic but also assumes the following things in the world:
7SURBHI SAROHA
Cont….
 Objects: A, B, people, numbers, colors, wars, theories,
squares, pits, …
 Relations: It can be unary relation such as: red, round,
is adjacent, or n-any relation such as: the sister of,
brother of, has color, comes between
 Function: Father of, best friend, third inning of, end of,
......
 As a natural language, first-order logic also has two main
parts:
 Syntax
 Semantics
8SURBHI SAROHA
Inference in First order logic
 Inference in First-Order Logic is used to deduce new facts
or sentences from existing sentences.
 Before understanding the FOL inference rule, let's
understand some basic terminologies used in FOL.
 Substitution:
 Substitution is a fundamental operation performed on
terms and formulas.
 It occurs in all inference systems in first-order logic.
 The substitution is complex in the presence of quantifiers
in FOL.
 If we write F[a/x], so it refers to substitute a constant "a" in
place of variable "x".
9SURBHI SAROHA
Cont….
 Equality:
 First-Order logic does not only use predicate and terms for
making atomic sentences but also uses another way, which
is equality in FOL. For this, we can use equality
symbols which specify that the two terms refer to the same
object.
 Example: Brother (John) = Smith.
 As in the above example, the object referred by
the Brother (John) is similar to the object referred
by Smith. The equality symbol can also be used with
negation to represent that two terms are not the same
objects.
 Example: ¬(x=y) which is equivalent to x ≠y.
10SURBHI SAROHA
FOL inference rules for quantifier:
 As propositional logic we also have inference rules in
first-order logic, so following are some basic inference
rules in FOL:
 Universal Generalization
 Universal Instantiation
 Existential Instantiation
 Existential introduction
11SURBHI SAROHA
1. Universal Generalization
 Universal generalization is a valid inference rule which
states that if premise P(c) is true for any arbitrary element c
in the universe of discourse, then we can have a conclusion
as ∀ x P(x).
 It can be represented as: .
 This rule can be used if we want to show that every element
has a similar property.
 In this rule, x must not appear as a free variable.
 Example: Let's represent, P(c): "A byte contains 8 bits",
so for ∀ x P(x) "All bytes contain 8 bits.", it will also be
true.
12SURBHI SAROHA
2.Universal Instantiation:
 Universal instantiation is also called as universal elimination or
UI is a valid inference rule. It can be applied multiple times to
add new sentences.
 As per UI, we can infer any sentence obtained by
substituting a ground term for the variable.
 The UI rule state that we can infer any sentence P(c) by
substituting a ground term c (a constant within domain x)
from ∀ x P(x) for any object in the universe of discourse.
 It can be represented as:.
 Example:1.
 IF "Every person like ice-cream"=> ∀x P(x) so we can infer that
"John likes ice-cream" => P(c)
13SURBHI SAROHA
Cont…
 Example: 2.
 Let's take a famous example,
 "All kings who are greedy are Evil." So let our knowledge
base contains this detail as in the form of FOL:
 ∀x king(x) ∧ greedy (x) → Evil (x),
 So from this information, we can infer any of the following
statements using Universal Instantiation:
 King(John) ∧ Greedy (John) → Evil (John),
 King(Richard) ∧ Greedy (Richard) → Evil (Richard),
 King(Father(John)) ∧ Greedy (Father(John)) → Evil
(Father(John)),
14SURBHI SAROHA
3.Existential Instantiation:
 Existential instantiation is also called as Existential
Elimination, which is a valid inference rule in first-order
logic.
 It can be applied only once to replace the existential
sentence.
 The new KB is not logically equivalent to old KB, but it will
be satisfiable if old KB was satisfiable.
 This rule states that one can infer P(c) from the formula
given in the form of ∃x P(x) for a new constant symbol c.
 The restriction with this rule is that c used in the rule must
be a new term for which P(c ) is true.
 It can be represented as:
15SURBHI SAROHA
Cont….
 Example:
 From the given sentence: ∃x Crown(x) ∧ OnHead(x,
John),
 So we can infer: Crown(K) ∧ OnHead( K, John), as
long as K does not appear in the knowledge base.
 The above used K is a constant symbol, which is
called Skolem constant.
 The Existential instantiation is a special case
of Skolemization process.
16SURBHI SAROHA
4.Existential introduction
 An existential introduction is also known as an
existential generalization, which is a valid inference
rule in first-order logic.
 This rule states that if there is some element c in the
universe of discourse which has a property P, then we
can infer that there exists something in the universe
which has the property P.
 It can be represented as:
 Example: Let's say that,
"Priyanka got good marks in English."
"Therefore, someone got good marks in English."
17SURBHI SAROHA
Forward & Backward chaining
 Forward chaining as the name suggests, start from
the known facts and move forward by applying
inference rules to extract more data, and it continues
until it reaches to the goal, whereas backward
chaining starts from the goal, move backward by
using inference rules to determine the facts that satisfy
the goal.
18SURBHI SAROHA
Forward Chaining
 Forward chaining is also known as a forward
deduction or forward reasoning method when using
an inference engine.
 Forward chaining is a form of reasoning which start
with atomic sentences in the knowledge base and
applies inference rules (Modus Ponens) in the forward
direction to extract more data until a goal is reached.
 The Forward-chaining algorithm starts from known
facts, triggers all rules whose premises are satisfied,
and add their conclusion to the known facts. This
process repeats until the problem is solved.
19SURBHI SAROHA
Properties of Forward-Chaining:
 It is a down-up approach, as it moves from bottom to
top.
 It is a process of making a conclusion based on known
facts or data, by starting from the initial state and
reaches the goal state.
 Forward-chaining approach is also called as data-
driven as we reach to the goal using available data.
 Forward -chaining approach is commonly used in the
expert system, such as CLIPS, business, and
production rule systems.
20SURBHI SAROHA
Example:
 "As per the law, it is a crime for an American to sell
weapons to hostile nations. Country A, an enemy
of America, has some missiles, and all the
missiles were sold to it by Robert, who is an
American citizen."
 Prove that "Robert is criminal."
 To solve the above problem, first, we will convert all
the above facts into first-order definite clauses, and
then we will use a forward-chaining algorithm to reach
the goal.
21SURBHI SAROHA
Facts Conversion into FOL:
 It is a crime for an American to sell weapons to hostile nations. (Let's
say p, q, and r are variables)
American (p) ∧ weapon(q) ∧ sells (p, q, r) ∧ hostile(r) →
Criminal(p) ...(1)
 Country A has some missiles. ?p Owns(A, p) ∧ Missile(p). It can be
written in two definite clauses by using Existential Instantiation,
introducing new Constant T1.
Owns(A, T1) ......(2)
Missile(T1) .......(3)
 All of the missiles were sold to country A by Robert.
?p Missiles(p) ∧ Owns (A, p) → Sells (Robert, p, A) ......(4)
 Missiles are weapons.
Missile(p) → Weapons (p) .......(5)
 Enemy of America is known as hostile.
Enemy(p, America) →Hostile(p) ........(6)
22SURBHI SAROHA
Cont….
 Country A is an enemy of America.
Enemy (A, America) .........(7)
 Robert is American
American(Robert). ..........(8)
23SURBHI SAROHA
Forward chaining proof:
 Step-1:
 In the first step we will start with the known facts and
will choose the sentences which do not have
implications, such as:
 American(Robert), Enemy(A, America), Owns(A,
T1), and Missile(T1). All these facts will be
represented as below.
24SURBHI SAROHA
Cont…
 Step-2:
 At the second step, we will see those facts which infer from
available facts and with satisfied premises.
 Rule-(1) does not satisfy premises, so it will not be added in
the first iteration.
 Rule-(2) and (3) are already added.
 Rule-(4) satisfy with the substitution {p/T1}, so Sells
(Robert, T1, A) is added, which infers from the
conjunction of Rule (2) and (3).
 Rule-(6) is satisfied with the substitution(p/A), so
Hostile(A) is added and which infers from Rule-(7).
25SURBHI SAROHA
26SURBHI SAROHA
Cont….
 Step-3:
 At step-3, as we can check Rule-(1) is satisfied with the
substitution {p/Robert, q/T1, r/A}, so we can add
Criminal(Robert) which infers all the available facts.
And hence we reached our goal statement.
27SURBHI SAROHA
28SURBHI SAROHA
Cont….
 Hence it is proved that Robert is Criminal using
forward chaining approach.
29SURBHI SAROHA
Backward chaining
 A backward chaining algorithm is a form of
reasoning, which starts with the goal and
works backward, chaining through rules to find
known facts that support the goal.
 Properties of backward chaining: It is known as a
top-down approach.
 Backward-chaining is based on modus ponens
inference rule.
30SURBHI SAROHA
Cont…
 The Forward-thinking approach is used in AI to help an AI
agent solve logical problems by inspecting the data from
the previous learnings and then coming to a conclusion full
of solutions.
 That’s not all, Forward Chaining might as well be used to
explore the available information or answer a question or
solve a problem.
 Forward chaining is extensively used to break down a long
and complex logical approach by attaching each step once
the previous one is completed.
 This way, it goes from beginning to the end with relative
ease.
31SURBHI SAROHA
Steps for working of Forwarding
Chaining
 Step 1: We start from the already stated facts, and then,
we’ll subsequently choose the facts that do not have
any implications at all.
 Step 2: Now, we will state those facts that can be
inferred from available facts with satisfied premises.
 Step 3: In step 3 we can check the given statement that
needs to be checked and check whether it is satisfied
with the substitution which infers all the previously
stated facts.
 Thus we reach our goal.
32SURBHI SAROHA
Resolution
 Resolution is a theorem proving technique that proceeds
by building refutation proofs, i.e., proofs by contradictions.
 It was invented by a Mathematician John Alan Robinson in
the year 1965.
 Resolution is used, if there are various statements are
given, and we need to prove a conclusion of those
statements.
 Unification is a key concept in proofs by resolutions.
 Resolution is a single inference rule which can efficiently
operate on the conjunctive normal form or clausal
form.
 Clause: Disjunction of literals (an atomic sentence) is
called a clause. It is also known as a unit clause.
33SURBHI SAROHA
Cont…
 Conjunctive Normal Form: A sentence represented as a
conjunction of clauses is said to be conjunctive normal
form or CNF.
 Steps for Resolution:
 Conversion of facts into first-order logic.
 Convert FOL statements into CNF
 Negate the statement which needs to prove (proof by
contradiction)
 Draw resolution graph (unification).
 To better understand all the above steps, we will take an
example in which we will apply resolution.
34SURBHI SAROHA
Cont….
 Example:
 John likes all kind of food.
 Apple and vegetable are food
 Anything anyone eats and not killed is food.
 Anil eats peanuts and still alive
 Harry eats everything that Anil eats.
Prove by resolution that:
 John likes peanuts.
35SURBHI SAROHA
Probabilistic reasoning
 Probabilistic reasoning is a way of knowledge
representation where we apply the concept
of probability to indicate the uncertainty in
knowledge.
 In probabilistic reasoning,
 we combine probability theory with logic to handle
the uncertainty.
36SURBHI SAROHA
Cont….
 Probabilistic reasoning is a method of representation
of knowledge where the concept of probability is
applied to indicate the uncertainty in knowledge.
 Probabilistic reasoning is used in AI:
 When we are unsure of the predicates
 When the possibilities of predicates become too large
to list down
 When it is known that an error occurs during an
experiment
37SURBHI SAROHA
Bayesian Networks
 Bayesian network is a directed acyclic graph model which
helps us represent probabilistic data.
 A Bayesian network contains two basic components:
 Nodes: These represent attributes/functions or data.
 Arcs: These represent the dependencies between the
nodes. If an arc is present between the two nodes, then
those two nodes have some sort of relation between them.
 An absence of an arc denotes the absence of any relation
between the nodes.
 When designing a Bayesian network, we keep the local
probability table at each node.
38SURBHI SAROHA
Utility theory
 Utility Theory is the discipline that lays out the
foundation to create and evaluate Utility Functions.
 Typically, Utility Theory uses the notion of
Expected Utility (EU) as a value that represents the
average utility of all possible outcomes of a state,
weighted by the probability that the outcome occurs.
 The agents use the utility theory for making decisions.
It is the mapping from lotteries to the real numbers.
 An agent is supposed to have various preferences and
can choose the one which best fits his necessity.
39SURBHI SAROHA
Hidden Markov Models(HMM)
 Hidden Markov Model is a statistical Markov model in which the
system being modeled is assumed to be a Markov process – call it
– with unobservable states.
 HMM assumes that there is another process whose behavior
"depends" on.
 The goal is to learn about by observing.
 Hidden Markov Model (HMM) is a statistical Markov model in
which the system being modeled is assumed to be a Markov
process – call it {displaystyle X} – with unobservable ("hidden")
states.
 HMM assumes that there is another process {displaystyle
Y} whose behavior "depends" on {displaystyle X}.
 The goal is to learn about {displaystyle X} by
observing {displaystyle Y}
40SURBHI SAROHA
Thank you 
41SURBHI SAROHA

More Related Content

PPTX
Propositional logic
PPT
Propositional And First-Order Logic
PPTX
Artificial Intelligence (AI) | Prepositional logic (PL)and first order predic...
PPTX
Knowledge Engineering in FOL.
PPTX
Unification and Lifting
PDF
AI PPT-ALR_Unit-3-1.pdf
PPTX
Frames
PPT
Predicate logic_2(Artificial Intelligence)
Propositional logic
Propositional And First-Order Logic
Artificial Intelligence (AI) | Prepositional logic (PL)and first order predic...
Knowledge Engineering in FOL.
Unification and Lifting
AI PPT-ALR_Unit-3-1.pdf
Frames
Predicate logic_2(Artificial Intelligence)

What's hot (20)

PPTX
Knowledge representation in AI
PDF
I.INFORMED SEARCH IN ARTIFICIAL INTELLIGENCE II. HEURISTIC FUNCTION IN AI III...
PPTX
Artificial Intelligence Searching Techniques
PPT
KNOWLEDGE REPRESENTATION ISSUES.ppt
PDF
I.BEST FIRST SEARCH IN AI
PPTX
knowledge representation using rules
PPTX
Semantic Networks
PPTX
Artificial Intelligence Notes Unit 3
PPT
AI Lecture 7 (uncertainty)
PPT
3. mining frequent patterns
PPT
First order logic
PPT
Knowledge Representation in Artificial intelligence
PPTX
Knowledge representation In Artificial Intelligence
PPTX
Kr using rules
PPT
Iterative deepening search
PPTX
Rule based system
PPT
AI Lecture 3 (solving problems by searching)
PPTX
Knowledge representation
PPTX
First order logic
PPTX
Structured Knowledge Representation
Knowledge representation in AI
I.INFORMED SEARCH IN ARTIFICIAL INTELLIGENCE II. HEURISTIC FUNCTION IN AI III...
Artificial Intelligence Searching Techniques
KNOWLEDGE REPRESENTATION ISSUES.ppt
I.BEST FIRST SEARCH IN AI
knowledge representation using rules
Semantic Networks
Artificial Intelligence Notes Unit 3
AI Lecture 7 (uncertainty)
3. mining frequent patterns
First order logic
Knowledge Representation in Artificial intelligence
Knowledge representation In Artificial Intelligence
Kr using rules
Iterative deepening search
Rule based system
AI Lecture 3 (solving problems by searching)
Knowledge representation
First order logic
Structured Knowledge Representation
Ad

Similar to Knowledge Representation & Reasoning AI UNIT 3 (20)

PPTX
Module4_AI 4th semester engineering.pptx
PDF
Ai lecture 10(unit03)
PPTX
Foundations of Knowledge Representation in Artificial Intelligence.pptx
PDF
22PCOAM11_IAI_Unit III Notes Full Notesmerged.pdf
PPTX
lecture-inference-in-first-order-logic.pptx
PDF
lecture-inference-in-first-order-logic.pdf
PPTX
Propositional logic(part 2)
PPTX
Untitled presentation in first order logic .pptx
PPTX
Inference in First-Order Logic
PDF
First Order Logic resolution
PPT
KNOWLEDGE Representation unit 3 for data mining
PDF
unit-3 First half..pdf nice ppt helps in ai intelligence
PPTX
First order logic
PPTX
ARTIFICIAL INTELLIGENCE---UNIT 4.pptx
PPTX
Basic Knowledge Representation and Reasonong
PPTX
Natural language processing: word senses and relations
PPTX
Jarrar: First Order Logic- Inference Methods
PDF
AI3391 Artificial intelligence Unit IV Notes _ merged.pdf
PPTX
Knowledge Representation and Reasoning.pptx
PDF
First order logic
Module4_AI 4th semester engineering.pptx
Ai lecture 10(unit03)
Foundations of Knowledge Representation in Artificial Intelligence.pptx
22PCOAM11_IAI_Unit III Notes Full Notesmerged.pdf
lecture-inference-in-first-order-logic.pptx
lecture-inference-in-first-order-logic.pdf
Propositional logic(part 2)
Untitled presentation in first order logic .pptx
Inference in First-Order Logic
First Order Logic resolution
KNOWLEDGE Representation unit 3 for data mining
unit-3 First half..pdf nice ppt helps in ai intelligence
First order logic
ARTIFICIAL INTELLIGENCE---UNIT 4.pptx
Basic Knowledge Representation and Reasonong
Natural language processing: word senses and relations
Jarrar: First Order Logic- Inference Methods
AI3391 Artificial intelligence Unit IV Notes _ merged.pdf
Knowledge Representation and Reasoning.pptx
First order logic
Ad

More from Dr. SURBHI SAROHA (20)

PPTX
Deep learning(UNIT 3) BY Ms SURBHI SAROHA
PPTX
MOBILE COMPUTING UNIT 2 by surbhi saroha
PPTX
Mobile Computing UNIT 1 by surbhi saroha
PPTX
DEEP LEARNING (UNIT 2 ) by surbhi saroha
PPTX
Introduction to Deep Leaning(UNIT 1).pptx
PPTX
Cloud Computing (Infrastructure as a Service)UNIT 2
PPTX
Management Information System(Unit 2).pptx
PPTX
Searching in Data Structure(Linear search and Binary search)
PPTX
Management Information System(UNIT 1).pptx
PPTX
Introduction to Cloud Computing(UNIT 1).pptx
PPTX
JAVA (UNIT 5)
PPTX
DBMS (UNIT 5)
PPTX
DBMS UNIT 4
PPTX
JAVA(UNIT 4)
PPTX
OOPs & C++(UNIT 5)
PPTX
OOPS & C++(UNIT 4)
PPTX
DBMS UNIT 3
PPTX
JAVA (UNIT 3)
PPTX
Keys in dbms(UNIT 2)
PPTX
DBMS (UNIT 2)
Deep learning(UNIT 3) BY Ms SURBHI SAROHA
MOBILE COMPUTING UNIT 2 by surbhi saroha
Mobile Computing UNIT 1 by surbhi saroha
DEEP LEARNING (UNIT 2 ) by surbhi saroha
Introduction to Deep Leaning(UNIT 1).pptx
Cloud Computing (Infrastructure as a Service)UNIT 2
Management Information System(Unit 2).pptx
Searching in Data Structure(Linear search and Binary search)
Management Information System(UNIT 1).pptx
Introduction to Cloud Computing(UNIT 1).pptx
JAVA (UNIT 5)
DBMS (UNIT 5)
DBMS UNIT 4
JAVA(UNIT 4)
OOPs & C++(UNIT 5)
OOPS & C++(UNIT 4)
DBMS UNIT 3
JAVA (UNIT 3)
Keys in dbms(UNIT 2)
DBMS (UNIT 2)

Recently uploaded (20)

PDF
BÀI TẬP BỔ TRỢ 4 KỸ NĂNG TIẾNG ANH 9 GLOBAL SUCCESS - CẢ NĂM - BÁM SÁT FORM Đ...
PPTX
Microbial diseases, their pathogenesis and prophylaxis
PDF
STATICS OF THE RIGID BODIES Hibbelers.pdf
PDF
Insiders guide to clinical Medicine.pdf
PDF
Abdominal Access Techniques with Prof. Dr. R K Mishra
PDF
Black Hat USA 2025 - Micro ICS Summit - ICS/OT Threat Landscape
PPTX
Final Presentation General Medicine 03-08-2024.pptx
PDF
Physiotherapy_for_Respiratory_and_Cardiac_Problems WEBBER.pdf
PDF
2.FourierTransform-ShortQuestionswithAnswers.pdf
PDF
Supply Chain Operations Speaking Notes -ICLT Program
PDF
grade 11-chemistry_fetena_net_5883.pdf teacher guide for all student
PDF
Saundersa Comprehensive Review for the NCLEX-RN Examination.pdf
PDF
FourierSeries-QuestionsWithAnswers(Part-A).pdf
PDF
Pre independence Education in Inndia.pdf
PDF
01-Introduction-to-Information-Management.pdf
PDF
O7-L3 Supply Chain Operations - ICLT Program
PDF
The Lost Whites of Pakistan by Jahanzaib Mughal.pdf
PDF
VCE English Exam - Section C Student Revision Booklet
PPTX
GDM (1) (1).pptx small presentation for students
PPTX
Cell Types and Its function , kingdom of life
BÀI TẬP BỔ TRỢ 4 KỸ NĂNG TIẾNG ANH 9 GLOBAL SUCCESS - CẢ NĂM - BÁM SÁT FORM Đ...
Microbial diseases, their pathogenesis and prophylaxis
STATICS OF THE RIGID BODIES Hibbelers.pdf
Insiders guide to clinical Medicine.pdf
Abdominal Access Techniques with Prof. Dr. R K Mishra
Black Hat USA 2025 - Micro ICS Summit - ICS/OT Threat Landscape
Final Presentation General Medicine 03-08-2024.pptx
Physiotherapy_for_Respiratory_and_Cardiac_Problems WEBBER.pdf
2.FourierTransform-ShortQuestionswithAnswers.pdf
Supply Chain Operations Speaking Notes -ICLT Program
grade 11-chemistry_fetena_net_5883.pdf teacher guide for all student
Saundersa Comprehensive Review for the NCLEX-RN Examination.pdf
FourierSeries-QuestionsWithAnswers(Part-A).pdf
Pre independence Education in Inndia.pdf
01-Introduction-to-Information-Management.pdf
O7-L3 Supply Chain Operations - ICLT Program
The Lost Whites of Pakistan by Jahanzaib Mughal.pdf
VCE English Exam - Section C Student Revision Booklet
GDM (1) (1).pptx small presentation for students
Cell Types and Its function , kingdom of life

Knowledge Representation & Reasoning AI UNIT 3

  • 2. SYLLABUS  Propositional logic  Theory of first order logic  Inference in First order logic  Forward & Backward chaining  Resolution  Probabilistic reasoning  Utility theory  Hidden Markov Models(HMM)  Bayesian Networks. 2SURBHI SAROHA
  • 3. Propositional logic  This logic is synonymous with sentential logic, provides ways and means of combining or changing the propositions to create a complicated structure and to build a new logical relationship and properties.  It combines logical connections of all the constituent statements and the true value of the complex statement is derived.  While doing, it considers other factors like reasoning, relationship, and interconnection between the constituent statements. 3SURBHI SAROHA
  • 4. Cont….  This logic provides better clarity on data and information in an incomplete environment by deeper analysis and inference of the limited information presented to it.  Due to this unique feature and powerful algorithms deployed in logical reasoning, it is widely adopted in advanced search in AI in solving complex problems. 4SURBHI SAROHA
  • 5. What is propositional logic?  The natural language words may have slightly different meanings.  Example: A ∧ B and B ∧ A should always have the same meaning.  But the sentences  She became sick and she went to the doctor.  and  She went to the doctor and she became sick. have different meanings. 5SURBHI SAROHA
  • 6. Theory of first order logic  In the topic of Propositional logic, we have seen that how to represent statements using propositional logic.  But unfortunately, in propositional logic, we can only represent the facts, which are either true or false.  PL is not sufficient to represent the complex sentences or natural language statements.  The propositional logic has very limited expressive power. Consider the following sentence, which we cannot represent using PL logic.  "Some humans are intelligent", or  "Sachin likes cricket."  To represent the above statements, PL logic is not sufficient, so we required some more powerful logic, such as first-order logic. 6SURBHI SAROHA
  • 7. First-Order logic:  First-order logic is another way of knowledge representation in artificial intelligence. It is an extension to propositional logic.  FOL is sufficiently expressive to represent the natural language statements in a concise way.  First-order logic is also known as Predicate logic or First- order predicate logic. First-order logic is a powerful language that develops information about the objects in a more easy way and can also express the relationship between those objects.  First-order logic (like natural language) does not only assume that the world contains facts like propositional logic but also assumes the following things in the world: 7SURBHI SAROHA
  • 8. Cont….  Objects: A, B, people, numbers, colors, wars, theories, squares, pits, …  Relations: It can be unary relation such as: red, round, is adjacent, or n-any relation such as: the sister of, brother of, has color, comes between  Function: Father of, best friend, third inning of, end of, ......  As a natural language, first-order logic also has two main parts:  Syntax  Semantics 8SURBHI SAROHA
  • 9. Inference in First order logic  Inference in First-Order Logic is used to deduce new facts or sentences from existing sentences.  Before understanding the FOL inference rule, let's understand some basic terminologies used in FOL.  Substitution:  Substitution is a fundamental operation performed on terms and formulas.  It occurs in all inference systems in first-order logic.  The substitution is complex in the presence of quantifiers in FOL.  If we write F[a/x], so it refers to substitute a constant "a" in place of variable "x". 9SURBHI SAROHA
  • 10. Cont….  Equality:  First-Order logic does not only use predicate and terms for making atomic sentences but also uses another way, which is equality in FOL. For this, we can use equality symbols which specify that the two terms refer to the same object.  Example: Brother (John) = Smith.  As in the above example, the object referred by the Brother (John) is similar to the object referred by Smith. The equality symbol can also be used with negation to represent that two terms are not the same objects.  Example: ¬(x=y) which is equivalent to x ≠y. 10SURBHI SAROHA
  • 11. FOL inference rules for quantifier:  As propositional logic we also have inference rules in first-order logic, so following are some basic inference rules in FOL:  Universal Generalization  Universal Instantiation  Existential Instantiation  Existential introduction 11SURBHI SAROHA
  • 12. 1. Universal Generalization  Universal generalization is a valid inference rule which states that if premise P(c) is true for any arbitrary element c in the universe of discourse, then we can have a conclusion as ∀ x P(x).  It can be represented as: .  This rule can be used if we want to show that every element has a similar property.  In this rule, x must not appear as a free variable.  Example: Let's represent, P(c): "A byte contains 8 bits", so for ∀ x P(x) "All bytes contain 8 bits.", it will also be true. 12SURBHI SAROHA
  • 13. 2.Universal Instantiation:  Universal instantiation is also called as universal elimination or UI is a valid inference rule. It can be applied multiple times to add new sentences.  As per UI, we can infer any sentence obtained by substituting a ground term for the variable.  The UI rule state that we can infer any sentence P(c) by substituting a ground term c (a constant within domain x) from ∀ x P(x) for any object in the universe of discourse.  It can be represented as:.  Example:1.  IF "Every person like ice-cream"=> ∀x P(x) so we can infer that "John likes ice-cream" => P(c) 13SURBHI SAROHA
  • 14. Cont…  Example: 2.  Let's take a famous example,  "All kings who are greedy are Evil." So let our knowledge base contains this detail as in the form of FOL:  ∀x king(x) ∧ greedy (x) → Evil (x),  So from this information, we can infer any of the following statements using Universal Instantiation:  King(John) ∧ Greedy (John) → Evil (John),  King(Richard) ∧ Greedy (Richard) → Evil (Richard),  King(Father(John)) ∧ Greedy (Father(John)) → Evil (Father(John)), 14SURBHI SAROHA
  • 15. 3.Existential Instantiation:  Existential instantiation is also called as Existential Elimination, which is a valid inference rule in first-order logic.  It can be applied only once to replace the existential sentence.  The new KB is not logically equivalent to old KB, but it will be satisfiable if old KB was satisfiable.  This rule states that one can infer P(c) from the formula given in the form of ∃x P(x) for a new constant symbol c.  The restriction with this rule is that c used in the rule must be a new term for which P(c ) is true.  It can be represented as: 15SURBHI SAROHA
  • 16. Cont….  Example:  From the given sentence: ∃x Crown(x) ∧ OnHead(x, John),  So we can infer: Crown(K) ∧ OnHead( K, John), as long as K does not appear in the knowledge base.  The above used K is a constant symbol, which is called Skolem constant.  The Existential instantiation is a special case of Skolemization process. 16SURBHI SAROHA
  • 17. 4.Existential introduction  An existential introduction is also known as an existential generalization, which is a valid inference rule in first-order logic.  This rule states that if there is some element c in the universe of discourse which has a property P, then we can infer that there exists something in the universe which has the property P.  It can be represented as:  Example: Let's say that, "Priyanka got good marks in English." "Therefore, someone got good marks in English." 17SURBHI SAROHA
  • 18. Forward & Backward chaining  Forward chaining as the name suggests, start from the known facts and move forward by applying inference rules to extract more data, and it continues until it reaches to the goal, whereas backward chaining starts from the goal, move backward by using inference rules to determine the facts that satisfy the goal. 18SURBHI SAROHA
  • 19. Forward Chaining  Forward chaining is also known as a forward deduction or forward reasoning method when using an inference engine.  Forward chaining is a form of reasoning which start with atomic sentences in the knowledge base and applies inference rules (Modus Ponens) in the forward direction to extract more data until a goal is reached.  The Forward-chaining algorithm starts from known facts, triggers all rules whose premises are satisfied, and add their conclusion to the known facts. This process repeats until the problem is solved. 19SURBHI SAROHA
  • 20. Properties of Forward-Chaining:  It is a down-up approach, as it moves from bottom to top.  It is a process of making a conclusion based on known facts or data, by starting from the initial state and reaches the goal state.  Forward-chaining approach is also called as data- driven as we reach to the goal using available data.  Forward -chaining approach is commonly used in the expert system, such as CLIPS, business, and production rule systems. 20SURBHI SAROHA
  • 21. Example:  "As per the law, it is a crime for an American to sell weapons to hostile nations. Country A, an enemy of America, has some missiles, and all the missiles were sold to it by Robert, who is an American citizen."  Prove that "Robert is criminal."  To solve the above problem, first, we will convert all the above facts into first-order definite clauses, and then we will use a forward-chaining algorithm to reach the goal. 21SURBHI SAROHA
  • 22. Facts Conversion into FOL:  It is a crime for an American to sell weapons to hostile nations. (Let's say p, q, and r are variables) American (p) ∧ weapon(q) ∧ sells (p, q, r) ∧ hostile(r) → Criminal(p) ...(1)  Country A has some missiles. ?p Owns(A, p) ∧ Missile(p). It can be written in two definite clauses by using Existential Instantiation, introducing new Constant T1. Owns(A, T1) ......(2) Missile(T1) .......(3)  All of the missiles were sold to country A by Robert. ?p Missiles(p) ∧ Owns (A, p) → Sells (Robert, p, A) ......(4)  Missiles are weapons. Missile(p) → Weapons (p) .......(5)  Enemy of America is known as hostile. Enemy(p, America) →Hostile(p) ........(6) 22SURBHI SAROHA
  • 23. Cont….  Country A is an enemy of America. Enemy (A, America) .........(7)  Robert is American American(Robert). ..........(8) 23SURBHI SAROHA
  • 24. Forward chaining proof:  Step-1:  In the first step we will start with the known facts and will choose the sentences which do not have implications, such as:  American(Robert), Enemy(A, America), Owns(A, T1), and Missile(T1). All these facts will be represented as below. 24SURBHI SAROHA
  • 25. Cont…  Step-2:  At the second step, we will see those facts which infer from available facts and with satisfied premises.  Rule-(1) does not satisfy premises, so it will not be added in the first iteration.  Rule-(2) and (3) are already added.  Rule-(4) satisfy with the substitution {p/T1}, so Sells (Robert, T1, A) is added, which infers from the conjunction of Rule (2) and (3).  Rule-(6) is satisfied with the substitution(p/A), so Hostile(A) is added and which infers from Rule-(7). 25SURBHI SAROHA
  • 27. Cont….  Step-3:  At step-3, as we can check Rule-(1) is satisfied with the substitution {p/Robert, q/T1, r/A}, so we can add Criminal(Robert) which infers all the available facts. And hence we reached our goal statement. 27SURBHI SAROHA
  • 29. Cont….  Hence it is proved that Robert is Criminal using forward chaining approach. 29SURBHI SAROHA
  • 30. Backward chaining  A backward chaining algorithm is a form of reasoning, which starts with the goal and works backward, chaining through rules to find known facts that support the goal.  Properties of backward chaining: It is known as a top-down approach.  Backward-chaining is based on modus ponens inference rule. 30SURBHI SAROHA
  • 31. Cont…  The Forward-thinking approach is used in AI to help an AI agent solve logical problems by inspecting the data from the previous learnings and then coming to a conclusion full of solutions.  That’s not all, Forward Chaining might as well be used to explore the available information or answer a question or solve a problem.  Forward chaining is extensively used to break down a long and complex logical approach by attaching each step once the previous one is completed.  This way, it goes from beginning to the end with relative ease. 31SURBHI SAROHA
  • 32. Steps for working of Forwarding Chaining  Step 1: We start from the already stated facts, and then, we’ll subsequently choose the facts that do not have any implications at all.  Step 2: Now, we will state those facts that can be inferred from available facts with satisfied premises.  Step 3: In step 3 we can check the given statement that needs to be checked and check whether it is satisfied with the substitution which infers all the previously stated facts.  Thus we reach our goal. 32SURBHI SAROHA
  • 33. Resolution  Resolution is a theorem proving technique that proceeds by building refutation proofs, i.e., proofs by contradictions.  It was invented by a Mathematician John Alan Robinson in the year 1965.  Resolution is used, if there are various statements are given, and we need to prove a conclusion of those statements.  Unification is a key concept in proofs by resolutions.  Resolution is a single inference rule which can efficiently operate on the conjunctive normal form or clausal form.  Clause: Disjunction of literals (an atomic sentence) is called a clause. It is also known as a unit clause. 33SURBHI SAROHA
  • 34. Cont…  Conjunctive Normal Form: A sentence represented as a conjunction of clauses is said to be conjunctive normal form or CNF.  Steps for Resolution:  Conversion of facts into first-order logic.  Convert FOL statements into CNF  Negate the statement which needs to prove (proof by contradiction)  Draw resolution graph (unification).  To better understand all the above steps, we will take an example in which we will apply resolution. 34SURBHI SAROHA
  • 35. Cont….  Example:  John likes all kind of food.  Apple and vegetable are food  Anything anyone eats and not killed is food.  Anil eats peanuts and still alive  Harry eats everything that Anil eats. Prove by resolution that:  John likes peanuts. 35SURBHI SAROHA
  • 36. Probabilistic reasoning  Probabilistic reasoning is a way of knowledge representation where we apply the concept of probability to indicate the uncertainty in knowledge.  In probabilistic reasoning,  we combine probability theory with logic to handle the uncertainty. 36SURBHI SAROHA
  • 37. Cont….  Probabilistic reasoning is a method of representation of knowledge where the concept of probability is applied to indicate the uncertainty in knowledge.  Probabilistic reasoning is used in AI:  When we are unsure of the predicates  When the possibilities of predicates become too large to list down  When it is known that an error occurs during an experiment 37SURBHI SAROHA
  • 38. Bayesian Networks  Bayesian network is a directed acyclic graph model which helps us represent probabilistic data.  A Bayesian network contains two basic components:  Nodes: These represent attributes/functions or data.  Arcs: These represent the dependencies between the nodes. If an arc is present between the two nodes, then those two nodes have some sort of relation between them.  An absence of an arc denotes the absence of any relation between the nodes.  When designing a Bayesian network, we keep the local probability table at each node. 38SURBHI SAROHA
  • 39. Utility theory  Utility Theory is the discipline that lays out the foundation to create and evaluate Utility Functions.  Typically, Utility Theory uses the notion of Expected Utility (EU) as a value that represents the average utility of all possible outcomes of a state, weighted by the probability that the outcome occurs.  The agents use the utility theory for making decisions. It is the mapping from lotteries to the real numbers.  An agent is supposed to have various preferences and can choose the one which best fits his necessity. 39SURBHI SAROHA
  • 40. Hidden Markov Models(HMM)  Hidden Markov Model is a statistical Markov model in which the system being modeled is assumed to be a Markov process – call it – with unobservable states.  HMM assumes that there is another process whose behavior "depends" on.  The goal is to learn about by observing.  Hidden Markov Model (HMM) is a statistical Markov model in which the system being modeled is assumed to be a Markov process – call it {displaystyle X} – with unobservable ("hidden") states.  HMM assumes that there is another process {displaystyle Y} whose behavior "depends" on {displaystyle X}.  The goal is to learn about {displaystyle X} by observing {displaystyle Y} 40SURBHI SAROHA