SlideShare a Scribd company logo
An interactive approach for cleaning noisy
observations in Bayesian networks with the
help of an expert
Andrés R. Masegosa and Serafín Moral
Department of Computer Science and Artificial Intelligence
University of Granada
Granada, September 2012
PGM 2012 Granada (Spain) 1/31
Outline
1 Introduction
2 Modelling noisy observations and expert knowledge
3 Cleaning noisy observations with the help of expert
knowledge
Entropy-based approach
Cost-based approach
EM algorithm to estimate the noise rate
4 Experimental evaluation
5 Conclusions and future works
PGM 2012 Granada (Spain) 2/31
Introduction
Part I
Introduction
PGM 2012 Granada (Spain) 3/31
Introduction
Bayesian Networks in Intelligent Software Systems
PGM 2012 Granada (Spain) 4/31
Introduction
Evidence Gathering Process
Sensor failure (GPS, Vision, etc).
Noisy transmissions in a communication channel.
Outliers is a particular case of a noisy observation.
Human errors with the GUI of the system.
...
PGM 2012 Granada (Spain) 5/31
Introduction
Problem of Noisy Observations
PGM 2012 Granada (Spain) 6/31
Introduction
Automatic Data Cleaning Methods
Mandatory in most of the practical data mining problems: poor quality data.
Many different approaches are proposed to address this problem (Zhu and Wu, 2006).
They are automatic methods where humans/experts do not play an active role.
PGM 2012 Granada (Spain) 7/31
Introduction
New Cleaning Methods: misspelled words in smartphones
System detects a corrupted noisy observation.
System displays alternative words (i.e. fixed observations).
The user ultimately decides which is the correct one.
The system interacts with the user.
PGM 2012 Granada (Spain) 8/31
Introduction
An Interactive Data Cleaning Method
Noisy observations: some values of the observations are noisy (i.e. different from its actual value).
Our data model is a Bayesian network over multinomial data.
The noisy process needs to be explicitly modelled.
We assume the existence of an expert able to provide knowledge about specific parts of the observation
vector.
PGM 2012 Granada (Spain) 9/31
Modelling Noisy Observations and Expert Knowledge
Part II
Modelling Noisy Observations and
Expert Knowledge
PGM 2012 Granada (Spain) 10/31
Modelling Noisy Observations and Expert Knowledge
Notation
We assume that we have a set of observable variables: O = {O1, ..., Op}.
o is a particular observation vector.
P(O) is modelled by a given Bayesian network.
We assume there is noise when observing these variables.
PGM 2012 Granada (Spain) 11/31
Modelling Noisy Observations and Expert Knowledge
Notation
We assume that we have a set of observable variables: O = {O1, ..., Op}.
o is a particular observation vector.
P(O) is modelled by a given Bayesian network.
We assume there is noise when observing these variables.
O = {O1, ..., Op} noisy observable variables
o is a particular noisy observation vector.
PGM 2012 Granada (Spain) 11/31
Modelling Noisy Observations and Expert Knowledge
Notation
We assume that we have a set of observable variables: O = {O1, ..., Op}.
o is a particular observation vector.
P(O) is modelled by a given Bayesian network.
We assume there is noise when observing these variables.
O = {O1, ..., Op} noisy observable variables
o is a particular noisy observation vector.
Our goal
To detect the noisy observations: oi = oi .
To recover the true observations: o = {o1, ..., op}.
... with the help of an expert.
PGM 2012 Granada (Spain) 11/31
Modelling Noisy Observations and Expert Knowledge
Modelling noisy observations
We artificially add two new variables to each observable variable:
Oi is the noisy observable variable associated to Oi .
PGM 2012 Granada (Spain) 12/31
Modelling Noisy Observations and Expert Knowledge
Modelling noisy observations
We artificially add two new variables to each observable variable:
Oi is the noisy observable variable associated to Oi .
Ni indicates if there is a noisy observation with probability ηi , which is
assumed to be a known value.
PGM 2012 Granada (Spain) 12/31
Modelling Noisy Observations and Expert Knowledge
Modelling noisy observations
We artificially add two new variables to each observable variable:
Oi is the noisy observable variable associated to Oi .
Ni indicates if there is a noisy observation with probability ηi , which is
assumed to be a known value.
The conditional P(Oi |Oi , Ni ) defines the noise model.
PGM 2012 Granada (Spain) 12/31
Modelling Noisy Observations and Expert Knowledge
Modelling expert knowledge
We also add two new variables to each observable variable:
Oe
i is the variable which receives the expert knowledge associated to Oi .
PGM 2012 Granada (Spain) 13/31
Modelling Noisy Observations and Expert Knowledge
Modelling expert knowledge
We also add two new variables to each observable variable:
Oe
i is the variable which receives the expert knowledge associated to Oi .
Ei indicates if the knowledge is correct with probability τi , which is
assumed to be a known value.
PGM 2012 Granada (Spain) 13/31
Modelling Noisy Observations and Expert Knowledge
Modelling expert knowledge
We also add two new variables to each observable variable:
Oe
i is the variable which receives the expert knowledge associated to Oi .
Ei indicates if the knowledge is correct with probability τi , which is
assumed to be a known value.
The conditional P(Oe
i |Oi , Ei ) defines how the expert gives wrong answers.
PGM 2012 Granada (Spain) 13/31
Modelling Noisy Observations and Expert Knowledge
A model for noisy observations and expert knowledge
Automatic cleaning method
Recover the most probable explanation of the observable variables given the
noisy observations:
oMPE
= arg max
O=o
P(O = o|O = o )
PGM 2012 Granada (Spain) 14/31
Modelling Noisy Observations and Expert Knowledge
A model for noisy observations and expert knowledge
Automatic cleaning method
Recover the most probable explanation of the observable variables given the
noisy observations:
oMPE
= arg max
O=o
P(O = o|O = o )
Not a good solution.
There exist alternative explanations, O = o, with
non-negligible probability.
PGM 2012 Granada (Spain) 14/31
Modelling Noisy Observations and Expert Knowledge
A model for noisy observations and expert knowledge
Automatic cleaning method
Recover the most probable explanation of the observable variables given the
noisy observations:
oMPE
= arg max
O=o
P(O = o|O = o )
Not a good solution.
There exist alternative explanations, O = o, with
non-negligible probability.
Use expert knowledge to discard those alternative
explanations.
PGM 2012 Granada (Spain) 14/31
Interactive Cleaning: Entropy Based Approach
Part III
Interactive Cleaning: Entropy
Based Approach
PGM 2012 Granada (Spain) 15/31
Interactive Cleaning: Entropy Based Approach
Cleaning noisy observations with the help of expert knowledge
Entropy Based Approach
Reduce the conditional entropy of the true observations:
H(O|O = o )
The lower this entropy, the stronger our confidence in the oMPE .
PGM 2012 Granada (Spain) 16/31
Interactive Cleaning: Entropy Based Approach
Cleaning noisy observations with the help of expert knowledge
Entropy Based Approach
Reduce the conditional entropy of the true observations:
H(O|O = o )
The lower this entropy, the stronger our confidence in the oMPE .
Our strategy is to request to the expert the knowledge which most reduces
the above entropy (the highest information gain):
IG(O, Oe
i |o )
Expert should submit his/her belief about the true value of Oi .
PGM 2012 Granada (Spain) 16/31
Interactive Cleaning: Entropy Based Approach
Cleaning noisy observations with the help of expert knowledge
Entropy Based Approach
Reduce the conditional entropy of the true observations:
H(O|O = o )
The lower this entropy, the stronger our confidence in the oMPE .
Our strategy is to request to the expert the knowledge which most reduces
the above entropy (the highest information gain):
IG(O, Oe
i |o )
Expert should submit his/her belief about the true value of Oi .
The Oe
i with the highest information gain is the one with the highest entropy:
arg max
Oe
i
IG(O, Oe
i |o ) = arg max
Oe
i
(H(Oe
i |O = o ) − H(Ei ))
PGM 2012 Granada (Spain) 16/31
Interactive Cleaning: Entropy Based Approach
Entropy Based Approach
Algorithm
1: oe
= ∅.
PGM 2012 Granada (Spain) 17/31
Interactive Cleaning: Entropy Based Approach
Entropy Based Approach
Algorithm
1: oe
= ∅.
2: repeat
3: Compute the Oe
i variable with the highest information gain:
Oe
max = arg max
Oe
i
IG(O; Oe
i |o , oe
)
4: if IG(O; Oe
i |o , oe
) > λ then
5: Ask the expert about Oi .
6: oe
= oe
∪ oe
max .
7: end if
8: until end
PGM 2012 Granada (Spain) 17/31
Interactive Cleaning: Entropy Based Approach
Entropy Based Approach
Algorithm
1: oe
= ∅.
2: repeat
3: Compute the Oe
i variable with the highest information gain:
Oe
max = arg max
Oe
i
IG(O; Oe
i |o , oe
)
4: if IG(O; Oe
i |o , oe
) > λ then
5: Ask the expert about Oi .
6: oe
= oe
∪ oe
max .
7: end if
8: until end
9:
10: return oMPE
= arg maxO=o P(O = o|o , oe
);
PGM 2012 Granada (Spain) 17/31
Interactive Cleaning: Cost Based Approach
Part IV
Interactive Cleaning: Cost Based
Approach
PGM 2012 Granada (Spain) 18/31
Interactive Cleaning: Cost Based Approach
Cleaning noisy observations with the help of expert knowledge
Cost Based Approach
We model the problem using as a decision making problem:
Fixing Decisions, Fi , about observation Oi with overall cost, CF.
Asking Decisions, Ai , about observation Oi with cost, CAi .
Problem: find the optimal set of fixing decisions F and asking decisions A with
lowest expected cost.
PGM 2012 Granada (Spain) 19/31
Interactive Cleaning: Cost Based Approach
Cleaning noisy observations with the help of expert knowledge
Cost Based Approach
We model the problem using as a decision making problem:
Fixing Decisions, Fi , about observation Oi with overall cost, CF.
Asking Decisions, Ai , about observation Oi with cost, CAi .
Problem: find the optimal set of fixing decisions F and asking decisions A with
lowest expected cost.
Solving this decision problem
It is computationally unfeasible if the number variables is large:
Unconstrained Decision Problem: no sequential order in the decisions.
Each decision depends on the information of the variables in O ∪ Oe.
Approximate solution based on several simplifications.
PGM 2012 Granada (Spain) 19/31
Interactive Cleaning: Cost Based Approach
Cost Based Approach
Simplifications
The decision problem is solved for a particular noisy observation vector o .
Cost of fixing computed as a sum of independent costs: CF =
p
i=1 CFi .
We assume that we have p different decision problems:
Problem Di involves decisions Ai and {F1, ..., Fp}.
When solving Di we do not ask for the rest of the variables.
PGM 2012 Granada (Spain) 20/31
Interactive Cleaning: Cost Based Approach
Cost Based Approach
Solving Problem Di
Expected cost gain: The difference in the expected cost between asking and
not asking about Oi :
CG(Ai |o , oe
) = c(Dn
i ) − c(Da
i )
c(Dn
i ): expected cost when no asking about Oi .
c(Da
i ): expected cost when asking about Oi .
PGM 2012 Granada (Spain) 21/31
Interactive Cleaning: Cost Based Approach
Cost Based Approach
Solving Problem Di
Expected cost gain: The difference in the expected cost between asking and
not asking about Oi :
CG(Ai |o , oe
) = c(Dn
i ) − c(Da
i )
c(Dn
i ): expected cost when no asking about Oi .
c(Da
i ): expected cost when asking about Oi .
Fixing Decisions: we compute for each decision Fi the value fj such that
fj = arg min
fj oj
CFj (fj , oj )P(oj |o , oe
)
The minimization problem can be solved in constant time after the observations
have been propagated:
0/1 cost error: select the oj with the highest probability.
PGM 2012 Granada (Spain) 21/31
Interactive Cleaning: Cost Based Approach
Cost Based Approach
Algorithm
1: oe
= ∅.
2: repeat
3: Compute the decision Ai the highest expected cost gain:
Amax = arg max
Ai
CG(Ai |o , oe
)
4: if CG(Amax |o , oe
) > 0 then
5: Ask the expert about Oi .
6: oe
= oe
∪ oe
max .
7: end if
8: until end
9:
10: return fj = arg minfj oj
CFj (fj , oj )P(oj |o , oe
);
PGM 2012 Granada (Spain) 22/31
EM algorithm to estimate the noise rate
Part V
EM algorithm to estimate the
noise rate
PGM 2012 Granada (Spain) 23/31
EM algorithm to estimate the noise rate
EM Algorithm to estimate the unknown noise rates τi
We are given a set of M noisy observations:D = {o(1)
, ..., o(M)
}.
PGM 2012 Granada (Spain) 24/31
EM algorithm to estimate the noise rate
EM Algorithm to estimate the unknown noise rates τi
We are given a set of M noisy observations:D = {o(1)
, ..., o(M)
}.
The EM algorithm is applied to estimate the MAP estimate of the parameters
τ = (τ1, ..., τp).
O = {O1, ..., Op} and N = {N1, ..., Np} are the hidden variables.
PGM 2012 Granada (Spain) 24/31
EM algorithm to estimate the noise rate
EM Algorithm to estimate the unknown noise rates τi
We are given a set of M noisy observations:D = {o(1)
, ..., o(M)
}.
The EM algorithm is applied to estimate the MAP estimate of the parameters
τ = (τ1, ..., τp).
O = {O1, ..., Op} and N = {N1, ..., Np} are the hidden variables.
Expectation step: Given a current estimate of τ<k>.
Compute P(Ni = noise|o (j), τ<k>) propagating in the extended BN for
j-th data sample.
Maximization step:
τ<k+1>
i =
j P(Ni = noise|o (j); τ<k>)
M
PGM 2012 Granada (Spain) 24/31
Experimental Evaluation
Part VI
Experimental Evaluation
PGM 2012 Granada (Spain) 25/31
Experimental Evaluation
Experimental Set-up
PGM 2012 Granada (Spain) 26/31
Experimental Evaluation
Experiments on Alarm with 5% noise rate
Noise Rate
With no expert knowledge only a minor proportion of the errors are identified
and we might introduce new errors.
The introduction of the expert knowledge boost the precision of the detected
errors.
PGM 2012 Granada (Spain) 27/31
Experimental Evaluation
Experiments on Alarm with 5% noise rate
Precision
With no expert knowledge only a minor proportion of the errors are identified
and we might introduce new errors.
The introduction of the expert knowledge boost the precision and the recall of
the detected errors.
PGM 2012 Granada (Spain) 28/31
Conclusions and Future Works
Part VII
Conclusions and Future Works
PGM 2012 Granada (Spain) 29/31
Conclusions and Future Works
Conclusions and Future Works
Conclusions:
We have proposed an interactive method for cleaning noisy
observations in Bayesian networks.
One specific module in our intelligent software system.
The performance strongly depends of the particular model.
PGM 2012 Granada (Spain) 30/31
Conclusions and Future Works
Conclusions and Future Works
Conclusions:
We have proposed an interactive method for cleaning noisy
observations in Bayesian networks.
One specific module in our intelligent software system.
The performance strongly depends of the particular model.
Future Works:
Implement this approach in a real intelligent system.
Extend this method assuming that we do not know neither the
parameters of the network nor the structure.
PGM 2012 Granada (Spain) 30/31
Conclusions and Future Works
Thanks for you attention!!!
PGM 2012 Granada (Spain) 31/31

More Related Content

PDF
Large strain solid dynamics in OpenFOAM
PDF
Volume computation and applications
PDF
Hamilton-Jacobi equation on networks: generalized Lax-Hopf formula
PDF
Large strain computational solid dynamics: An upwind cell centred Finite Volu...
PDF
An upwind cell centred Finite Volume Method for nearly incompressible explici...
PPTX
Towards Expertise Modelling for Routing Data Cleaning Tasks within a Communit...
PDF
Internet of things
PPT
Presentation on Data Cleansing
 
Large strain solid dynamics in OpenFOAM
Volume computation and applications
Hamilton-Jacobi equation on networks: generalized Lax-Hopf formula
Large strain computational solid dynamics: An upwind cell centred Finite Volu...
An upwind cell centred Finite Volume Method for nearly incompressible explici...
Towards Expertise Modelling for Routing Data Cleaning Tasks within a Communit...
Internet of things
Presentation on Data Cleansing
 

Viewers also liked (6)

PPT
Brief Introduction to the 12 Steps of Evaluation Data Cleaning
PDF
Data Cleansing introduction (for BigClean Prague 2011)
PDF
Lifecycle Approach to Cleaning Validation
PDF
Data Cleaning Process
PPTX
Data Cleaning Techniques
Brief Introduction to the 12 Steps of Evaluation Data Cleaning
Data Cleansing introduction (for BigClean Prague 2011)
Lifecycle Approach to Cleaning Validation
Data Cleaning Process
Data Cleaning Techniques
Ad

Similar to An interactive approach for cleaning noisy observations in Bayesian networks with the help of an expert (12)

PPTX
2014 CISAP6 Atmospheric Turbulent Dispersion Modeling Methods using Machine L...
PDF
MAXENTROPIC APPROACH TO DECOMPOUND AGGREGATE RISK LOSSES
PDF
Reinforcement Learning in Configurable Environments
PDF
T he SPL - IT Query by Example Search on Speech system for MediaEval 2014
PDF
Baocaoseminar2012
PPT
Hardness of approximation
PPTX
Master Thesis Defense
PDF
AggNet: Deep Learning from Crowds
PDF
An Optimal Iterative Algorithm for Extracting MUCs in a Black-box Constraint ...
PDF
Industry equilibrium with os and propietary firms
PDF
Adaptive Machine Learning for Credit Card Fraud Detection
PDF
SSBSE11b.ppt
2014 CISAP6 Atmospheric Turbulent Dispersion Modeling Methods using Machine L...
MAXENTROPIC APPROACH TO DECOMPOUND AGGREGATE RISK LOSSES
Reinforcement Learning in Configurable Environments
T he SPL - IT Query by Example Search on Speech system for MediaEval 2014
Baocaoseminar2012
Hardness of approximation
Master Thesis Defense
AggNet: Deep Learning from Crowds
An Optimal Iterative Algorithm for Extracting MUCs in a Black-box Constraint ...
Industry equilibrium with os and propietary firms
Adaptive Machine Learning for Credit Card Fraud Detection
SSBSE11b.ppt
Ad

More from NTNU (18)

PDF
Varying parameter in classification based on imprecise probabilities
PDF
An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...
PDF
Bagging Decision Trees on Data Sets with Classification Noise
PDF
lassification with decision trees from a nonparametric predictive inference p...
PDF
Locally Averaged Bayesian Dirichlet Metrics
PDF
Application of a Selective Gaussian Naïve Bayes Model for Diffuse-Large B-Cel...
PDF
Learning classifiers from discretized expression quantitative trait loci
PDF
Split Criterions for Variable Selection Using Decision Trees
PDF
A Semi-naive Bayes Classifier with Grouping of Cases
PDF
Combining Decision Trees Based on Imprecise Probabilities and Uncertainty Mea...
PDF
Interactive Learning of Bayesian Networks
PDF
A Bayesian approach to estimate probabilities in classification trees
PDF
A Bayesian Random Split to Build Ensembles of Classification Trees
PDF
An Experimental Study about Simple Decision Trees for Bagging Ensemble on Dat...
PDF
Selective Gaussian Naïve Bayes Model for Diffuse Large-B-Cell Lymphoma Classi...
PDF
Evaluating query-independent object features for relevancy prediction
PDF
Effects of Highly Agreed Documents in Relevancy Prediction
PDF
Conference poster 6
Varying parameter in classification based on imprecise probabilities
An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...
Bagging Decision Trees on Data Sets with Classification Noise
lassification with decision trees from a nonparametric predictive inference p...
Locally Averaged Bayesian Dirichlet Metrics
Application of a Selective Gaussian Naïve Bayes Model for Diffuse-Large B-Cel...
Learning classifiers from discretized expression quantitative trait loci
Split Criterions for Variable Selection Using Decision Trees
A Semi-naive Bayes Classifier with Grouping of Cases
Combining Decision Trees Based on Imprecise Probabilities and Uncertainty Mea...
Interactive Learning of Bayesian Networks
A Bayesian approach to estimate probabilities in classification trees
A Bayesian Random Split to Build Ensembles of Classification Trees
An Experimental Study about Simple Decision Trees for Bagging Ensemble on Dat...
Selective Gaussian Naïve Bayes Model for Diffuse Large-B-Cell Lymphoma Classi...
Evaluating query-independent object features for relevancy prediction
Effects of Highly Agreed Documents in Relevancy Prediction
Conference poster 6

Recently uploaded (20)

PDF
SEHH2274 Organic Chemistry Notes 1 Structure and Bonding.pdf
PPTX
SCIENCE10 Q1 5 WK8 Evidence Supporting Plate Movement.pptx
PDF
Phytochemical Investigation of Miliusa longipes.pdf
PPTX
DRUG THERAPY FOR SHOCK gjjjgfhhhhh.pptx.
PDF
. Radiology Case Scenariosssssssssssssss
PDF
Mastering Bioreactors and Media Sterilization: A Complete Guide to Sterile Fe...
PPTX
EPIDURAL ANESTHESIA ANATOMY AND PHYSIOLOGY.pptx
PPTX
INTRODUCTION TO EVS | Concept of sustainability
PDF
Unveiling a 36 billion solar mass black hole at the centre of the Cosmic Hors...
PPT
The World of Physical Science, • Labs: Safety Simulation, Measurement Practice
PPTX
2. Earth - The Living Planet Module 2ELS
PPTX
microscope-Lecturecjchchchchcuvuvhc.pptx
PPTX
Cell Membrane: Structure, Composition & Functions
PPTX
GEN. BIO 1 - CELL TYPES & CELL MODIFICATIONS
PDF
IFIT3 RNA-binding activity primores influenza A viruz infection and translati...
PPTX
Derivatives of integument scales, beaks, horns,.pptx
PDF
VARICELLA VACCINATION: A POTENTIAL STRATEGY FOR PREVENTING MULTIPLE SCLEROSIS
PPTX
Introduction to Fisheries Biotechnology_Lesson 1.pptx
PPT
Chemical bonding and molecular structure
PDF
MIRIDeepImagingSurvey(MIDIS)oftheHubbleUltraDeepField
SEHH2274 Organic Chemistry Notes 1 Structure and Bonding.pdf
SCIENCE10 Q1 5 WK8 Evidence Supporting Plate Movement.pptx
Phytochemical Investigation of Miliusa longipes.pdf
DRUG THERAPY FOR SHOCK gjjjgfhhhhh.pptx.
. Radiology Case Scenariosssssssssssssss
Mastering Bioreactors and Media Sterilization: A Complete Guide to Sterile Fe...
EPIDURAL ANESTHESIA ANATOMY AND PHYSIOLOGY.pptx
INTRODUCTION TO EVS | Concept of sustainability
Unveiling a 36 billion solar mass black hole at the centre of the Cosmic Hors...
The World of Physical Science, • Labs: Safety Simulation, Measurement Practice
2. Earth - The Living Planet Module 2ELS
microscope-Lecturecjchchchchcuvuvhc.pptx
Cell Membrane: Structure, Composition & Functions
GEN. BIO 1 - CELL TYPES & CELL MODIFICATIONS
IFIT3 RNA-binding activity primores influenza A viruz infection and translati...
Derivatives of integument scales, beaks, horns,.pptx
VARICELLA VACCINATION: A POTENTIAL STRATEGY FOR PREVENTING MULTIPLE SCLEROSIS
Introduction to Fisheries Biotechnology_Lesson 1.pptx
Chemical bonding and molecular structure
MIRIDeepImagingSurvey(MIDIS)oftheHubbleUltraDeepField

An interactive approach for cleaning noisy observations in Bayesian networks with the help of an expert

  • 1. An interactive approach for cleaning noisy observations in Bayesian networks with the help of an expert Andrés R. Masegosa and Serafín Moral Department of Computer Science and Artificial Intelligence University of Granada Granada, September 2012 PGM 2012 Granada (Spain) 1/31
  • 2. Outline 1 Introduction 2 Modelling noisy observations and expert knowledge 3 Cleaning noisy observations with the help of expert knowledge Entropy-based approach Cost-based approach EM algorithm to estimate the noise rate 4 Experimental evaluation 5 Conclusions and future works PGM 2012 Granada (Spain) 2/31
  • 4. Introduction Bayesian Networks in Intelligent Software Systems PGM 2012 Granada (Spain) 4/31
  • 5. Introduction Evidence Gathering Process Sensor failure (GPS, Vision, etc). Noisy transmissions in a communication channel. Outliers is a particular case of a noisy observation. Human errors with the GUI of the system. ... PGM 2012 Granada (Spain) 5/31
  • 6. Introduction Problem of Noisy Observations PGM 2012 Granada (Spain) 6/31
  • 7. Introduction Automatic Data Cleaning Methods Mandatory in most of the practical data mining problems: poor quality data. Many different approaches are proposed to address this problem (Zhu and Wu, 2006). They are automatic methods where humans/experts do not play an active role. PGM 2012 Granada (Spain) 7/31
  • 8. Introduction New Cleaning Methods: misspelled words in smartphones System detects a corrupted noisy observation. System displays alternative words (i.e. fixed observations). The user ultimately decides which is the correct one. The system interacts with the user. PGM 2012 Granada (Spain) 8/31
  • 9. Introduction An Interactive Data Cleaning Method Noisy observations: some values of the observations are noisy (i.e. different from its actual value). Our data model is a Bayesian network over multinomial data. The noisy process needs to be explicitly modelled. We assume the existence of an expert able to provide knowledge about specific parts of the observation vector. PGM 2012 Granada (Spain) 9/31
  • 10. Modelling Noisy Observations and Expert Knowledge Part II Modelling Noisy Observations and Expert Knowledge PGM 2012 Granada (Spain) 10/31
  • 11. Modelling Noisy Observations and Expert Knowledge Notation We assume that we have a set of observable variables: O = {O1, ..., Op}. o is a particular observation vector. P(O) is modelled by a given Bayesian network. We assume there is noise when observing these variables. PGM 2012 Granada (Spain) 11/31
  • 12. Modelling Noisy Observations and Expert Knowledge Notation We assume that we have a set of observable variables: O = {O1, ..., Op}. o is a particular observation vector. P(O) is modelled by a given Bayesian network. We assume there is noise when observing these variables. O = {O1, ..., Op} noisy observable variables o is a particular noisy observation vector. PGM 2012 Granada (Spain) 11/31
  • 13. Modelling Noisy Observations and Expert Knowledge Notation We assume that we have a set of observable variables: O = {O1, ..., Op}. o is a particular observation vector. P(O) is modelled by a given Bayesian network. We assume there is noise when observing these variables. O = {O1, ..., Op} noisy observable variables o is a particular noisy observation vector. Our goal To detect the noisy observations: oi = oi . To recover the true observations: o = {o1, ..., op}. ... with the help of an expert. PGM 2012 Granada (Spain) 11/31
  • 14. Modelling Noisy Observations and Expert Knowledge Modelling noisy observations We artificially add two new variables to each observable variable: Oi is the noisy observable variable associated to Oi . PGM 2012 Granada (Spain) 12/31
  • 15. Modelling Noisy Observations and Expert Knowledge Modelling noisy observations We artificially add two new variables to each observable variable: Oi is the noisy observable variable associated to Oi . Ni indicates if there is a noisy observation with probability ηi , which is assumed to be a known value. PGM 2012 Granada (Spain) 12/31
  • 16. Modelling Noisy Observations and Expert Knowledge Modelling noisy observations We artificially add two new variables to each observable variable: Oi is the noisy observable variable associated to Oi . Ni indicates if there is a noisy observation with probability ηi , which is assumed to be a known value. The conditional P(Oi |Oi , Ni ) defines the noise model. PGM 2012 Granada (Spain) 12/31
  • 17. Modelling Noisy Observations and Expert Knowledge Modelling expert knowledge We also add two new variables to each observable variable: Oe i is the variable which receives the expert knowledge associated to Oi . PGM 2012 Granada (Spain) 13/31
  • 18. Modelling Noisy Observations and Expert Knowledge Modelling expert knowledge We also add two new variables to each observable variable: Oe i is the variable which receives the expert knowledge associated to Oi . Ei indicates if the knowledge is correct with probability τi , which is assumed to be a known value. PGM 2012 Granada (Spain) 13/31
  • 19. Modelling Noisy Observations and Expert Knowledge Modelling expert knowledge We also add two new variables to each observable variable: Oe i is the variable which receives the expert knowledge associated to Oi . Ei indicates if the knowledge is correct with probability τi , which is assumed to be a known value. The conditional P(Oe i |Oi , Ei ) defines how the expert gives wrong answers. PGM 2012 Granada (Spain) 13/31
  • 20. Modelling Noisy Observations and Expert Knowledge A model for noisy observations and expert knowledge Automatic cleaning method Recover the most probable explanation of the observable variables given the noisy observations: oMPE = arg max O=o P(O = o|O = o ) PGM 2012 Granada (Spain) 14/31
  • 21. Modelling Noisy Observations and Expert Knowledge A model for noisy observations and expert knowledge Automatic cleaning method Recover the most probable explanation of the observable variables given the noisy observations: oMPE = arg max O=o P(O = o|O = o ) Not a good solution. There exist alternative explanations, O = o, with non-negligible probability. PGM 2012 Granada (Spain) 14/31
  • 22. Modelling Noisy Observations and Expert Knowledge A model for noisy observations and expert knowledge Automatic cleaning method Recover the most probable explanation of the observable variables given the noisy observations: oMPE = arg max O=o P(O = o|O = o ) Not a good solution. There exist alternative explanations, O = o, with non-negligible probability. Use expert knowledge to discard those alternative explanations. PGM 2012 Granada (Spain) 14/31
  • 23. Interactive Cleaning: Entropy Based Approach Part III Interactive Cleaning: Entropy Based Approach PGM 2012 Granada (Spain) 15/31
  • 24. Interactive Cleaning: Entropy Based Approach Cleaning noisy observations with the help of expert knowledge Entropy Based Approach Reduce the conditional entropy of the true observations: H(O|O = o ) The lower this entropy, the stronger our confidence in the oMPE . PGM 2012 Granada (Spain) 16/31
  • 25. Interactive Cleaning: Entropy Based Approach Cleaning noisy observations with the help of expert knowledge Entropy Based Approach Reduce the conditional entropy of the true observations: H(O|O = o ) The lower this entropy, the stronger our confidence in the oMPE . Our strategy is to request to the expert the knowledge which most reduces the above entropy (the highest information gain): IG(O, Oe i |o ) Expert should submit his/her belief about the true value of Oi . PGM 2012 Granada (Spain) 16/31
  • 26. Interactive Cleaning: Entropy Based Approach Cleaning noisy observations with the help of expert knowledge Entropy Based Approach Reduce the conditional entropy of the true observations: H(O|O = o ) The lower this entropy, the stronger our confidence in the oMPE . Our strategy is to request to the expert the knowledge which most reduces the above entropy (the highest information gain): IG(O, Oe i |o ) Expert should submit his/her belief about the true value of Oi . The Oe i with the highest information gain is the one with the highest entropy: arg max Oe i IG(O, Oe i |o ) = arg max Oe i (H(Oe i |O = o ) − H(Ei )) PGM 2012 Granada (Spain) 16/31
  • 27. Interactive Cleaning: Entropy Based Approach Entropy Based Approach Algorithm 1: oe = ∅. PGM 2012 Granada (Spain) 17/31
  • 28. Interactive Cleaning: Entropy Based Approach Entropy Based Approach Algorithm 1: oe = ∅. 2: repeat 3: Compute the Oe i variable with the highest information gain: Oe max = arg max Oe i IG(O; Oe i |o , oe ) 4: if IG(O; Oe i |o , oe ) > λ then 5: Ask the expert about Oi . 6: oe = oe ∪ oe max . 7: end if 8: until end PGM 2012 Granada (Spain) 17/31
  • 29. Interactive Cleaning: Entropy Based Approach Entropy Based Approach Algorithm 1: oe = ∅. 2: repeat 3: Compute the Oe i variable with the highest information gain: Oe max = arg max Oe i IG(O; Oe i |o , oe ) 4: if IG(O; Oe i |o , oe ) > λ then 5: Ask the expert about Oi . 6: oe = oe ∪ oe max . 7: end if 8: until end 9: 10: return oMPE = arg maxO=o P(O = o|o , oe ); PGM 2012 Granada (Spain) 17/31
  • 30. Interactive Cleaning: Cost Based Approach Part IV Interactive Cleaning: Cost Based Approach PGM 2012 Granada (Spain) 18/31
  • 31. Interactive Cleaning: Cost Based Approach Cleaning noisy observations with the help of expert knowledge Cost Based Approach We model the problem using as a decision making problem: Fixing Decisions, Fi , about observation Oi with overall cost, CF. Asking Decisions, Ai , about observation Oi with cost, CAi . Problem: find the optimal set of fixing decisions F and asking decisions A with lowest expected cost. PGM 2012 Granada (Spain) 19/31
  • 32. Interactive Cleaning: Cost Based Approach Cleaning noisy observations with the help of expert knowledge Cost Based Approach We model the problem using as a decision making problem: Fixing Decisions, Fi , about observation Oi with overall cost, CF. Asking Decisions, Ai , about observation Oi with cost, CAi . Problem: find the optimal set of fixing decisions F and asking decisions A with lowest expected cost. Solving this decision problem It is computationally unfeasible if the number variables is large: Unconstrained Decision Problem: no sequential order in the decisions. Each decision depends on the information of the variables in O ∪ Oe. Approximate solution based on several simplifications. PGM 2012 Granada (Spain) 19/31
  • 33. Interactive Cleaning: Cost Based Approach Cost Based Approach Simplifications The decision problem is solved for a particular noisy observation vector o . Cost of fixing computed as a sum of independent costs: CF = p i=1 CFi . We assume that we have p different decision problems: Problem Di involves decisions Ai and {F1, ..., Fp}. When solving Di we do not ask for the rest of the variables. PGM 2012 Granada (Spain) 20/31
  • 34. Interactive Cleaning: Cost Based Approach Cost Based Approach Solving Problem Di Expected cost gain: The difference in the expected cost between asking and not asking about Oi : CG(Ai |o , oe ) = c(Dn i ) − c(Da i ) c(Dn i ): expected cost when no asking about Oi . c(Da i ): expected cost when asking about Oi . PGM 2012 Granada (Spain) 21/31
  • 35. Interactive Cleaning: Cost Based Approach Cost Based Approach Solving Problem Di Expected cost gain: The difference in the expected cost between asking and not asking about Oi : CG(Ai |o , oe ) = c(Dn i ) − c(Da i ) c(Dn i ): expected cost when no asking about Oi . c(Da i ): expected cost when asking about Oi . Fixing Decisions: we compute for each decision Fi the value fj such that fj = arg min fj oj CFj (fj , oj )P(oj |o , oe ) The minimization problem can be solved in constant time after the observations have been propagated: 0/1 cost error: select the oj with the highest probability. PGM 2012 Granada (Spain) 21/31
  • 36. Interactive Cleaning: Cost Based Approach Cost Based Approach Algorithm 1: oe = ∅. 2: repeat 3: Compute the decision Ai the highest expected cost gain: Amax = arg max Ai CG(Ai |o , oe ) 4: if CG(Amax |o , oe ) > 0 then 5: Ask the expert about Oi . 6: oe = oe ∪ oe max . 7: end if 8: until end 9: 10: return fj = arg minfj oj CFj (fj , oj )P(oj |o , oe ); PGM 2012 Granada (Spain) 22/31
  • 37. EM algorithm to estimate the noise rate Part V EM algorithm to estimate the noise rate PGM 2012 Granada (Spain) 23/31
  • 38. EM algorithm to estimate the noise rate EM Algorithm to estimate the unknown noise rates τi We are given a set of M noisy observations:D = {o(1) , ..., o(M) }. PGM 2012 Granada (Spain) 24/31
  • 39. EM algorithm to estimate the noise rate EM Algorithm to estimate the unknown noise rates τi We are given a set of M noisy observations:D = {o(1) , ..., o(M) }. The EM algorithm is applied to estimate the MAP estimate of the parameters τ = (τ1, ..., τp). O = {O1, ..., Op} and N = {N1, ..., Np} are the hidden variables. PGM 2012 Granada (Spain) 24/31
  • 40. EM algorithm to estimate the noise rate EM Algorithm to estimate the unknown noise rates τi We are given a set of M noisy observations:D = {o(1) , ..., o(M) }. The EM algorithm is applied to estimate the MAP estimate of the parameters τ = (τ1, ..., τp). O = {O1, ..., Op} and N = {N1, ..., Np} are the hidden variables. Expectation step: Given a current estimate of τ<k>. Compute P(Ni = noise|o (j), τ<k>) propagating in the extended BN for j-th data sample. Maximization step: τ<k+1> i = j P(Ni = noise|o (j); τ<k>) M PGM 2012 Granada (Spain) 24/31
  • 41. Experimental Evaluation Part VI Experimental Evaluation PGM 2012 Granada (Spain) 25/31
  • 43. Experimental Evaluation Experiments on Alarm with 5% noise rate Noise Rate With no expert knowledge only a minor proportion of the errors are identified and we might introduce new errors. The introduction of the expert knowledge boost the precision of the detected errors. PGM 2012 Granada (Spain) 27/31
  • 44. Experimental Evaluation Experiments on Alarm with 5% noise rate Precision With no expert knowledge only a minor proportion of the errors are identified and we might introduce new errors. The introduction of the expert knowledge boost the precision and the recall of the detected errors. PGM 2012 Granada (Spain) 28/31
  • 45. Conclusions and Future Works Part VII Conclusions and Future Works PGM 2012 Granada (Spain) 29/31
  • 46. Conclusions and Future Works Conclusions and Future Works Conclusions: We have proposed an interactive method for cleaning noisy observations in Bayesian networks. One specific module in our intelligent software system. The performance strongly depends of the particular model. PGM 2012 Granada (Spain) 30/31
  • 47. Conclusions and Future Works Conclusions and Future Works Conclusions: We have proposed an interactive method for cleaning noisy observations in Bayesian networks. One specific module in our intelligent software system. The performance strongly depends of the particular model. Future Works: Implement this approach in a real intelligent system. Extend this method assuming that we do not know neither the parameters of the network nor the structure. PGM 2012 Granada (Spain) 30/31
  • 48. Conclusions and Future Works Thanks for you attention!!! PGM 2012 Granada (Spain) 31/31