SlideShare a Scribd company logo
A Semi-naive Bayes
Classifier with Grouping
of Cases
J. Abellán, A. Cano, A. R. Masegosa, S. Moral
Department of Computer Science and A.I.
University of Granada
Spain
2
Outline
1. Introduction.
2. Semi-Naive Bayes Classifier with
Grouping of Cases.
 General Description
 The Joining Criterions
 The Grouping Criterions
3. Experimental Evaluation.
4. Conclusions and Future Work.
3
Introduction
Information from a data base
Attribute variables Class variable
Data Base
Calcium Tumor Coma Migraine Cancer
normal a1 absent absent absent
high a1 present absent present
normal a1 absent absent absent
normal a1 absent absent absent
high ao present present absent
...... ...... ...... ...... ......
4
Introduction
Naive Bayes (Duda & Hart, 1973)
 Attribute variables {Xi | i=1,..,r}
 Class variable C={c1,..,ck}.
 New observation z=(z1,..,zr) 
(X1=z1,..,Xr=zr).
 Select state of C:
arg maxci
(P(ci|Z)).
 Supposition of independecy
known the class variable:
arg maxci
(P(ci) ∏r
j=1
P(zj|ci))
…
C
X1 X2 Xr
Graphical Structure
5
Introduction
Naive Bayes Classifiers
 Naive Bayesian Classifiers:
NB’s performance is comparable with some
state-of-the-art classifiers even when its
independency assumption does not hold in
normal cases.
 Question:
“Can the performance be better when the
conditional independency assumption of NB is
relaxed?”
6
 Semi-Naive Bayesian Classifiers(SNB)
 A looser assumption than NB.
 Independency occurs among the joined
variables given the class variable C.
Introduction
Semi-Naive Bayes Classifiers
7
Introduction
Semi-Naive Bayes Classifiers
 Main problems of Semi-NB approach:
 When to join two variables? Joining Criterion
 Kononenko’s criterion is entropy based.
 Pazzani’s criterion is accuracy based.
 Wrapper estimation.
 Very high complexity with high number of variables.
Class entropy reduction
8
A SNB with Grouping of Cases
Joining Method
 Three new proposals for Joining Criterions.
 BDe: Bayesian Dirichlet Equivalent.
 L10: The Expected Log-likelihood under
leaving-one-out.
 LRT: Log-likelihood Ratio Test.
9
A SNB with Grouping of Cases
Grouping Method
 Increment in Parameter Estimations
 Solution: “Grouping cases of the new variable”.
Independent
P (Xi | C)P(Xj | C)
Nº Parameters:
#(C) (#(Xi) + #(Xj))
Dependent
P (Xi, Xj | C)
Nº Parameters:
#(C) #(Xi) #(Xj)
Similar Information
10
A SNB with Grouping of Cases
Example
…
C
X1 X2 Xr
Joining Phase
…
C
X5 x X9 X1 Xr
Each pair of Variables
is evaluated using a JC
Grouping Phase
Similar Information
Each pair of Cases
is evaluated using a GC
…
C
X5 x X9 X1 Xr
11
Joining Criterions
BDe criterion
 Bayesian Dirichlet equivalent Metric (BDe)
“Bayesian scores measure the quality of a
model, M, as the posterior probability of
the model given the learning data D”
JC(BDe) = Score (M1:D) – Score(M2:D)
C
X Y
C
X x Y
M1 M2
12
Joining Criterions
L1O criterion
 Expected Log-Likelihood Under Leave-
One-Out (L1O).
Leave-one-out EstimationLaplace Estimation
“The estimation of the log-likelihood of the class
is carried out with a leave-one-out scheme
computed with a closed equation”
13
Joining Criterions
LRT criterion
 Log-likelihood Ratio Test (LRT):
Corrector Factor:
“Comparison of two nested models: M1 with
merged variables and M2 variables are independent”
Number of total
comparisons over
n active variables
14
Grouping Method
Hypotheses
 Hypotheses: Model Selection Problem
 Sample data D is restricted to X=xi or X=xj.
 Consider xi and xj the only possible cases of X.
 Grouping xi and xj implies X has only one case.
Similar Information
15
Grouping Method
Criterions
 BDe score:
 L10 score:
 LRT score:
16
Experimental Evaluation
Details
 SNG was implemented in Elvira.
 Integrated in Weka for evaluation.
 Tested in 13 data bases without missing
values from UCI repository.
 10 fold-cross validation repeated 10 times.
 Comparison with a corrected paired t-test
to 5%.
17
 The trade-off between Accuracy and log-
likelihood is better for LRT.
 L10 works badly as joining criterion.
Evaluating Joining Criterions
Naive Bayes Comparison
18
Evaluating Joining Criterions
Pazzani’s semi-NB comparison
LRT works slightly better than BDe.
Similar performance with a lower time
complexity.
LRT is the best joining criterion
19
Evaluating Grouping Criterions
Naive Bayes Comparison
 LRT Joining + Grouping Method
Not strong differences among criterions.
L10 slightly better.
L1O is the best grouping criterion
20
Pazzani’s Semi-NB Comparison
SNB-G = LRT Joining + L10 Grouping
 Similar performance:
 Dramatic building time reduction:
21
State-of-the-art Classifiers
AODE, TAN and LBR comparison
 Three wins against
NB.
 1 W vs 1 D against
AODE.
 None difference
against TAN and
LBR.
 One Win against
Pazzani’s Semi-NB.
22
Conclusions and Future Work
 A preprocessing step for Naive Bayes:
 Method for joining variables.
 Combined method for grouping cases.
 Very efficient with similar performance
respect to Pazzani’s Semi-NB classifier.
 Application to high-dimensionality data sets.
 Generalization of the methodology to
another models: decision trees and TAN
model.

More Related Content

PDF
New approaches for boosting to uniformity
PDF
Lecture10 - Naïve Bayes
PDF
Design Pattern Explained CH1
PPTX
Naive bayes
PDF
06 Machine Learning - Naive Bayes
PDF
Wikipedia, Dead Authors, Naive Bayes and Python
PPT
Modified naive bayes model for improved web page classification
PDF
Bayesian Machine Learning & Python – Naïve Bayes (PyData SV 2013)
New approaches for boosting to uniformity
Lecture10 - Naïve Bayes
Design Pattern Explained CH1
Naive bayes
06 Machine Learning - Naive Bayes
Wikipedia, Dead Authors, Naive Bayes and Python
Modified naive bayes model for improved web page classification
Bayesian Machine Learning & Python – Naïve Bayes (PyData SV 2013)

Viewers also liked (11)

PPT
Naive Bayes with Conditionally Dependent Data
PPTX
"Naive Bayes Classifier" @ Papers We Love Bucharest
PDF
02. naive bayes classifier revision
PDF
Naive Bayes Classifier
PDF
DMTM 2015 - 05 Association Rules
PDF
Naive Bayes
PPTX
Sentiment analysis using naive bayes classifier
PPTX
Naive Bayes Presentation
PDF
Modeling Social Data, Lecture 6: Classification with Naive Bayes
PPTX
Classification with Naive Bayes
PDF
2013-1 Machine Learning Lecture 03 - Naïve Bayes Classifiers
Naive Bayes with Conditionally Dependent Data
"Naive Bayes Classifier" @ Papers We Love Bucharest
02. naive bayes classifier revision
Naive Bayes Classifier
DMTM 2015 - 05 Association Rules
Naive Bayes
Sentiment analysis using naive bayes classifier
Naive Bayes Presentation
Modeling Social Data, Lecture 6: Classification with Naive Bayes
Classification with Naive Bayes
2013-1 Machine Learning Lecture 03 - Naïve Bayes Classifiers
Ad

Similar to A Semi-naive Bayes Classifier with Grouping of Cases (20)

PPT
UNIT2_NaiveBayes algorithms used in machine learning
PDF
Machine learning naive bayes and svm.pdf
PPT
Lecture07_ Naive Bayes Classifier Machine Learning
PDF
Bayes 6
PPTX
SVM - Functional Verification
PPT
bayes answer jejisiowwoowwksknejejrjejej
PPT
bayesNaive.ppt
PPT
bayesNaive.ppt
PPT
bayesNaive algorithm in machine learning
PDF
classification_clean.pdf
PPT
9-Decision Tree Induction-23-01-2025.ppt
PDF
Artificial Neural Network Lect4 : Single Layer Perceptron Classifiers
PDF
NBaysian classifier, Naive Bayes classifier
PPT
Machine learning by Dr. Vivek Vijay and Dr. Sandeep Yadav
PPTX
Introduction to Machine Learning Concepts
PPT
NaiveBayes_machine-learning(basic_ppt).ppt
PDF
Data classification sammer
PDF
Bayesian data analysis1
PDF
18 ijcse-01232
PDF
lecture 5 about lecture 5 about lecture lecture
UNIT2_NaiveBayes algorithms used in machine learning
Machine learning naive bayes and svm.pdf
Lecture07_ Naive Bayes Classifier Machine Learning
Bayes 6
SVM - Functional Verification
bayes answer jejisiowwoowwksknejejrjejej
bayesNaive.ppt
bayesNaive.ppt
bayesNaive algorithm in machine learning
classification_clean.pdf
9-Decision Tree Induction-23-01-2025.ppt
Artificial Neural Network Lect4 : Single Layer Perceptron Classifiers
NBaysian classifier, Naive Bayes classifier
Machine learning by Dr. Vivek Vijay and Dr. Sandeep Yadav
Introduction to Machine Learning Concepts
NaiveBayes_machine-learning(basic_ppt).ppt
Data classification sammer
Bayesian data analysis1
18 ijcse-01232
lecture 5 about lecture 5 about lecture lecture
Ad

More from NTNU (18)

PDF
Varying parameter in classification based on imprecise probabilities
PDF
An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...
PDF
Bagging Decision Trees on Data Sets with Classification Noise
PDF
lassification with decision trees from a nonparametric predictive inference p...
PDF
Locally Averaged Bayesian Dirichlet Metrics
PDF
Application of a Selective Gaussian Naïve Bayes Model for Diffuse-Large B-Cel...
PDF
An interactive approach for cleaning noisy observations in Bayesian networks ...
PDF
Learning classifiers from discretized expression quantitative trait loci
PDF
Split Criterions for Variable Selection Using Decision Trees
PDF
Combining Decision Trees Based on Imprecise Probabilities and Uncertainty Mea...
PDF
Interactive Learning of Bayesian Networks
PDF
A Bayesian approach to estimate probabilities in classification trees
PDF
A Bayesian Random Split to Build Ensembles of Classification Trees
PDF
An Experimental Study about Simple Decision Trees for Bagging Ensemble on Dat...
PDF
Selective Gaussian Naïve Bayes Model for Diffuse Large-B-Cell Lymphoma Classi...
PDF
Evaluating query-independent object features for relevancy prediction
PDF
Effects of Highly Agreed Documents in Relevancy Prediction
PDF
Conference poster 6
Varying parameter in classification based on imprecise probabilities
An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...
Bagging Decision Trees on Data Sets with Classification Noise
lassification with decision trees from a nonparametric predictive inference p...
Locally Averaged Bayesian Dirichlet Metrics
Application of a Selective Gaussian Naïve Bayes Model for Diffuse-Large B-Cel...
An interactive approach for cleaning noisy observations in Bayesian networks ...
Learning classifiers from discretized expression quantitative trait loci
Split Criterions for Variable Selection Using Decision Trees
Combining Decision Trees Based on Imprecise Probabilities and Uncertainty Mea...
Interactive Learning of Bayesian Networks
A Bayesian approach to estimate probabilities in classification trees
A Bayesian Random Split to Build Ensembles of Classification Trees
An Experimental Study about Simple Decision Trees for Bagging Ensemble on Dat...
Selective Gaussian Naïve Bayes Model for Diffuse Large-B-Cell Lymphoma Classi...
Evaluating query-independent object features for relevancy prediction
Effects of Highly Agreed Documents in Relevancy Prediction
Conference poster 6

Recently uploaded (20)

PPT
The World of Physical Science, • Labs: Safety Simulation, Measurement Practice
PPTX
Comparative Structure of Integument in Vertebrates.pptx
PPTX
Cell Membrane: Structure, Composition & Functions
PPTX
Classification Systems_TAXONOMY_SCIENCE8.pptx
PPTX
GEN. BIO 1 - CELL TYPES & CELL MODIFICATIONS
PPTX
Introduction to Fisheries Biotechnology_Lesson 1.pptx
PPTX
ognitive-behavioral therapy, mindfulness-based approaches, coping skills trai...
PPTX
DRUG THERAPY FOR SHOCK gjjjgfhhhhh.pptx.
PPTX
neck nodes and dissection types and lymph nodes levels
PPTX
EPIDURAL ANESTHESIA ANATOMY AND PHYSIOLOGY.pptx
PDF
MIRIDeepImagingSurvey(MIDIS)oftheHubbleUltraDeepField
PDF
Unveiling a 36 billion solar mass black hole at the centre of the Cosmic Hors...
PPTX
SCIENCE10 Q1 5 WK8 Evidence Supporting Plate Movement.pptx
PPTX
ANEMIA WITH LEUKOPENIA MDS 07_25.pptx htggtftgt fredrctvg
PPTX
microscope-Lecturecjchchchchcuvuvhc.pptx
PDF
bbec55_b34400a7914c42429908233dbd381773.pdf
PDF
diccionario toefl examen de ingles para principiante
PPTX
Microbiology with diagram medical studies .pptx
PPTX
Taita Taveta Laboratory Technician Workshop Presentation.pptx
PPTX
famous lake in india and its disturibution and importance
The World of Physical Science, • Labs: Safety Simulation, Measurement Practice
Comparative Structure of Integument in Vertebrates.pptx
Cell Membrane: Structure, Composition & Functions
Classification Systems_TAXONOMY_SCIENCE8.pptx
GEN. BIO 1 - CELL TYPES & CELL MODIFICATIONS
Introduction to Fisheries Biotechnology_Lesson 1.pptx
ognitive-behavioral therapy, mindfulness-based approaches, coping skills trai...
DRUG THERAPY FOR SHOCK gjjjgfhhhhh.pptx.
neck nodes and dissection types and lymph nodes levels
EPIDURAL ANESTHESIA ANATOMY AND PHYSIOLOGY.pptx
MIRIDeepImagingSurvey(MIDIS)oftheHubbleUltraDeepField
Unveiling a 36 billion solar mass black hole at the centre of the Cosmic Hors...
SCIENCE10 Q1 5 WK8 Evidence Supporting Plate Movement.pptx
ANEMIA WITH LEUKOPENIA MDS 07_25.pptx htggtftgt fredrctvg
microscope-Lecturecjchchchchcuvuvhc.pptx
bbec55_b34400a7914c42429908233dbd381773.pdf
diccionario toefl examen de ingles para principiante
Microbiology with diagram medical studies .pptx
Taita Taveta Laboratory Technician Workshop Presentation.pptx
famous lake in india and its disturibution and importance

A Semi-naive Bayes Classifier with Grouping of Cases

  • 1. A Semi-naive Bayes Classifier with Grouping of Cases J. Abellán, A. Cano, A. R. Masegosa, S. Moral Department of Computer Science and A.I. University of Granada Spain
  • 2. 2 Outline 1. Introduction. 2. Semi-Naive Bayes Classifier with Grouping of Cases.  General Description  The Joining Criterions  The Grouping Criterions 3. Experimental Evaluation. 4. Conclusions and Future Work.
  • 3. 3 Introduction Information from a data base Attribute variables Class variable Data Base Calcium Tumor Coma Migraine Cancer normal a1 absent absent absent high a1 present absent present normal a1 absent absent absent normal a1 absent absent absent high ao present present absent ...... ...... ...... ...... ......
  • 4. 4 Introduction Naive Bayes (Duda & Hart, 1973)  Attribute variables {Xi | i=1,..,r}  Class variable C={c1,..,ck}.  New observation z=(z1,..,zr)  (X1=z1,..,Xr=zr).  Select state of C: arg maxci (P(ci|Z)).  Supposition of independecy known the class variable: arg maxci (P(ci) ∏r j=1 P(zj|ci)) … C X1 X2 Xr Graphical Structure
  • 5. 5 Introduction Naive Bayes Classifiers  Naive Bayesian Classifiers: NB’s performance is comparable with some state-of-the-art classifiers even when its independency assumption does not hold in normal cases.  Question: “Can the performance be better when the conditional independency assumption of NB is relaxed?”
  • 6. 6  Semi-Naive Bayesian Classifiers(SNB)  A looser assumption than NB.  Independency occurs among the joined variables given the class variable C. Introduction Semi-Naive Bayes Classifiers
  • 7. 7 Introduction Semi-Naive Bayes Classifiers  Main problems of Semi-NB approach:  When to join two variables? Joining Criterion  Kononenko’s criterion is entropy based.  Pazzani’s criterion is accuracy based.  Wrapper estimation.  Very high complexity with high number of variables. Class entropy reduction
  • 8. 8 A SNB with Grouping of Cases Joining Method  Three new proposals for Joining Criterions.  BDe: Bayesian Dirichlet Equivalent.  L10: The Expected Log-likelihood under leaving-one-out.  LRT: Log-likelihood Ratio Test.
  • 9. 9 A SNB with Grouping of Cases Grouping Method  Increment in Parameter Estimations  Solution: “Grouping cases of the new variable”. Independent P (Xi | C)P(Xj | C) Nº Parameters: #(C) (#(Xi) + #(Xj)) Dependent P (Xi, Xj | C) Nº Parameters: #(C) #(Xi) #(Xj) Similar Information
  • 10. 10 A SNB with Grouping of Cases Example … C X1 X2 Xr Joining Phase … C X5 x X9 X1 Xr Each pair of Variables is evaluated using a JC Grouping Phase Similar Information Each pair of Cases is evaluated using a GC … C X5 x X9 X1 Xr
  • 11. 11 Joining Criterions BDe criterion  Bayesian Dirichlet equivalent Metric (BDe) “Bayesian scores measure the quality of a model, M, as the posterior probability of the model given the learning data D” JC(BDe) = Score (M1:D) – Score(M2:D) C X Y C X x Y M1 M2
  • 12. 12 Joining Criterions L1O criterion  Expected Log-Likelihood Under Leave- One-Out (L1O). Leave-one-out EstimationLaplace Estimation “The estimation of the log-likelihood of the class is carried out with a leave-one-out scheme computed with a closed equation”
  • 13. 13 Joining Criterions LRT criterion  Log-likelihood Ratio Test (LRT): Corrector Factor: “Comparison of two nested models: M1 with merged variables and M2 variables are independent” Number of total comparisons over n active variables
  • 14. 14 Grouping Method Hypotheses  Hypotheses: Model Selection Problem  Sample data D is restricted to X=xi or X=xj.  Consider xi and xj the only possible cases of X.  Grouping xi and xj implies X has only one case. Similar Information
  • 15. 15 Grouping Method Criterions  BDe score:  L10 score:  LRT score:
  • 16. 16 Experimental Evaluation Details  SNG was implemented in Elvira.  Integrated in Weka for evaluation.  Tested in 13 data bases without missing values from UCI repository.  10 fold-cross validation repeated 10 times.  Comparison with a corrected paired t-test to 5%.
  • 17. 17  The trade-off between Accuracy and log- likelihood is better for LRT.  L10 works badly as joining criterion. Evaluating Joining Criterions Naive Bayes Comparison
  • 18. 18 Evaluating Joining Criterions Pazzani’s semi-NB comparison LRT works slightly better than BDe. Similar performance with a lower time complexity. LRT is the best joining criterion
  • 19. 19 Evaluating Grouping Criterions Naive Bayes Comparison  LRT Joining + Grouping Method Not strong differences among criterions. L10 slightly better. L1O is the best grouping criterion
  • 20. 20 Pazzani’s Semi-NB Comparison SNB-G = LRT Joining + L10 Grouping  Similar performance:  Dramatic building time reduction:
  • 21. 21 State-of-the-art Classifiers AODE, TAN and LBR comparison  Three wins against NB.  1 W vs 1 D against AODE.  None difference against TAN and LBR.  One Win against Pazzani’s Semi-NB.
  • 22. 22 Conclusions and Future Work  A preprocessing step for Naive Bayes:  Method for joining variables.  Combined method for grouping cases.  Very efficient with similar performance respect to Pazzani’s Semi-NB classifier.  Application to high-dimensionality data sets.  Generalization of the methodology to another models: decision trees and TAN model.