SlideShare a Scribd company logo
Post pruning methods
Mojtaba amiri
M.S of artificial intelligence
Approaches
• top-down approach :
proceed from the root towards the leaves of tree
• bottom-up approach :
starting the analysis from the leaves and finishing with the root node
Based methods
1. Reduced Error Pruning
2. Pessimistic error pruning
3. Error complexity pruning
4. Minimum Error pruning
5. Cost based pruning
6. Iterative growing and pruning algorithm
Reduced Error Pruning
• This method was proposed by Quinlan .
• This method considers each of the decision nodes in the tree to be
candidates for pruning, consist of removing the subtree rooted at that
node, making it a leaf node .
• The available data is divided into three parts:
training examples , validation examples for pruning , a set of test
examples for estimate of accuracy over future unseen examples
• Next step calculating new tree accuracy
Reduced Error Pruning
• The problems related to the method of reduced error pruning are
basically two:
1. The use of a pruning set distinct from the training set is inadequate
when a small number of observations are available
2. the parts of the original tree that correspond to special cases outside
the test set may be lost after pruning. Therefore trees pruned via that
pruning method may fail in correctly classifying exceptional cases.
Reduced Error Pruning
• Pros :
linear computational complexity
• Cons :
over pruning , if the test set is much smaller than the training set
Pessimistic error pruning
• This pruning method, proposed by Quinlan
• it avoids using an independent pruning set
• The misclassification rate, estimated by means of the training set
• If results to be optimistic , thus it always happens that pruning
operations based on the same data set produce trees that are larger
than it is needed .
Pessimistic error pruning
• Pros :
This method has a linear complexity in the number of leaves
• Cons:
the worst case is that in which the tree has not to be pruned at all
Error complexity pruning
• It finds error complexity at each node .
1. Calculating the error cost of the node
2. r(t) is error rate of a node
3. p(t) is probability of occurrence of a node
4. If node t was not pruned then error
cost of subtree T , rooted at t:
5. Then , The error complexity of the node
Error complexity pruning
• The method consists of following steps:
1. a ( error complexity ) is computed for each node.
2. the minimum a node is pruned.
3. the above is repeated and a forest of pruned tree is formed.
4. the tree with best accuracy is selected.
Minimum Error pruning
• This method was developed by Niblett and Brotko
• bottom-up approach
• seeks a single tree that minimizes the expected error rate on an
independent data set
• If it is predicted that all future examples will be in class c, the
following equation is used to predict the expected error rate of
pruning at node t :
1. K : # of class
2. nt : # of examples in node t
3. nt,c : # of examples assigned to class c in node t
Minimum Error pruning
• The method consists of following steps
1. At each non leaf node in the tree, calculate expected error rate if
that subtree is pruned.
2. Calculate the expected error rate for that node if subtree is not
pruned.
3. If pruning the node leads to greater expected error rate, then keep
the subtree; otherwise, prune it .
Cost based pruning
• In this method not only an error rate is considered at each node but
also a cost is considered .
• That is for pruning decision tree error rate and cost of deciding
selection of one or more class-label attribute is considered .
Iterative growing and pruning algorithm
• Gelfand et al
• These goals are reached by splitting the data set into two subsets
• then by repeatedly growing and pruning a tree on different subsets
• a tree is grown by using the first subset
• then it is pruned by using the second subset
Tanks for reading

More Related Content

PPTX
Random forest
PPTX
Issues in knowledge representation
PPTX
Ensemble learning
DOC
Dbms lab questions
PPTX
Metrics for project size estimation
PPTX
Tree pruning
PPTX
Classification and Regression
PPTX
Overfitting & Underfitting
Random forest
Issues in knowledge representation
Ensemble learning
Dbms lab questions
Metrics for project size estimation
Tree pruning
Classification and Regression
Overfitting & Underfitting

What's hot (20)

PPTX
Lecture 6: Ensemble Methods
PPTX
Neural network
PDF
Artificial Intelligence - Hill climbing.
PPTX
Decision Trees
PPTX
ML - Multiple Linear Regression
PDF
Decision trees in Machine Learning
PDF
PAC Learning
PPTX
Machine Learning: Bias and Variance Trade-off
PDF
K - Nearest neighbor ( KNN )
PPT
Back propagation
PPTX
K-Nearest Neighbor Classifier
PPTX
Performance analysis(Time & Space Complexity)
PDF
Machine learning Summer Training report
PPTX
OLAP operations
ODP
Machine Learning With Logistic Regression
PDF
I. AO* SEARCH ALGORITHM
PDF
Vc dimension in Machine Learning
PPTX
Dijkstra's algorithm presentation
PDF
Understanding Bagging and Boosting
Lecture 6: Ensemble Methods
Neural network
Artificial Intelligence - Hill climbing.
Decision Trees
ML - Multiple Linear Regression
Decision trees in Machine Learning
PAC Learning
Machine Learning: Bias and Variance Trade-off
K - Nearest neighbor ( KNN )
Back propagation
K-Nearest Neighbor Classifier
Performance analysis(Time & Space Complexity)
Machine learning Summer Training report
OLAP operations
Machine Learning With Logistic Regression
I. AO* SEARCH ALGORITHM
Vc dimension in Machine Learning
Dijkstra's algorithm presentation
Understanding Bagging and Boosting
Ad

Viewers also liked (20)

PPT
Jose_TH1_T09_5.ppt
PDF
Bonsai Networking: pruning your professional learning network (VU Seminar)
PPT
Hyperspectral unmixing using novel conversion model.ppt
PPT
Subspace_Discriminant_Approach_Hyperspectral.ppt
PDF
MCC_PhDDefense
PDF
Signal Processing Course : Compressed Sensing
PPTX
B.sc. agri i po h unit 3 method of pruning, growth regulators and nursery
PDF
Learning Sparse Representation
PPT
Components Of Executive Information System
PDF
Trade finance post and pre shipment
DOC
Executive information system
PDF
C. executive information systems
PPTX
Executive information system (EIS)
PPTX
Spandana image processing and compression techniques (7840228)
PPT
Pre & post_shipment_finance[1]
PPT
Executive Information System
PPT
Competitive advantage
PDF
PPTX
Distintas ciudades
PPTX
Matemática y tic
Jose_TH1_T09_5.ppt
Bonsai Networking: pruning your professional learning network (VU Seminar)
Hyperspectral unmixing using novel conversion model.ppt
Subspace_Discriminant_Approach_Hyperspectral.ppt
MCC_PhDDefense
Signal Processing Course : Compressed Sensing
B.sc. agri i po h unit 3 method of pruning, growth regulators and nursery
Learning Sparse Representation
Components Of Executive Information System
Trade finance post and pre shipment
Executive information system
C. executive information systems
Executive information system (EIS)
Spandana image processing and compression techniques (7840228)
Pre & post_shipment_finance[1]
Executive Information System
Competitive advantage
Distintas ciudades
Matemática y tic
Ad

Similar to Post pruning (20)

PPTX
random forest.pptx
PPTX
Random forest Algorithm using machine learnig.pptx
PDF
Lecture 9 - Decision Trees and Ensemble Methods, a lecture in subject module ...
PPTX
Random Forest Decision Tree.pptx
PPTX
23-512(Decision Tree) machine learning ppt
PDF
Decision Tree in Machine Learning
PPTX
Supervised Learning Algorithm Slide.pptx
PDF
Machine Learning Algorithm - Decision Trees
PPTX
Machine learning tree models for classification
PPTX
Feature selection with imbalanced data in agriculture
PDF
Decision tree
PPTX
Decision Tree.pptx
PPTX
Decision_Tree_Presentation_with_indepth_analysis.pptx
PPTX
Macine learning algorithms - K means, KNN
PPTX
Decision-trees basic decryptions DT .pptx
PDF
From decision trees to random forests
PPTX
Unit 2-ML.pptx
PPT
Lecture4.ppt
PPTX
Comparitive Analysis .pptx Footprinting, Enumeration, Scanning, Sniffing, Soc...
PPTX
Decision Tree.pptx
random forest.pptx
Random forest Algorithm using machine learnig.pptx
Lecture 9 - Decision Trees and Ensemble Methods, a lecture in subject module ...
Random Forest Decision Tree.pptx
23-512(Decision Tree) machine learning ppt
Decision Tree in Machine Learning
Supervised Learning Algorithm Slide.pptx
Machine Learning Algorithm - Decision Trees
Machine learning tree models for classification
Feature selection with imbalanced data in agriculture
Decision tree
Decision Tree.pptx
Decision_Tree_Presentation_with_indepth_analysis.pptx
Macine learning algorithms - K means, KNN
Decision-trees basic decryptions DT .pptx
From decision trees to random forests
Unit 2-ML.pptx
Lecture4.ppt
Comparitive Analysis .pptx Footprinting, Enumeration, Scanning, Sniffing, Soc...
Decision Tree.pptx

Recently uploaded (20)

PDF
What if we spent less time fighting change, and more time building what’s rig...
PDF
Complications of Minimal Access Surgery at WLH
PDF
Chinmaya Tiranga quiz Grand Finale.pdf
PDF
Classroom Observation Tools for Teachers
PPTX
CHAPTER IV. MAN AND BIOSPHERE AND ITS TOTALITY.pptx
PDF
1_English_Language_Set_2.pdf probationary
PPTX
Tissue processing ( HISTOPATHOLOGICAL TECHNIQUE
PDF
Weekly quiz Compilation Jan -July 25.pdf
PDF
احياء السادس العلمي - الفصل الثالث (التكاثر) منهج متميزين/كلية بغداد/موهوبين
PDF
SOIL: Factor, Horizon, Process, Classification, Degradation, Conservation
PDF
medical_surgical_nursing_10th_edition_ignatavicius_TEST_BANK_pdf.pdf
PPTX
Lesson notes of climatology university.
PPTX
History, Philosophy and sociology of education (1).pptx
PDF
Computing-Curriculum for Schools in Ghana
PDF
Practical Manual AGRO-233 Principles and Practices of Natural Farming
PDF
Paper A Mock Exam 9_ Attempt review.pdf.
PDF
IGGE1 Understanding the Self1234567891011
PPTX
Radiologic_Anatomy_of_the_Brachial_plexus [final].pptx
PPTX
Cell Types and Its function , kingdom of life
PPTX
UV-Visible spectroscopy..pptx UV-Visible Spectroscopy – Electronic Transition...
What if we spent less time fighting change, and more time building what’s rig...
Complications of Minimal Access Surgery at WLH
Chinmaya Tiranga quiz Grand Finale.pdf
Classroom Observation Tools for Teachers
CHAPTER IV. MAN AND BIOSPHERE AND ITS TOTALITY.pptx
1_English_Language_Set_2.pdf probationary
Tissue processing ( HISTOPATHOLOGICAL TECHNIQUE
Weekly quiz Compilation Jan -July 25.pdf
احياء السادس العلمي - الفصل الثالث (التكاثر) منهج متميزين/كلية بغداد/موهوبين
SOIL: Factor, Horizon, Process, Classification, Degradation, Conservation
medical_surgical_nursing_10th_edition_ignatavicius_TEST_BANK_pdf.pdf
Lesson notes of climatology university.
History, Philosophy and sociology of education (1).pptx
Computing-Curriculum for Schools in Ghana
Practical Manual AGRO-233 Principles and Practices of Natural Farming
Paper A Mock Exam 9_ Attempt review.pdf.
IGGE1 Understanding the Self1234567891011
Radiologic_Anatomy_of_the_Brachial_plexus [final].pptx
Cell Types and Its function , kingdom of life
UV-Visible spectroscopy..pptx UV-Visible Spectroscopy – Electronic Transition...

Post pruning

  • 1. Post pruning methods Mojtaba amiri M.S of artificial intelligence
  • 2. Approaches • top-down approach : proceed from the root towards the leaves of tree • bottom-up approach : starting the analysis from the leaves and finishing with the root node
  • 3. Based methods 1. Reduced Error Pruning 2. Pessimistic error pruning 3. Error complexity pruning 4. Minimum Error pruning 5. Cost based pruning 6. Iterative growing and pruning algorithm
  • 4. Reduced Error Pruning • This method was proposed by Quinlan . • This method considers each of the decision nodes in the tree to be candidates for pruning, consist of removing the subtree rooted at that node, making it a leaf node . • The available data is divided into three parts: training examples , validation examples for pruning , a set of test examples for estimate of accuracy over future unseen examples • Next step calculating new tree accuracy
  • 5. Reduced Error Pruning • The problems related to the method of reduced error pruning are basically two: 1. The use of a pruning set distinct from the training set is inadequate when a small number of observations are available 2. the parts of the original tree that correspond to special cases outside the test set may be lost after pruning. Therefore trees pruned via that pruning method may fail in correctly classifying exceptional cases.
  • 6. Reduced Error Pruning • Pros : linear computational complexity • Cons : over pruning , if the test set is much smaller than the training set
  • 7. Pessimistic error pruning • This pruning method, proposed by Quinlan • it avoids using an independent pruning set • The misclassification rate, estimated by means of the training set • If results to be optimistic , thus it always happens that pruning operations based on the same data set produce trees that are larger than it is needed .
  • 8. Pessimistic error pruning • Pros : This method has a linear complexity in the number of leaves • Cons: the worst case is that in which the tree has not to be pruned at all
  • 9. Error complexity pruning • It finds error complexity at each node . 1. Calculating the error cost of the node 2. r(t) is error rate of a node 3. p(t) is probability of occurrence of a node 4. If node t was not pruned then error cost of subtree T , rooted at t: 5. Then , The error complexity of the node
  • 10. Error complexity pruning • The method consists of following steps: 1. a ( error complexity ) is computed for each node. 2. the minimum a node is pruned. 3. the above is repeated and a forest of pruned tree is formed. 4. the tree with best accuracy is selected.
  • 11. Minimum Error pruning • This method was developed by Niblett and Brotko • bottom-up approach • seeks a single tree that minimizes the expected error rate on an independent data set • If it is predicted that all future examples will be in class c, the following equation is used to predict the expected error rate of pruning at node t : 1. K : # of class 2. nt : # of examples in node t 3. nt,c : # of examples assigned to class c in node t
  • 12. Minimum Error pruning • The method consists of following steps 1. At each non leaf node in the tree, calculate expected error rate if that subtree is pruned. 2. Calculate the expected error rate for that node if subtree is not pruned. 3. If pruning the node leads to greater expected error rate, then keep the subtree; otherwise, prune it .
  • 13. Cost based pruning • In this method not only an error rate is considered at each node but also a cost is considered . • That is for pruning decision tree error rate and cost of deciding selection of one or more class-label attribute is considered .
  • 14. Iterative growing and pruning algorithm • Gelfand et al • These goals are reached by splitting the data set into two subsets • then by repeatedly growing and pruning a tree on different subsets • a tree is grown by using the first subset • then it is pruned by using the second subset