Voting-Based Learning
   Classifier System
for multi-label classification
   Kaveh Ahmadi-Abhari (Presenter)
             Ali Hamzeh
           Sattar Hashemi

  IWLCS 2011 – Dublin, Ireland, 13th July 2011
Multi-label Classification

   Single Label Classification
      Exclusive classes: each
       example belongs to one
       class

   Multi-label Classification
     Each instance can belong
      to more than one class




Kaveh Ahmadi-Abhari               2   Shiraz University, Soft Computing Group
Multi-label Classification
                                                           Sky
                                      People
   Single Label Classification
      Exclusive classes: each
       example belongs to one
       class

   Multi-label Classification
     Each instance can belong
      to more than one class

                                         Sand




Kaveh Ahmadi-Abhari               3       Shiraz University, Soft Computing Group
Current Methods


                    Problem        • Transfer problem to a single-
                 Transformation      label classification problem




                      Algorithm    • Adapt single-label classifiers
                      Adaptation     to Solve the problem




                                                          [Tsoumakas & Katakis, 2007]
Kaveh Ahmadi-Abhari                     4                    Shiraz University, Soft Computing Group
Problem Transformation Approaches
                                            Ex.   Label- set
                                            1a      λ1
                  Copy Transformation       1b      λ4
                                            2a       λ3
                                            2b       λ4
                                            3        λ1
                                            4a       λ2
       Ex.             Label- set           4b       λ3
                                            4c       λ4
        1               {λ1 , λ4 }
        2               {λ3 , λ4 }
        3                {λ1}
        4             {λ2 , λ3 , λ4 }
                                                      [Tsoumakas et al., 2009]
Kaveh Ahmadi-Abhari                     5             Shiraz University, Soft Computing Group
Algorithm Adaptation Approaches
   Multi-label lazy algorithm
         ML-kNN [Zhang & Zhou, PRJ07]
   Multi-label decision trees
         ADTBoost.MH [DeComité et al. MLDM03]
         Multi-Label C4.5 [Clare & King, LNCS2168]
   Multi-label kernel methods
         Rank-SVM [Elisseeff & Weston, NIPS02]
         ML-SVM [M.R. Boutell, et al. PR04]
   Multi-label text categorization algorithms
         BoosTexter [Schapire & Singer, MLJ00]
         Maximal Margin Labeling [Kazawa et al., NIPS04]
         Probabilistic generative models [McCallum, AAAI99] [Ueda & Saito, NIPS03]
         BP-MLL [Zhang & Zhou, TKDE06]



Kaveh Ahmadi-Abhari                          6                    Shiraz University, Soft Computing Group
Motivation

       A lot has been done in terms of classifications
       using LCSs

       Most of these studies have been conducted for
       single-label classification problems

       Multi-label classification is in its inception [Vallim
       et al., IWLCS 08]




Kaveh Ahmadi-Abhari               7               Shiraz University, Soft Computing Group
Voting Based Learning Classifier System


                      How can we guide the discovery mechanism
                      (e.g. evolutionary operators) in LCSs?




Kaveh Ahmadi-Abhari                  8               Shiraz University, Soft Computing Group
Voting Based Learning Classifier System


                      How can we guide the discovery mechanism
                      (e.g. evolutionary operators) in LCSs?



                                  Using the prior knowledge gained from
                                  past experiences




Kaveh Ahmadi-Abhari                  9                  Shiraz University, Soft Computing Group
Voting Based Learning Classifier System


                        How can we guide the discovery mechanism
                        (e.g. evolutionary operators) in LCSs?



                                         Using the prior knowledge gained from
                                         past experiences


    Training instances vote their matched rules
    according to how correct the rule is




Kaveh Ahmadi-Abhari                          10                Shiraz University, Soft Computing Group
Voting Based Learning Classifier System


                        How can we guide the discovery mechanism
                        (e.g. evolutionary operators) in LCSs?



                                         Using the prior knowledge gained from
                                         past experiences


    Training instances vote their matched rules
    according to how correct the rule is



                               Fitness measure

Kaveh Ahmadi-Abhari                          11                Shiraz University, Soft Computing Group
Voting Defining Rule Types


                          How can the given votes describe the
                          quality of the rules accurately?




           Define different types for the rules such that each of these types
           describes the quality status the rule might have.




Kaveh Ahmadi-Abhari                        12                   Shiraz University, Soft Computing Group
Rule Types

           Example:
                  in a single-label classification problem, rule types
                  might be correct or wrong.




            Each rule might receive a “correct” or “wrong” vote from each
            matched training instance.

            A rule receives a combination of “correct” and “wrong” votes from its
            matched training instances




Kaveh Ahmadi-Abhari                          13                     Shiraz University, Soft Computing Group
Votes as Fitness Measure


                      • Given votes
                        • Describe the quality of the rules
                        • Use as a fitness measure for
                          guiding the discovery mechanism.


                      • For example, a rule with more “wrong”
                        votes, should be discovered with a high
                        probability to achieve a meaningful rule




Kaveh Ahmadi-Abhari                14                  Shiraz University, Soft Computing Group
Rules Definition


                         Antecedent / Consequent
                               ###1 / 110
                               0011 / 001




   Antecedent part matches with the feature vector.
   Consequent part are the classes predicted by the rule.
   One bit for each class in the consequent part.
         Value 1 in the bit indicates existence of the respective class.

Kaveh Ahmadi-Abhari                       15                    Shiraz University, Soft Computing Group
VLCS Vote Types for Multi-label Problem


                                    Correct



                      Wrong                         Subset
                               Multi-label
                              Vote Types for
                                  VLCS


                          Partial             Superset



Kaveh Ahmadi-Abhari                   16                 Shiraz University, Soft Computing Group
Multi-Label Simple Dataset

                                                  000
                              111
                                                               001
                                                  1, 4
                      110           1, 3

                                                                     010
                                                        2, 4
                                     1, 2
                        101

                                                            011
                                       100

                                                        Expand from [Vallim et al., GECCO’ 08]
Kaveh Ahmadi-Abhari                          17                      Shiraz University, Soft Computing Group
VLCS Voting Options for Multi-label Problem

   Correct Rules (C)                                       111
                                                                          000
                                                                                       001
                                                                          1, 4
                                                      110
                                                              1, 3
                      00# /1001                                              2, 4          010
                                                               1, 2
                                                       101
            • Is correct when it matches with:                                     011
              • 000 or                                             100
              • 001




Kaveh Ahmadi-Abhari                              18           Shiraz University, Soft Computing Group
VLCS Voting Options for Multi-label Problem

   Wrong Rules (W)                                       111
                                                                        000
                                                                                     001
                                                                        1, 4
                                                    110
                                                            1, 3
                      0#0/0010                                             2, 4          010
                                                             1, 2
                                                     101
            • Is wrong when it matches with:                                     011
              • 000 or                                           100
              • 010




Kaveh Ahmadi-Abhari                            19           Shiraz University, Soft Computing Group
VLCS Voting Options for Multi-label Problem

   Subset Rules                                           111
                                                                         000
                                                                                      001
                                                                         1, 4
                                                     110
                                                             1, 3
                      #01/1000                                              2, 4          010
                                                              1, 2
                                                      101
            • Is subset when it matches with:                                     011
              • 001 or                                            100
              • 101




Kaveh Ahmadi-Abhari                             20           Shiraz University, Soft Computing Group
VLCS Voting Options for Multi-label Problem

   Subset Rules                                           111
                                                                         000
                                                                                      001
                                                                         1, 4
                                                     110
                                                             1, 3
                      #01/1000                                              2, 4          010
                                                              1, 2
                                                      101
            • Is subset when it matches with:                                     011
              • 001 or                                            100
              • 101

                        Excepted Classes:

                              1, 4

Kaveh Ahmadi-Abhari                             21           Shiraz University, Soft Computing Group
VLCS Voting Options for Multi-label Problem

   Superset Rules                                       111
                                                                       000
                                                                                    001
                                                                       1, 4
                                                   110
                                                           1, 3
                      #00/1101                                            2, 4          010
                                                            1, 2
                                                    101
            • Is superset when it matches with:                                 011
              • 001 or                                          100
              • 101




Kaveh Ahmadi-Abhari                           22           Shiraz University, Soft Computing Group
VLCS Voting Options for Multi-label Problem

   Superset Rules                                       111
                                                                       000
                                                                                    001
                                                                       1, 4
                                                   110
                                                           1, 3
                      #00/1101                                            2, 4          010
                                                            1, 2
                                                    101
            • Is superset when it matches with:                                 011
              • 001 or                                          100
              • 101

                        Excepted Classes:

                              1, 4

Kaveh Ahmadi-Abhari                           23           Shiraz University, Soft Computing Group
VLCS Voting Options for Multi-label Problem

   Partial-set Rules                                    111
                                                                       000
                                                                                    001
                                                                       1, 4
                                                   110
                                                           1, 3
                      #1# / 0110                                          2, 4          010
                                                            1, 2
                                                    101
            • Is superset when it matches with:                                 011
              • 010 or                                          100
              • 111




Kaveh Ahmadi-Abhari                           24           Shiraz University, Soft Computing Group
VLCS Voting Options for Multi-label Problem

   Partial-set Rules                                    111
                                                                       000
                                                                                    001
                                                                       1, 4
                                                   110
                                                           1, 3
                      #1# / 0110                                          2, 4          010
                                                            1, 2
                                                    101
            • Is superset when it matches with:                                 011
              • 010 or                                          100
              • 111

                        Excepted Classes:

                              2, 4

Kaveh Ahmadi-Abhari                           25           Shiraz University, Soft Computing Group
VLCS Voting Options for Multi-label Problem

                                                              000
   Rules might receive different votes         111
                                                                           001
    during the time                                           1, 4
                                          110
                                                  1, 3

                                                                 2, 4          010
                                                   1, 2
                      #0# / 1001           101
                                                                       011
                                                       100




Kaveh Ahmadi-Abhari                26             Shiraz University, Soft Computing Group
VLCS Voting Options for Multi-label Problem

                                                              000
   Rules might receive different votes         111
                                                                           001
    during the time                                           1, 4
                                          110
                                                  1, 3

                                                                 2, 4          010
                                                   1, 2
                       #0# / 1001          101
                                                                       011
                                                       100


      Is correct for
      instance 000




Kaveh Ahmadi-Abhari                 27            Shiraz University, Soft Computing Group
VLCS Voting Options for Multi-label Problem

                                                                   000
   Rules might receive different votes              111
                                                                                001
    during the time                                                1, 4
                                               110
                                                       1, 3

                                                                      2, 4          010
                                                        1, 2
                       #0# / 1001               101
                                                                            011
                                                            100


      Is correct for          Is partial-set
      instance 000            for instance
                                   101




Kaveh Ahmadi-Abhari                  28                Shiraz University, Soft Computing Group
Using Stored Prior Knowledge
   Consider a rule that all received votes
   are superset                                            }     Information




                                                                        }
         The rule is covering an appropriate area
         of the problem




                                                                                         Inference
              The rule is predicting greater number
              of classes for the matched input
              instance

                      The number of the classes the rule
                      predicts should be subtracted


Kaveh Ahmadi-Abhari                        29              Shiraz University, Soft Computing Group
Discovery Operators

   In the discovery mechanism an evolutionary algorithm with
    four mutation operators is defined:




Kaveh Ahmadi-Abhari             30              Shiraz University, Soft Computing Group
Discovery Operators
   Mutation operators on rule’s antecedent part



                      Generalize the rule by flipping the 0
              MA-G    or 1 bits to #



                      Specializes the rule by flipping #
               MA-S   bits to 1 or 0




Kaveh Ahmadi-Abhari                31                Shiraz University, Soft Computing Group
Discovery Operators
   Mutation operators on rule’s consequent part



                      Subtract the number of predicted
               MC-S   classes by flipping 1 bits to 0



                      Adds more classes to predicted
               MC-A   classes by flipping 0 bits to 1




Kaveh Ahmadi-Abhari               32               Shiraz University, Soft Computing Group
Which Discovery Operator?

          The votes each rule has received guide which mutation
          operator should act.




Kaveh Ahmadi-Abhari                33               Shiraz University, Soft Computing Group
Which Discovery Operator?

          The votes each rule has received guide which mutation
          operator should act.




                                 Wrongly            Subtract the
                              assigned some          number of
          Superset Rule
                              non-expected            predicted
                                  classes          classes (MC-S)




Kaveh Ahmadi-Abhari                34               Shiraz University, Soft Computing Group
Which Discovery Operator?
                                                    Activated Mutation
                       Rule Received Votes
                                                         Operator
                              Correct                     MA-G
                              Subset                      MC-A
                             Superset                     MC-S
                            Partial-Set                MC-A, MC-S
                              Wrong                    MC-A, MC-S
                         Correct, Subset                  MA-S
                        Correct, Superset                 MA-G
                        Correct, Partial-Set              MA-S
                         Correct, Wrong                   MA-S
                          Wrong, Subset                MA-S, MC-A
                          Wrong, Partial                  MA-S

                      Correct, Subset, Wrong           MA-S, MA-G

Kaveh Ahmadi-Abhari                            35                    Shiraz University, Soft Computing Group
Mutation Rate
                      • Mutation operator performs bit flipping
                        using a probability, which is the mutation
                        rate.

                      • The strength of a rule is the amount of
                        reward we predict the system to receive if
                        the rule acts.


                      • The more the strength, the less the mutation
                        rate.




Kaveh Ahmadi-Abhari                   36                  Shiraz University, Soft Computing Group
Strength of a Rule
   The mean of the rewards the rule gets over time.



              Reward Function:

                                     C rule ∆C expected
                            R = 1−
                                     C rule  C expected




                                                 Alteration of [Vallim et al., GECCO’ 08]
Kaveh Ahmadi-Abhari                     37                     Shiraz University, Soft Computing Group
Strength of a Rule
   The mean of the rewards the rule gets over time.



              Reward Function:

                                      C rule ∆C expected
                            R = 1−
                                     C rule  C expected


                      A ∆B
                         =       {x : ( x ∈ A ) ⊕ ( x ∈ B )}

                                                 Alteration of [Vallim et al., GECCO’ 08]
Kaveh Ahmadi-Abhari                      38                    Shiraz University, Soft Computing Group
Rules Rewards


                  Input    Expected   Selected     Received
                                                                  Reward
                Instance    output      Rule         Vote
                  0001       1, 2     ###1 / 110    Correct            1
                  0101      1, 2, 3   ###1 / 110    Subset           0.66
                  0111        1       ###1 / 110   Superset          0.50
                  1111       1,3      ###1 / 110   Partial-set       0.33
                  0011        3       ###1 / 110    Wrong              0




Kaveh Ahmadi-Abhari                       39                     Shiraz University, Soft Computing Group
Experimental Results
   Data Sets:
         Two binary datasets in the bioinformatics domain
              [Chan and Freitas, GECCO’ 06 ]
              Extracted from [Alves et al., 2009]




Kaveh Ahmadi-Abhari                  40              Shiraz University, Soft Computing Group
Experimental Results
   Quality Metrics:

                 Accuracy

                • Proportion of predicted classes among all predicted or
                  true classes

                 Precision

                • Proportion of true classes among all predicted classes

                 Recall

                • Proportion of predicted classes among all true classes


                                                              [Tsoumakas & Katakis, 2007]
Kaveh Ahmadi-Abhari                         41                    Shiraz University, Soft Computing Group
Experimental Results
   For the VLCS, we use a 5-fold cross validation in which the
    training part is used to evaluate the rules using the voting
    mechanism described above.
   Fixed size population
         initially are the most general possible rules.
   In each generation, each rule is voted by its matched
    instances
         reward is assigned
   Defined mutation operators to discover new rules
   The combination of the best rules among the parents and the
    off-springs make the next generation.
   We stop the training phase if the mean strength of the rules
    decreases in a number of consecutive generations.
Kaveh Ahmadi-Abhari                       42               Shiraz University, Soft Computing Group
Experimental Results
   [Chan and Freitas, GECCO’ 06 ]
         135 instances
         152 attributes
         Two classes
             • Each instance could have one or both of the available class labels.


                 Method      Accuracy       Precision        Recall

                      BR       0.89            0.89           0.87

                 ML-KNN        0.91            0.93           0.91

                      VLCS     0.89            0.89           0.89



Kaveh Ahmadi-Abhari                       43                   Shiraz University, Soft Computing Group
Experimental Results
   Extracted from [Alves et al., 2009]
         7877 proteins
         40 attributes
         Six classes
             • Each instance could have some of the available class labels.


                 Method      Accuracy       Precision       Recall

                      BR       0.78            0.77          0.78

                 ML-KNN        0.80            0.81          0.80

                      VLCS     0.81            0.83          0.82



Kaveh Ahmadi-Abhari                       44                   Shiraz University, Soft Computing Group
Conclusion




                      Guiding the discovery mechanism
                      with a prior knowledge, such that is
                      used in VLCS, can help us solve
                      applicable problems




Kaveh Ahmadi-Abhari                     45               Shiraz University, Soft Computing Group
Future Work
   A representation for dealing with numeric and nominal
    datasets.
   Future studies on scalability and stability of the system is
    necessary.
   Additional studies on system performance in dealing with
    imbalanced data and noise is also required.
   Improving evolutionary operators, guiding mechanism and
    rule refinement.




Kaveh Ahmadi-Abhari               46               Shiraz University, Soft Computing Group
Any Question?




                      The most exciting phrase to hear in
                      science, the one that heralds new
                      discoveries is not “Eureka”! (I found
                      it!) but “That's funny...”
                                               - Isaac Asimov




Kaveh Ahmadi-Abhari               47                Shiraz University, Soft Computing Group

More Related Content

PDF
Information For Candidates Booklet
PDF
Vietnam Exam Ielts Information To Candidates 230910
PPTX
MrKNN_Soft Relevance for Multi-label Classification
PDF
Tag Extraction Final Presentation - CS185CSpring2014
PDF
Multi-label, Multi-class Classification Using Polylingual Embeddings
PDF
Multi-label Classification with Meta-labels
PPTX
Multi-Class Classification on Cartographic Data(Forest Cover)
PDF
Svm implementation for Health Data
Information For Candidates Booklet
Vietnam Exam Ielts Information To Candidates 230910
MrKNN_Soft Relevance for Multi-label Classification
Tag Extraction Final Presentation - CS185CSpring2014
Multi-label, Multi-class Classification Using Polylingual Embeddings
Multi-label Classification with Meta-labels
Multi-Class Classification on Cartographic Data(Forest Cover)
Svm implementation for Health Data

Viewers also liked (15)

PDF
CNN-RNN: A Unified Framework for Multi-label Image Classification@CV勉強会35回CVP...
PPTX
Presentation of Alaa Abi Haidar at the BnF Information Day
PPTX
2013 KDD conference presentation--"Multi-Label Relational Neighbor Classifica...
PPTX
Naïve multi label classification of you tube comments using
PPTX
Text extraction using document structure features and support vector machines
PDF
Analyse de données fonctionnelles par Machines à Vecteurs de Support (SVM)
PDF
Théorie de l’apprentissage et SVM : présentation rapide et premières idées da...
PPT
Support Vector machine
PPTX
Decision trees
PPTX
Decision Tree Analysis
PDF
Decision tree example problem
PDF
(Deep) Neural Networks在 NLP 和 Text Mining 总结
PDF
Decision tree
PDF
2013-1 Machine Learning Lecture 05 - Andrew Moore - Support Vector Machines
CNN-RNN: A Unified Framework for Multi-label Image Classification@CV勉強会35回CVP...
Presentation of Alaa Abi Haidar at the BnF Information Day
2013 KDD conference presentation--"Multi-Label Relational Neighbor Classifica...
Naïve multi label classification of you tube comments using
Text extraction using document structure features and support vector machines
Analyse de données fonctionnelles par Machines à Vecteurs de Support (SVM)
Théorie de l’apprentissage et SVM : présentation rapide et premières idées da...
Support Vector machine
Decision trees
Decision Tree Analysis
Decision tree example problem
(Deep) Neural Networks在 NLP 和 Text Mining 总结
Decision tree
2013-1 Machine Learning Lecture 05 - Andrew Moore - Support Vector Machines
Ad

Similar to Voting Based Learning Classifier System for Multi-Label Classification (20)

PPT
Lazy Association Classification
PPTX
Edura_Presentation.pptx contains information about education roadmap
PDF
IRJET - A Survey on Machine Learning Algorithms, Techniques and Applications
PDF
CSEDU 2012: Source code validation and plagiarism detection: technology-rich ...
PPTX
Classification
PDF
Novel Class Detection Using RBF SVM Kernel from Feature Evolving Data Streams
PPTX
Edu Tools2
PDF
A New Active Learning Technique Using Furthest Nearest Neighbour Criterion fo...
PPTX
Java ppt
PDF
IEEE Final Year Projects 2011-2012 :: Elysium Technologies Pvt Ltd::Knowledge...
PDF
A Coursework Support System for Offering Challenges and Assistance by Analyzi...
PDF
IRJET- Analysis of Question and Answering Recommendation System
PDF
Active learning for ranking through expected loss optimization
PDF
Are evolutionary rule learning algorithms appropriate for malware detection
PDF
1111111111111111111111111111111111111.pdf
PPSX
Chapter3-evaluation techniques HCI
PDF
Answer Selection and Validation for Arabic Questions
PDF
Regularized Weighted Ensemble of Deep Classifiers
PPTX
NSL KDD Cup 99 dataset Anomaly Detection using Machine Learning Technique
PDF
QER : query entity recognition
Lazy Association Classification
Edura_Presentation.pptx contains information about education roadmap
IRJET - A Survey on Machine Learning Algorithms, Techniques and Applications
CSEDU 2012: Source code validation and plagiarism detection: technology-rich ...
Classification
Novel Class Detection Using RBF SVM Kernel from Feature Evolving Data Streams
Edu Tools2
A New Active Learning Technique Using Furthest Nearest Neighbour Criterion fo...
Java ppt
IEEE Final Year Projects 2011-2012 :: Elysium Technologies Pvt Ltd::Knowledge...
A Coursework Support System for Offering Challenges and Assistance by Analyzi...
IRJET- Analysis of Question and Answering Recommendation System
Active learning for ranking through expected loss optimization
Are evolutionary rule learning algorithms appropriate for malware detection
1111111111111111111111111111111111111.pdf
Chapter3-evaluation techniques HCI
Answer Selection and Validation for Arabic Questions
Regularized Weighted Ensemble of Deep Classifiers
NSL KDD Cup 99 dataset Anomaly Detection using Machine Learning Technique
QER : query entity recognition
Ad

More from Daniele Loiacono (20)

PDF
GPUs for GEC Competition @ GECCO-2013
PDF
EvoRobocode Competition @ GECCO-2013
PDF
2013 Simulated Car Racing @ GECCO-2013
PDF
2012 Simulated Car Racing Championship @ CIG-2012
PDF
2012 Simulated Car Racing Championship @ GECCO-2012
PDF
2012 Simulated Car Racing Championship @ Evo*-2012
PDF
Computational Intelligence in Games Tutorial @GECCO2012
PPTX
XCSF with Local Deletion: Preventing Detrimental Forgetting
PDF
Testing learning classifier systems
PDF
Random Artificial Incorporation of Noise in a Learning Classifier System Envi...
PPTX
One Step Fits All
PDF
Introducing LCS to Digital Design Verification
PPTX
A temporal classifier system using spiking neural networks
PDF
Confusion Matrices for Improving Performance of Feature Pattern Classifier Sy...
PDF
Automatically Defined Functions for Learning Classifier Systems
PDF
2011 Simulated Car Racing Championship @ GECCO-2011
PDF
2010 Simulated Car Racing Championship @ CIG-2010
PDF
2010 Simulated Car Racing Championship @ GECCO-2010
PDF
2010 Simulated Car Racing Championship @ WCCI-2010
PDF
Car Setup Optimization Competition @ EvoStar 2010
GPUs for GEC Competition @ GECCO-2013
EvoRobocode Competition @ GECCO-2013
2013 Simulated Car Racing @ GECCO-2013
2012 Simulated Car Racing Championship @ CIG-2012
2012 Simulated Car Racing Championship @ GECCO-2012
2012 Simulated Car Racing Championship @ Evo*-2012
Computational Intelligence in Games Tutorial @GECCO2012
XCSF with Local Deletion: Preventing Detrimental Forgetting
Testing learning classifier systems
Random Artificial Incorporation of Noise in a Learning Classifier System Envi...
One Step Fits All
Introducing LCS to Digital Design Verification
A temporal classifier system using spiking neural networks
Confusion Matrices for Improving Performance of Feature Pattern Classifier Sy...
Automatically Defined Functions for Learning Classifier Systems
2011 Simulated Car Racing Championship @ GECCO-2011
2010 Simulated Car Racing Championship @ CIG-2010
2010 Simulated Car Racing Championship @ GECCO-2010
2010 Simulated Car Racing Championship @ WCCI-2010
Car Setup Optimization Competition @ EvoStar 2010

Recently uploaded (20)

PPTX
TRAINNING, DEVELOPMENT AND APPRAISAL.pptx
PDF
#1 Safe and Secure Verified Cash App Accounts for Purchase.pdf
DOCX
Hand book of Entrepreneurship 4 Chapters.docx
PPTX
33ABJFA6556B1ZP researhchzfrsdfasdfsadzd
PPTX
operations management : demand supply ch
PDF
Middle East's Most Impactful Business Leaders to Follow in 2025
PPTX
CTG - Business Update 2Q2025 & 6M2025.pptx
PDF
MBA2024 CGE 1.pdf file presentation 2025
PDF
Second Hand Fashion Call to Action March 2025
PPTX
interschool scomp.pptxzdkjhdjvdjvdjdhjhieij
PDF
Kishore Vora - Best CFO in India to watch in 2025.pdf
PDF
Immigration Law and Communication: Challenges and Solutions {www.kiu.ac.ug)
PDF
Robin Fischer: A Visionary Leader Making a Difference in Healthcare, One Day ...
DOCX
Handbook of Entrepreneurship- Chapter 5: Identifying business opportunity.docx
PDF
Tortilla Mexican Grill 发射点犯得上发射点发生发射点犯得上发生
PPTX
Project Management_ SMART Projects Class.pptx
PPTX
2 - Self & Personality 587689213yiuedhwejbmansbeakjrk
PPTX
Understanding Procurement Strategies.pptx Your score increases as you pick a ...
PDF
Vinod Bhatt - Most Inspiring Supply Chain Leader in India 2025.pdf
DOCX
Center Enamel A Strategic Partner for the Modernization of Georgia's Chemical...
TRAINNING, DEVELOPMENT AND APPRAISAL.pptx
#1 Safe and Secure Verified Cash App Accounts for Purchase.pdf
Hand book of Entrepreneurship 4 Chapters.docx
33ABJFA6556B1ZP researhchzfrsdfasdfsadzd
operations management : demand supply ch
Middle East's Most Impactful Business Leaders to Follow in 2025
CTG - Business Update 2Q2025 & 6M2025.pptx
MBA2024 CGE 1.pdf file presentation 2025
Second Hand Fashion Call to Action March 2025
interschool scomp.pptxzdkjhdjvdjvdjdhjhieij
Kishore Vora - Best CFO in India to watch in 2025.pdf
Immigration Law and Communication: Challenges and Solutions {www.kiu.ac.ug)
Robin Fischer: A Visionary Leader Making a Difference in Healthcare, One Day ...
Handbook of Entrepreneurship- Chapter 5: Identifying business opportunity.docx
Tortilla Mexican Grill 发射点犯得上发射点发生发射点犯得上发生
Project Management_ SMART Projects Class.pptx
2 - Self & Personality 587689213yiuedhwejbmansbeakjrk
Understanding Procurement Strategies.pptx Your score increases as you pick a ...
Vinod Bhatt - Most Inspiring Supply Chain Leader in India 2025.pdf
Center Enamel A Strategic Partner for the Modernization of Georgia's Chemical...

Voting Based Learning Classifier System for Multi-Label Classification

  • 1. Voting-Based Learning Classifier System for multi-label classification Kaveh Ahmadi-Abhari (Presenter) Ali Hamzeh Sattar Hashemi IWLCS 2011 – Dublin, Ireland, 13th July 2011
  • 2. Multi-label Classification  Single Label Classification  Exclusive classes: each example belongs to one class  Multi-label Classification  Each instance can belong to more than one class Kaveh Ahmadi-Abhari 2 Shiraz University, Soft Computing Group
  • 3. Multi-label Classification Sky People  Single Label Classification  Exclusive classes: each example belongs to one class  Multi-label Classification  Each instance can belong to more than one class Sand Kaveh Ahmadi-Abhari 3 Shiraz University, Soft Computing Group
  • 4. Current Methods Problem • Transfer problem to a single- Transformation label classification problem Algorithm • Adapt single-label classifiers Adaptation to Solve the problem [Tsoumakas & Katakis, 2007] Kaveh Ahmadi-Abhari 4 Shiraz University, Soft Computing Group
  • 5. Problem Transformation Approaches Ex. Label- set 1a λ1 Copy Transformation 1b λ4 2a λ3 2b λ4 3 λ1 4a λ2 Ex. Label- set 4b λ3 4c λ4 1 {λ1 , λ4 } 2 {λ3 , λ4 } 3 {λ1} 4 {λ2 , λ3 , λ4 } [Tsoumakas et al., 2009] Kaveh Ahmadi-Abhari 5 Shiraz University, Soft Computing Group
  • 6. Algorithm Adaptation Approaches  Multi-label lazy algorithm  ML-kNN [Zhang & Zhou, PRJ07]  Multi-label decision trees  ADTBoost.MH [DeComité et al. MLDM03]  Multi-Label C4.5 [Clare & King, LNCS2168]  Multi-label kernel methods  Rank-SVM [Elisseeff & Weston, NIPS02]  ML-SVM [M.R. Boutell, et al. PR04]  Multi-label text categorization algorithms  BoosTexter [Schapire & Singer, MLJ00]  Maximal Margin Labeling [Kazawa et al., NIPS04]  Probabilistic generative models [McCallum, AAAI99] [Ueda & Saito, NIPS03]  BP-MLL [Zhang & Zhou, TKDE06] Kaveh Ahmadi-Abhari 6 Shiraz University, Soft Computing Group
  • 7. Motivation A lot has been done in terms of classifications using LCSs Most of these studies have been conducted for single-label classification problems Multi-label classification is in its inception [Vallim et al., IWLCS 08] Kaveh Ahmadi-Abhari 7 Shiraz University, Soft Computing Group
  • 8. Voting Based Learning Classifier System How can we guide the discovery mechanism (e.g. evolutionary operators) in LCSs? Kaveh Ahmadi-Abhari 8 Shiraz University, Soft Computing Group
  • 9. Voting Based Learning Classifier System How can we guide the discovery mechanism (e.g. evolutionary operators) in LCSs? Using the prior knowledge gained from past experiences Kaveh Ahmadi-Abhari 9 Shiraz University, Soft Computing Group
  • 10. Voting Based Learning Classifier System How can we guide the discovery mechanism (e.g. evolutionary operators) in LCSs? Using the prior knowledge gained from past experiences Training instances vote their matched rules according to how correct the rule is Kaveh Ahmadi-Abhari 10 Shiraz University, Soft Computing Group
  • 11. Voting Based Learning Classifier System How can we guide the discovery mechanism (e.g. evolutionary operators) in LCSs? Using the prior knowledge gained from past experiences Training instances vote their matched rules according to how correct the rule is Fitness measure Kaveh Ahmadi-Abhari 11 Shiraz University, Soft Computing Group
  • 12. Voting Defining Rule Types How can the given votes describe the quality of the rules accurately? Define different types for the rules such that each of these types describes the quality status the rule might have. Kaveh Ahmadi-Abhari 12 Shiraz University, Soft Computing Group
  • 13. Rule Types Example: in a single-label classification problem, rule types might be correct or wrong. Each rule might receive a “correct” or “wrong” vote from each matched training instance. A rule receives a combination of “correct” and “wrong” votes from its matched training instances Kaveh Ahmadi-Abhari 13 Shiraz University, Soft Computing Group
  • 14. Votes as Fitness Measure • Given votes • Describe the quality of the rules • Use as a fitness measure for guiding the discovery mechanism. • For example, a rule with more “wrong” votes, should be discovered with a high probability to achieve a meaningful rule Kaveh Ahmadi-Abhari 14 Shiraz University, Soft Computing Group
  • 15. Rules Definition Antecedent / Consequent ###1 / 110 0011 / 001  Antecedent part matches with the feature vector.  Consequent part are the classes predicted by the rule.  One bit for each class in the consequent part.  Value 1 in the bit indicates existence of the respective class. Kaveh Ahmadi-Abhari 15 Shiraz University, Soft Computing Group
  • 16. VLCS Vote Types for Multi-label Problem Correct Wrong Subset Multi-label Vote Types for VLCS Partial Superset Kaveh Ahmadi-Abhari 16 Shiraz University, Soft Computing Group
  • 17. Multi-Label Simple Dataset 000 111 001 1, 4 110 1, 3 010 2, 4 1, 2 101 011 100 Expand from [Vallim et al., GECCO’ 08] Kaveh Ahmadi-Abhari 17 Shiraz University, Soft Computing Group
  • 18. VLCS Voting Options for Multi-label Problem  Correct Rules (C) 111 000 001 1, 4 110 1, 3 00# /1001 2, 4 010 1, 2 101 • Is correct when it matches with: 011 • 000 or 100 • 001 Kaveh Ahmadi-Abhari 18 Shiraz University, Soft Computing Group
  • 19. VLCS Voting Options for Multi-label Problem  Wrong Rules (W) 111 000 001 1, 4 110 1, 3 0#0/0010 2, 4 010 1, 2 101 • Is wrong when it matches with: 011 • 000 or 100 • 010 Kaveh Ahmadi-Abhari 19 Shiraz University, Soft Computing Group
  • 20. VLCS Voting Options for Multi-label Problem  Subset Rules 111 000 001 1, 4 110 1, 3 #01/1000 2, 4 010 1, 2 101 • Is subset when it matches with: 011 • 001 or 100 • 101 Kaveh Ahmadi-Abhari 20 Shiraz University, Soft Computing Group
  • 21. VLCS Voting Options for Multi-label Problem  Subset Rules 111 000 001 1, 4 110 1, 3 #01/1000 2, 4 010 1, 2 101 • Is subset when it matches with: 011 • 001 or 100 • 101 Excepted Classes: 1, 4 Kaveh Ahmadi-Abhari 21 Shiraz University, Soft Computing Group
  • 22. VLCS Voting Options for Multi-label Problem  Superset Rules 111 000 001 1, 4 110 1, 3 #00/1101 2, 4 010 1, 2 101 • Is superset when it matches with: 011 • 001 or 100 • 101 Kaveh Ahmadi-Abhari 22 Shiraz University, Soft Computing Group
  • 23. VLCS Voting Options for Multi-label Problem  Superset Rules 111 000 001 1, 4 110 1, 3 #00/1101 2, 4 010 1, 2 101 • Is superset when it matches with: 011 • 001 or 100 • 101 Excepted Classes: 1, 4 Kaveh Ahmadi-Abhari 23 Shiraz University, Soft Computing Group
  • 24. VLCS Voting Options for Multi-label Problem  Partial-set Rules 111 000 001 1, 4 110 1, 3 #1# / 0110 2, 4 010 1, 2 101 • Is superset when it matches with: 011 • 010 or 100 • 111 Kaveh Ahmadi-Abhari 24 Shiraz University, Soft Computing Group
  • 25. VLCS Voting Options for Multi-label Problem  Partial-set Rules 111 000 001 1, 4 110 1, 3 #1# / 0110 2, 4 010 1, 2 101 • Is superset when it matches with: 011 • 010 or 100 • 111 Excepted Classes: 2, 4 Kaveh Ahmadi-Abhari 25 Shiraz University, Soft Computing Group
  • 26. VLCS Voting Options for Multi-label Problem 000  Rules might receive different votes 111 001 during the time 1, 4 110 1, 3 2, 4 010 1, 2 #0# / 1001 101 011 100 Kaveh Ahmadi-Abhari 26 Shiraz University, Soft Computing Group
  • 27. VLCS Voting Options for Multi-label Problem 000  Rules might receive different votes 111 001 during the time 1, 4 110 1, 3 2, 4 010 1, 2 #0# / 1001 101 011 100 Is correct for instance 000 Kaveh Ahmadi-Abhari 27 Shiraz University, Soft Computing Group
  • 28. VLCS Voting Options for Multi-label Problem 000  Rules might receive different votes 111 001 during the time 1, 4 110 1, 3 2, 4 010 1, 2 #0# / 1001 101 011 100 Is correct for Is partial-set instance 000 for instance 101 Kaveh Ahmadi-Abhari 28 Shiraz University, Soft Computing Group
  • 29. Using Stored Prior Knowledge Consider a rule that all received votes are superset } Information } The rule is covering an appropriate area of the problem Inference The rule is predicting greater number of classes for the matched input instance The number of the classes the rule predicts should be subtracted Kaveh Ahmadi-Abhari 29 Shiraz University, Soft Computing Group
  • 30. Discovery Operators  In the discovery mechanism an evolutionary algorithm with four mutation operators is defined: Kaveh Ahmadi-Abhari 30 Shiraz University, Soft Computing Group
  • 31. Discovery Operators  Mutation operators on rule’s antecedent part Generalize the rule by flipping the 0 MA-G or 1 bits to # Specializes the rule by flipping # MA-S bits to 1 or 0 Kaveh Ahmadi-Abhari 31 Shiraz University, Soft Computing Group
  • 32. Discovery Operators  Mutation operators on rule’s consequent part Subtract the number of predicted MC-S classes by flipping 1 bits to 0 Adds more classes to predicted MC-A classes by flipping 0 bits to 1 Kaveh Ahmadi-Abhari 32 Shiraz University, Soft Computing Group
  • 33. Which Discovery Operator? The votes each rule has received guide which mutation operator should act. Kaveh Ahmadi-Abhari 33 Shiraz University, Soft Computing Group
  • 34. Which Discovery Operator? The votes each rule has received guide which mutation operator should act. Wrongly Subtract the assigned some number of Superset Rule non-expected predicted classes classes (MC-S) Kaveh Ahmadi-Abhari 34 Shiraz University, Soft Computing Group
  • 35. Which Discovery Operator? Activated Mutation Rule Received Votes Operator Correct MA-G Subset MC-A Superset MC-S Partial-Set MC-A, MC-S Wrong MC-A, MC-S Correct, Subset MA-S Correct, Superset MA-G Correct, Partial-Set MA-S Correct, Wrong MA-S Wrong, Subset MA-S, MC-A Wrong, Partial MA-S Correct, Subset, Wrong MA-S, MA-G Kaveh Ahmadi-Abhari 35 Shiraz University, Soft Computing Group
  • 36. Mutation Rate • Mutation operator performs bit flipping using a probability, which is the mutation rate. • The strength of a rule is the amount of reward we predict the system to receive if the rule acts. • The more the strength, the less the mutation rate. Kaveh Ahmadi-Abhari 36 Shiraz University, Soft Computing Group
  • 37. Strength of a Rule  The mean of the rewards the rule gets over time. Reward Function: C rule ∆C expected R = 1− C rule  C expected Alteration of [Vallim et al., GECCO’ 08] Kaveh Ahmadi-Abhari 37 Shiraz University, Soft Computing Group
  • 38. Strength of a Rule  The mean of the rewards the rule gets over time. Reward Function: C rule ∆C expected R = 1− C rule  C expected A ∆B = {x : ( x ∈ A ) ⊕ ( x ∈ B )} Alteration of [Vallim et al., GECCO’ 08] Kaveh Ahmadi-Abhari 38 Shiraz University, Soft Computing Group
  • 39. Rules Rewards Input Expected Selected Received Reward Instance output Rule Vote 0001 1, 2 ###1 / 110 Correct 1 0101 1, 2, 3 ###1 / 110 Subset 0.66 0111 1 ###1 / 110 Superset 0.50 1111 1,3 ###1 / 110 Partial-set 0.33 0011 3 ###1 / 110 Wrong 0 Kaveh Ahmadi-Abhari 39 Shiraz University, Soft Computing Group
  • 40. Experimental Results  Data Sets:  Two binary datasets in the bioinformatics domain  [Chan and Freitas, GECCO’ 06 ]  Extracted from [Alves et al., 2009] Kaveh Ahmadi-Abhari 40 Shiraz University, Soft Computing Group
  • 41. Experimental Results  Quality Metrics: Accuracy • Proportion of predicted classes among all predicted or true classes Precision • Proportion of true classes among all predicted classes Recall • Proportion of predicted classes among all true classes [Tsoumakas & Katakis, 2007] Kaveh Ahmadi-Abhari 41 Shiraz University, Soft Computing Group
  • 42. Experimental Results  For the VLCS, we use a 5-fold cross validation in which the training part is used to evaluate the rules using the voting mechanism described above.  Fixed size population  initially are the most general possible rules.  In each generation, each rule is voted by its matched instances  reward is assigned  Defined mutation operators to discover new rules  The combination of the best rules among the parents and the off-springs make the next generation.  We stop the training phase if the mean strength of the rules decreases in a number of consecutive generations. Kaveh Ahmadi-Abhari 42 Shiraz University, Soft Computing Group
  • 43. Experimental Results  [Chan and Freitas, GECCO’ 06 ]  135 instances  152 attributes  Two classes • Each instance could have one or both of the available class labels. Method Accuracy Precision Recall BR 0.89 0.89 0.87 ML-KNN 0.91 0.93 0.91 VLCS 0.89 0.89 0.89 Kaveh Ahmadi-Abhari 43 Shiraz University, Soft Computing Group
  • 44. Experimental Results  Extracted from [Alves et al., 2009]  7877 proteins  40 attributes  Six classes • Each instance could have some of the available class labels. Method Accuracy Precision Recall BR 0.78 0.77 0.78 ML-KNN 0.80 0.81 0.80 VLCS 0.81 0.83 0.82 Kaveh Ahmadi-Abhari 44 Shiraz University, Soft Computing Group
  • 45. Conclusion Guiding the discovery mechanism with a prior knowledge, such that is used in VLCS, can help us solve applicable problems Kaveh Ahmadi-Abhari 45 Shiraz University, Soft Computing Group
  • 46. Future Work  A representation for dealing with numeric and nominal datasets.  Future studies on scalability and stability of the system is necessary.  Additional studies on system performance in dealing with imbalanced data and noise is also required.  Improving evolutionary operators, guiding mechanism and rule refinement. Kaveh Ahmadi-Abhari 46 Shiraz University, Soft Computing Group
  • 47. Any Question? The most exciting phrase to hear in science, the one that heralds new discoveries is not “Eureka”! (I found it!) but “That's funny...” - Isaac Asimov Kaveh Ahmadi-Abhari 47 Shiraz University, Soft Computing Group