SlideShare a Scribd company logo
Learning incoherent dictionaries for sparse
approximation using iterative projections and
                 rotations

       Daniele Barchiesi and Mark D. Plumbley

                       Centre for Digital Music
       School of Electronic Engineering and Computer Science
                  Queen Mary University of London

              daniele.barchiesi@eecs.qmul.ac.uk
               mark.plumbley@eecs.qmul.ac.uk


                       30th June 2012
Overview


     Background
           Dictionary learning model and algorithms
           Learning incoherent dictionaries
           Previous work
     Learning incoherent dictionaries using iterative projections and
     rotations
           Constructing Grassmannian frames using iterative projections
           The rotation step
           Iterative projections and rotation algorithm
     Numerical experiments
           Incoherence results, comparison with existing methods
           Sparse approximation results
     Conclusions and future research
           Proposed applications
           Summary



                  D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Background: Dictionary Learning

  Problem Definition
  Let {ym ∈ RN }M be a set of M observed signals of dimension N. The
                m=1
                 goal of dictionary learning is to express:

                                            Y ≈ ΦX

      where Y contains the signals along its columns, Φ is a dictionary
   containing unit norm atoms and every column of X contains at most S
                           non-zero coefficients.




                  D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Background: Dictionary Learning

  Problem Definition
  Let {ym ∈ RN }M be a set of M observed signals of dimension N. The
                m=1
                 goal of dictionary learning is to express:

                                              Y ≈ ΦX

      where Y contains the signals along its columns, Φ is a dictionary
   containing unit norm atoms and every column of X contains at most S
                           non-zero coefficients.

  Optimisation

                      ˆ ˆ                                                      2
                     (Φ, X) = arg min                     ||Y − ΦX||2
                                          Φ,X

                                    such that             ||xm ||0 ≤ S             ∀m

  The problem is not convex even if the               0   pseudo-norm is relaxed by the      1   norm

                    D. Barchiesi and M. D. Plumbley       Learning incoherent dictionaries
Background: Dictionary Learning Algorithms

  Optimisation Strategy
      Start from an initial dictionary Φ(0)
      Repeat for t = {1, . . . , T } iterations:
      Sparse coding : given a fixed dictionary Φ(t) , find a sparse
                    approximation X(t) with any suitable algorithm.
      Dictionary update : given X(t) , update the dictionary Φ(t+1) to
                    minimise the DL objective (possibly subject to
                    additional constraints).




                   D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Background: Dictionary Learning Algorithms

  Optimisation Strategy
      Start from an initial dictionary Φ(0)
      Repeat for t = {1, . . . , T } iterations:
      Sparse coding : given a fixed dictionary Φ(t) , find a sparse
                    approximation X(t) with any suitable algorithm.
      Dictionary update : given X(t) , update the dictionary Φ(t+1) to
                    minimise the DL objective (possibly subject to
                    additional constraints).

  Previous Work
  Methods for dictionary learning include:
      Probabilistic models [Lewicki and Sejnowski]
      Method of optimal directions ( mod) [Engan et al.]
      k-svd [Aharon et al.]
      Online learning [Mairal et al.]

                   D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Learning Incoherent Dictionaries

  Mutual Coherence
  The coherence of a dictionary expresses the similarity between atoms or
  groups of atoms in the dictionary. The mutual coherence is defined as:
                                          def
                               µ(Φ) = max φi , φj
                                                    i=j

  Results on sparse recovery link the performance of sparse approximation
  algorithms to the coherence of the dictionary. For over-complete
  approximations, low µ leads to recovery guarantees.




                  D. Barchiesi and M. D. Plumbley         Learning incoherent dictionaries
Learning Incoherent Dictionaries

  Mutual Coherence
  The coherence of a dictionary expresses the similarity between atoms or
  groups of atoms in the dictionary. The mutual coherence is defined as:
                                           def
                                µ(Φ) = max φi , φj
                                                     i=j

  Results on sparse recovery link the performance of sparse approximation
  algorithms to the coherence of the dictionary. For over-complete
  approximations, low µ leads to recovery guarantees.

  Goal
  The objective is to learn dictionaries that are both:
      Well adapted to a set of training data Y
      Mutually incoherent


                   D. Barchiesi and M. D. Plumbley         Learning incoherent dictionaries
Learning Incoherent Dictionaries

  Advantages
  Advantages of incoherent dictionaries include:
      Sub-dictionaries have low condition number and their
      (pseudo)inverse computed by many sparse approximation algorithms
      is well-posed.
      Convergence of greedy algorithms is faster for incoherent dictionaries
      (experimental results).
      Application-oriented intuitions (future work).




                   D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Learning Incoherent Dictionaries

  Advantages
  Advantages of incoherent dictionaries include:
      Sub-dictionaries have low condition number and their
      (pseudo)inverse computed by many sparse approximation algorithms
      is well-posed.
      Convergence of greedy algorithms is faster for incoherent dictionaries
      (experimental results).
      Application-oriented intuitions (future work).

  Previous Work
      Method of coherence-constrained directions ( mocod) [Sapiro et al.]
      Incoherent k-svd ( ink-svd) [Mailh´ et al.]
                                        e
      Parametric dictionary design for sparse coding [Yaghoobi et al.]



                   D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Incoherent Dictionary Learning: Previous Work
  mocod
  Uncostrained, penalised optimisation:
              ˆ ˆ
             (Φ, X) =arg min            ||Y − ΦX||2 + τ
                                                  F                      log(|xkm | + β)+
                          Φ,X
                                                                   m,n
                                                      K
                                                                                 2
                        + ζ ||G − I||2 + η
                                     F                      ||φk ||2 − 1
                                                                   2
                                                      k=1

  where the factor multiplied by τ promotes sparsity and the factors multiplied by
  ζ and η promote incoherence and unit-norm atoms.




                    D. Barchiesi and M. D. Plumbley    Learning incoherent dictionaries
Incoherent Dictionary Learning: Previous Work
  mocod
  Uncostrained, penalised optimisation:
              ˆ ˆ
             (Φ, X) =arg min            ||Y − ΦX||2 + τ
                                                  F                      log(|xkm | + β)+
                          Φ,X
                                                                   m,n
                                                      K
                                                                                 2
                        + ζ ||G − I||2 + η
                                     F                      ||φk ||2 − 1
                                                                   2
                                                      k=1

  where the factor multiplied by τ promotes sparsity and the factors multiplied by
  ζ and η promote incoherence and unit-norm atoms.

  ink-svd
  Greedy algorithm that includes a dictionary de-correlation step after a
  k-svd dictionary update:
       Find pairs of coherent atoms
       De-correlate atoms two-by-two
       Repeat until a target mutual coherence is reached

                    D. Barchiesi and M. D. Plumbley    Learning incoherent dictionaries
ipr Algorithm: constructing Grassmannian frames

  A Grassmannian frame is a dictionary with minimal mutual coherence.
                                  K −N
  For a N × K dictionary, µ ≥ N(K −1) and this bound can be reached
  only for some (N, K ) pairs.




                   D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
ipr Algorithm: constructing Grassmannian frames

  A Grassmannian frame is a dictionary with minimal mutual coherence.
                                  K −N
  For a N × K dictionary, µ ≥ N(K −1) and this bound can be reached
  only for some (N, K ) pairs.

  Iterative Projections Algorithm
      Start from an initial dictionary Φ(0)
                                                     def         T
      Calculate its Gram matrix G(0) = Φ(0) Φ(0)
      Repeat for t = {0, . . . , T − 1} iterations:
      Project Gram matrix onto the structural constraint set
                     def
                Kµ0 = {K : K = KT , diag(K) = 1, max |ki,j | ≤ µ0 }.
                                                                           i>j

      Project Gram matrix onto the spectral constraint set
                       def
                   F =         F : F = FT , eig(F) ≥ 0, rank(F) ≤ N
                                                             T
      Factorise the Gram matrix as Φ(T −1) Φ(T −1) = G(T −1)

                   D. Barchiesi and M. D. Plumbley     Learning incoherent dictionaries
ipr Algorithm: the rotation step

  Idea!
  The factorisation at the end of the iterative projection algorithm is not
  unique, since for any orthonormal matrix W
                              T
                   (WΦ) (WΦ) = ΦT WT WΦ = ΦT Φ

  Therefore, we can optimise an orthonormal matrix for the DL objective!
  This is an (improper) rotation of the dictionary Φ.




                   D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
ipr Algorithm: the rotation step

  Idea!
  The factorisation at the end of the iterative projection algorithm is not
  unique, since for any orthonormal matrix W
                              T
                   (WΦ) (WΦ) = ΦT WT WΦ = ΦT Φ

  Therefore, we can optimise an orthonormal matrix for the DL objective!
  This is an (improper) rotation of the dictionary Φ.

  Dictionary Rotation
                      ˆ
                      W = arg min                    ||Y − WΦX||F
                                  W:WT W=I

  A closed-form solution to this problem can be found by computing the
                                                 def
  svd decomposition of the covariance matrix C = ΦXYT = UΣV T
  and setting:
                                  ˆ
                                 W = VUT

                   D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Iterative Projections and Rotations algorithm

     Start from a dictionary Φ(0) returned by the dictionary update step of any
     DL algorithm.
     Repeat for t = {0, . . . , T − 1} iterations:
                                                           T
     Calculate the Gram matrix: G(t) ← Φ(t) Φ(t)
     Project Gram matrix onto the structural constraint set:

          diag(G) ← 1
          G ← Limit(G, µ0 )
     Factorise Gram matrix and project it onto the spectral constraint set

          [Q, Λ] ← evd(G)
          Λ ← Thresh(Λ, N)
          Φ ← Λ1/2 QT
     Rotate the dictionary

          C ← Y(ΦX)T
          [U, Σ, V] ← svd(C)
          W ← VUT
          Φ ← WΦ
                   D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Numerical Experiments: The SMALLBox framework


  The SMALLBox is a Matlab framework for benchmarking and developing
  dictionary learning algorithms developed by a team at Queen Mary
  University of London.
      Latest version can be downloaded from
      http://code.soundsoftware.ac.uk/
      SMALLBox integrates many third-party toolboxes such as Sparco,
      SparseLab, CVX, SPAMS, etc.
      SMALLBox provides a unique interface for different DL algorithms
      that can be used for benchmark
      The new distribution of SMALLBox allows to program add-ons to
      expand the functionalities of the framework without interfering with
      the core code.
      IncoherentDL is a SMALLBox add-on and can be used to reproduce
      some of the results presented in this talk.



                  D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Numerical Experiments: Mutual coherence vs residual norm

  Test Conditions
       Tests on a 16kHz guitar audio signal divided in N = 256 long
       overlapping blocks
      A fixed number of active atoms was chosen S = 12 (around 5% of
      the dimension N).
      A twice-overcomplete dictionary was initialised with either:
           Randomly selected samples from the training set.
           An over-complete Gabor frame.
      DL algorithms were run for 50 iterations.

  Test Objective
  The mutual coherence achieved by every learned dictionary is paired with
  the approximation error defined as:

                                                            ||Y||F
                     snr(Φ, X) = 20 log10                            .
                                                         ||Y − ΦX||F

                   D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Numerical Experiments: mocod updates, data init.

                               mutual coherence µ                                                                                        reconstruction error




        1                                                                                                     25

      0.9
                                                                                                              20
      0.8
                                                                                                              15




                                                                                                   SNR (dB)
      0.7
  µ




      0.6                                                                                                     10

      0.5
                                                                                                               5
      0.4

       0.3                                                                                  10
                                                                                              −2               0                                                               −2
                                                                                                                                                                              10
        −2                                                                                                    −2
      10                                                                        0                       10                                                                0
              0                                                               10                                         0                                               10
             10                                                    2      η                                             10                                       2
                       2                                       10                                                                    2                          10   η
                  ζ   10                                                                                                     ζ      10
                                                4      4                                                                                             4      4
                                               10     10                                                                                            10   10

                                                                       mutual coherence−reconstruction error scatter plot
                                      25



                                      20            ← µmin                                                                                       ← µmax



                                      15
                           SNR (dB)




                                      10



                                       5



                                       0
                                           0                 0.2                    0.4              0.6                      0.8               1
                                                                                          mutual coherence µ




                                      D. Barchiesi and M. D. Plumbley                                          Learning incoherent dictionaries
Numerical Experiments: mocod updates, Gabor init.
                               mutual coherence µ
                                                                                                                                         reconstruction error



        1
                                                                                                              25
      0.9
                                                                                                              20
      0.8

      0.7                                                                                                     15




                                                                                                   SNR (dB)
  µ




      0.6
                                                                                                              10
      0.5
                                                                                                               5
      0.4

       0.3                                                                                  10
                                                                                              −2
                                                                                                               0                                                               −2
        −2                                                                                                    −2                                                              10
      10                                                                        0                       10                                                                0
              0                                                               10                                         0                                               10
             10                                                                                                         10
                  ζ    2                                       10
                                                                   2
                                                                          η                                                  ζ       2                           2   η
                      10                                                                                                            10                          10
                                                4      4                                                                                             4      4
                                               10     10                                                                                            10   10

                                                                       mutual coherence−reconstruction error scatter plot
                                      25



                                      20            ← µmin                                                                                       ← µmax



                                      15
                           SNR (dB)




                                      10



                                       5



                                       0
                                           0                 0.2                    0.4              0.6                      0.8               1
                                                                                          mutual coherence µ




                                      D. Barchiesi and M. D. Plumbley                                          Learning incoherent dictionaries
Numerical Experiments: ink-svd and ipr
                                                 Data Initialisation
                     25

                     20   ← µmin                                                         ← µmax



          SNR (dB)
                     15

                     10
                                                                                  IPR
                      5                                                           INK−SVD

                      0
                          0.05         0.1                              0.5             1
                                               mutual coherence µ
                                                Gabor Initialisation
                     25

                     20   ← µmin                                                         ← µmax
          SNR (dB)




                     15

                     10
                                                                                  IPR
                      5                                                           INK−SVD
                      0
                          0.05         0.1                              0.5             1
                                               mutual coherence µ

                      D. Barchiesi and M. D. Plumbley      Learning incoherent dictionaries
Numerical Experiments: Sparse Approximation

  Test Conditions
      Matching pursuit algorithm ( mp) run for 1000 iterations on the
      following signals:
           Training set.
           Different guitar recording taken from the rwc database.
           A piano recording taken from the rwc database.
      Dictionaries with different mutual coherences were selected as
      returned from the ipr algorithm with data initialisation.

  Test Objective
  The norm of the residual in decibel defined as:

                                     20 log10 ||y − Φx||2

  is computed and averaged over the number of signals M and 10
  dictionaries resulting from independent trials of the learning algorithm.


                    D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Numerical Experiments: Training set approximation

                                                                 guitar − training signal
                                   0


                                −20
   average residual norm (dB)




                                −40


                                −60

                                             µ = 0.72
                                −80
                                             µ = 0.37
                                             µ = 0.19
                                −100         µ = 0.1
                                             µ = 0.06
                                −120
                                       0       200                 400           600                            800   1000
                                                                  number of iterations


                                           D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Numerical Experiments: Guitar approximation

                                                                    guitar − test signal
                                   0


                                −20
   average residual norm (dB)




                                −40


                                −60

                                             µ = 0.72
                                −80          µ = 0.37
                                             µ = 0.19
                                −100         µ = 0.1
                                             µ = 0.06
                                −120
                                       0       200                 400           600                            800   1000
                                                                  number of iterations


                                           D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Numerical Experiments: Piano approximation

                                                                             piano
                                   0


                                −20
   average residual norm (dB)




                                −40


                                −60
                                             µ = 0.72
                                −80          µ = 0.37
                                             µ = 0.19
                                −100         µ = 0.1
                                             µ = 0.06
                                −120
                                       0       200                 400           600                             800   1000
                                                                  number of iterations


                                           D. Barchiesi and M. D. Plumbley    Learning incoherent dictionaries
Conclusions: Possible Applications


  Morphological Component Analysis
  Morphological component analysis is a dictionary learning approach to
  classification.
      Different dictionaries are learned on morphologically dissimilar
      training sets (e.g., edges and textures, percussive and steady state
      sounds)
      A test signal is classified according to the support or magnitude of
      the coefficients of its sparse approximation (i.e., what is the best
      dictionary to represent it?)
  ipr could be used to enforce incoherence between the atoms belonging
  to different morphological components and enhance classification and
  separation performance.




                  D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Conclusions: Possible Applications


  Blind Compressed Sensing
  Blind compressed sensing generalises compressed sensing to the case of
  an unknown dictionary that generates the signals to be recovered.
      A set of observations Z is acquired through the known measurement
      matrix M. Z = MY = MΦX
      Dictionary learning is used to optimise Ψ and factorize the observed
      data as Z ≈ ΨX.
                                                                  ˆ
      The learned dictionary is factorized as the product Ψ ≈ MΦ and
                                    ˆ    ˆ
      the signals reconstructed as Y = ΦX.
  The two factorisations are not unique and strong constraints on Φ are
  assumed to correctly reconstruct the signals. ipr might be used to
  constrain the factorisations and lead to a less ambiguous solution.




                  D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Conclusions: Summary


     The ipr algorithm can be used to learn dictionaries that are both
     adapted to a training set and mutually incoherent.
     The ipr algorithm can be used as a de-correlation step in any
     dictionary learning algorithm.
     Experimental data show that ipr performed generally better than
     benchmark techniques on audio signals.
     Incoherent dictionaries are useful for sparse approximation and could
     be used in a number of potential applications.




                 D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Conclusions: Summary


     The ipr algorithm can be used to learn dictionaries that are both
     adapted to a training set and mutually incoherent.
     The ipr algorithm can be used as a de-correlation step in any
     dictionary learning algorithm.
     Experimental data show that ipr performed generally better than
     benchmark techniques on audio signals.
     Incoherent dictionaries are useful for sparse approximation and could
     be used in a number of potential applications.


        Thank you for your attention
           and for any question!

                 D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries

More Related Content

PDF
Cooperating Techniques for Extracting Conceptual Taxonomies from Text
PDF
Dependent Types in Natural Language Semantics
PDF
Latent dirichletallocation presentation
PPT
Topic Models - LDA and Correlated Topic Models
PDF
Topic model an introduction
PDF
Composing (Im)politeness in Dependent Type Semantics
PDF
Data Complexity in EL Family of Description Logics
PDF
Latent Dirichlet Allocation
Cooperating Techniques for Extracting Conceptual Taxonomies from Text
Dependent Types in Natural Language Semantics
Latent dirichletallocation presentation
Topic Models - LDA and Correlated Topic Models
Topic model an introduction
Composing (Im)politeness in Dependent Type Semantics
Data Complexity in EL Family of Description Logics
Latent Dirichlet Allocation

What's hot (20)

PDF
Jarrar.lecture notes.aai.2012s.descriptionlogic
PDF
Harnessing Deep Neural Networks with Logic Rules
PPT
Topic Models
PDF
TopicModels_BleiPaper_Summary.pptx
PDF
slides for "Supervised Model Learning with Feature Grouping based on a Discre...
PDF
Tensor-based Models of Natural Language Semantics
PPTX
Formal language
PDF
Reduction Monads and Their Signatures
PDF
Rethinking Perturbations in Encoder-Decoders for Fast Training
PDF
Conditional preference queries and possibilistic logic
PDF
Learning Morphological Rules for Amharic Verbs Using Inductive Logic Programming
PDF
Learning phoneme mappings for transliteration without parallel data
PDF
Compositional Distributional Models of Meaning
PDF
A survey on parallel corpora alignment
PDF
W17 5406
PPTX
Jarrar: First Order Logic- Inference Methods
PDF
深層意味表現学習 (Deep Semantic Representations)
PDF
Csr2011 june14 12_00_hansen
PDF
Exempler approach
Jarrar.lecture notes.aai.2012s.descriptionlogic
Harnessing Deep Neural Networks with Logic Rules
Topic Models
TopicModels_BleiPaper_Summary.pptx
slides for "Supervised Model Learning with Feature Grouping based on a Discre...
Tensor-based Models of Natural Language Semantics
Formal language
Reduction Monads and Their Signatures
Rethinking Perturbations in Encoder-Decoders for Fast Training
Conditional preference queries and possibilistic logic
Learning Morphological Rules for Amharic Verbs Using Inductive Logic Programming
Learning phoneme mappings for transliteration without parallel data
Compositional Distributional Models of Meaning
A survey on parallel corpora alignment
W17 5406
Jarrar: First Order Logic- Inference Methods
深層意味表現学習 (Deep Semantic Representations)
Csr2011 june14 12_00_hansen
Exempler approach
Ad

Similar to Icml12 (20)

PDF
形式言語理論への 測度論的アプローチ
PDF
Algorithms and Complexity: Cryptography Theory
PDF
PDF
Topological Inference via Meshing
PDF
Canini09a
PDF
Cooperating Techniques for Extracting Conceptual Taxonomies from Text
PDF
Research Summary: Hidden Topic Markov Models, Gruber
PPT
KNOWLEDGE Representation unit 3 for data mining
PPTX
Dirichlet processes and Applications
PPTX
lec2cct computational cmplexity theory.pptx
PDF
Ml mle_bayes
PDF
IVR - Chapter 7 - Patch models and dictionary learning
PDF
Simple effective decipherment via combinatorial optimization
PDF
Higher dimensional image analysis using brunn minkowski theorem, convexity an...
PDF
Functional analysis in mechanics 2e
PDF
Functional analysis in mechanics
PDF
Astaño 4
PDF
Tro07 sparse-solutions-talk
PDF
CUMC talk notes v3
PDF
A course on mathematical logic
形式言語理論への 測度論的アプローチ
Algorithms and Complexity: Cryptography Theory
Topological Inference via Meshing
Canini09a
Cooperating Techniques for Extracting Conceptual Taxonomies from Text
Research Summary: Hidden Topic Markov Models, Gruber
KNOWLEDGE Representation unit 3 for data mining
Dirichlet processes and Applications
lec2cct computational cmplexity theory.pptx
Ml mle_bayes
IVR - Chapter 7 - Patch models and dictionary learning
Simple effective decipherment via combinatorial optimization
Higher dimensional image analysis using brunn minkowski theorem, convexity an...
Functional analysis in mechanics 2e
Functional analysis in mechanics
Astaño 4
Tro07 sparse-solutions-talk
CUMC talk notes v3
A course on mathematical logic
Ad

Recently uploaded (20)

PPTX
O2C Customer Invoices to Receipt V15A.pptx
PPTX
MicrosoftCybserSecurityReferenceArchitecture-April-2025.pptx
PDF
Univ-Connecticut-ChatGPT-Presentaion.pdf
PDF
Zenith AI: Advanced Artificial Intelligence
PDF
Architecture types and enterprise applications.pdf
PPTX
1. Introduction to Computer Programming.pptx
PPTX
Programs and apps: productivity, graphics, security and other tools
PDF
NewMind AI Weekly Chronicles – August ’25 Week III
PDF
Web App vs Mobile App What Should You Build First.pdf
PDF
Profit Center Accounting in SAP S/4HANA, S4F28 Col11
PDF
gpt5_lecture_notes_comprehensive_20250812015547.pdf
PDF
Transform Your ITIL® 4 & ITSM Strategy with AI in 2025.pdf
PDF
From MVP to Full-Scale Product A Startup’s Software Journey.pdf
PPT
Module 1.ppt Iot fundamentals and Architecture
PDF
A novel scalable deep ensemble learning framework for big data classification...
PDF
August Patch Tuesday
PDF
2021 HotChips TSMC Packaging Technologies for Chiplets and 3D_0819 publish_pu...
PDF
Hybrid model detection and classification of lung cancer
PPTX
TechTalks-8-2019-Service-Management-ITIL-Refresh-ITIL-4-Framework-Supports-Ou...
PDF
ENT215_Completing-a-large-scale-migration-and-modernization-with-AWS.pdf
O2C Customer Invoices to Receipt V15A.pptx
MicrosoftCybserSecurityReferenceArchitecture-April-2025.pptx
Univ-Connecticut-ChatGPT-Presentaion.pdf
Zenith AI: Advanced Artificial Intelligence
Architecture types and enterprise applications.pdf
1. Introduction to Computer Programming.pptx
Programs and apps: productivity, graphics, security and other tools
NewMind AI Weekly Chronicles – August ’25 Week III
Web App vs Mobile App What Should You Build First.pdf
Profit Center Accounting in SAP S/4HANA, S4F28 Col11
gpt5_lecture_notes_comprehensive_20250812015547.pdf
Transform Your ITIL® 4 & ITSM Strategy with AI in 2025.pdf
From MVP to Full-Scale Product A Startup’s Software Journey.pdf
Module 1.ppt Iot fundamentals and Architecture
A novel scalable deep ensemble learning framework for big data classification...
August Patch Tuesday
2021 HotChips TSMC Packaging Technologies for Chiplets and 3D_0819 publish_pu...
Hybrid model detection and classification of lung cancer
TechTalks-8-2019-Service-Management-ITIL-Refresh-ITIL-4-Framework-Supports-Ou...
ENT215_Completing-a-large-scale-migration-and-modernization-with-AWS.pdf

Icml12

  • 1. Learning incoherent dictionaries for sparse approximation using iterative projections and rotations Daniele Barchiesi and Mark D. Plumbley Centre for Digital Music School of Electronic Engineering and Computer Science Queen Mary University of London daniele.barchiesi@eecs.qmul.ac.uk mark.plumbley@eecs.qmul.ac.uk 30th June 2012
  • 2. Overview Background Dictionary learning model and algorithms Learning incoherent dictionaries Previous work Learning incoherent dictionaries using iterative projections and rotations Constructing Grassmannian frames using iterative projections The rotation step Iterative projections and rotation algorithm Numerical experiments Incoherence results, comparison with existing methods Sparse approximation results Conclusions and future research Proposed applications Summary D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 3. Background: Dictionary Learning Problem Definition Let {ym ∈ RN }M be a set of M observed signals of dimension N. The m=1 goal of dictionary learning is to express: Y ≈ ΦX where Y contains the signals along its columns, Φ is a dictionary containing unit norm atoms and every column of X contains at most S non-zero coefficients. D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 4. Background: Dictionary Learning Problem Definition Let {ym ∈ RN }M be a set of M observed signals of dimension N. The m=1 goal of dictionary learning is to express: Y ≈ ΦX where Y contains the signals along its columns, Φ is a dictionary containing unit norm atoms and every column of X contains at most S non-zero coefficients. Optimisation ˆ ˆ 2 (Φ, X) = arg min ||Y − ΦX||2 Φ,X such that ||xm ||0 ≤ S ∀m The problem is not convex even if the 0 pseudo-norm is relaxed by the 1 norm D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 5. Background: Dictionary Learning Algorithms Optimisation Strategy Start from an initial dictionary Φ(0) Repeat for t = {1, . . . , T } iterations: Sparse coding : given a fixed dictionary Φ(t) , find a sparse approximation X(t) with any suitable algorithm. Dictionary update : given X(t) , update the dictionary Φ(t+1) to minimise the DL objective (possibly subject to additional constraints). D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 6. Background: Dictionary Learning Algorithms Optimisation Strategy Start from an initial dictionary Φ(0) Repeat for t = {1, . . . , T } iterations: Sparse coding : given a fixed dictionary Φ(t) , find a sparse approximation X(t) with any suitable algorithm. Dictionary update : given X(t) , update the dictionary Φ(t+1) to minimise the DL objective (possibly subject to additional constraints). Previous Work Methods for dictionary learning include: Probabilistic models [Lewicki and Sejnowski] Method of optimal directions ( mod) [Engan et al.] k-svd [Aharon et al.] Online learning [Mairal et al.] D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 7. Learning Incoherent Dictionaries Mutual Coherence The coherence of a dictionary expresses the similarity between atoms or groups of atoms in the dictionary. The mutual coherence is defined as: def µ(Φ) = max φi , φj i=j Results on sparse recovery link the performance of sparse approximation algorithms to the coherence of the dictionary. For over-complete approximations, low µ leads to recovery guarantees. D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 8. Learning Incoherent Dictionaries Mutual Coherence The coherence of a dictionary expresses the similarity between atoms or groups of atoms in the dictionary. The mutual coherence is defined as: def µ(Φ) = max φi , φj i=j Results on sparse recovery link the performance of sparse approximation algorithms to the coherence of the dictionary. For over-complete approximations, low µ leads to recovery guarantees. Goal The objective is to learn dictionaries that are both: Well adapted to a set of training data Y Mutually incoherent D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 9. Learning Incoherent Dictionaries Advantages Advantages of incoherent dictionaries include: Sub-dictionaries have low condition number and their (pseudo)inverse computed by many sparse approximation algorithms is well-posed. Convergence of greedy algorithms is faster for incoherent dictionaries (experimental results). Application-oriented intuitions (future work). D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 10. Learning Incoherent Dictionaries Advantages Advantages of incoherent dictionaries include: Sub-dictionaries have low condition number and their (pseudo)inverse computed by many sparse approximation algorithms is well-posed. Convergence of greedy algorithms is faster for incoherent dictionaries (experimental results). Application-oriented intuitions (future work). Previous Work Method of coherence-constrained directions ( mocod) [Sapiro et al.] Incoherent k-svd ( ink-svd) [Mailh´ et al.] e Parametric dictionary design for sparse coding [Yaghoobi et al.] D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 11. Incoherent Dictionary Learning: Previous Work mocod Uncostrained, penalised optimisation: ˆ ˆ (Φ, X) =arg min ||Y − ΦX||2 + τ F log(|xkm | + β)+ Φ,X m,n K 2 + ζ ||G − I||2 + η F ||φk ||2 − 1 2 k=1 where the factor multiplied by τ promotes sparsity and the factors multiplied by ζ and η promote incoherence and unit-norm atoms. D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 12. Incoherent Dictionary Learning: Previous Work mocod Uncostrained, penalised optimisation: ˆ ˆ (Φ, X) =arg min ||Y − ΦX||2 + τ F log(|xkm | + β)+ Φ,X m,n K 2 + ζ ||G − I||2 + η F ||φk ||2 − 1 2 k=1 where the factor multiplied by τ promotes sparsity and the factors multiplied by ζ and η promote incoherence and unit-norm atoms. ink-svd Greedy algorithm that includes a dictionary de-correlation step after a k-svd dictionary update: Find pairs of coherent atoms De-correlate atoms two-by-two Repeat until a target mutual coherence is reached D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 13. ipr Algorithm: constructing Grassmannian frames A Grassmannian frame is a dictionary with minimal mutual coherence. K −N For a N × K dictionary, µ ≥ N(K −1) and this bound can be reached only for some (N, K ) pairs. D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 14. ipr Algorithm: constructing Grassmannian frames A Grassmannian frame is a dictionary with minimal mutual coherence. K −N For a N × K dictionary, µ ≥ N(K −1) and this bound can be reached only for some (N, K ) pairs. Iterative Projections Algorithm Start from an initial dictionary Φ(0) def T Calculate its Gram matrix G(0) = Φ(0) Φ(0) Repeat for t = {0, . . . , T − 1} iterations: Project Gram matrix onto the structural constraint set def Kµ0 = {K : K = KT , diag(K) = 1, max |ki,j | ≤ µ0 }. i>j Project Gram matrix onto the spectral constraint set def F = F : F = FT , eig(F) ≥ 0, rank(F) ≤ N T Factorise the Gram matrix as Φ(T −1) Φ(T −1) = G(T −1) D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 15. ipr Algorithm: the rotation step Idea! The factorisation at the end of the iterative projection algorithm is not unique, since for any orthonormal matrix W T (WΦ) (WΦ) = ΦT WT WΦ = ΦT Φ Therefore, we can optimise an orthonormal matrix for the DL objective! This is an (improper) rotation of the dictionary Φ. D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 16. ipr Algorithm: the rotation step Idea! The factorisation at the end of the iterative projection algorithm is not unique, since for any orthonormal matrix W T (WΦ) (WΦ) = ΦT WT WΦ = ΦT Φ Therefore, we can optimise an orthonormal matrix for the DL objective! This is an (improper) rotation of the dictionary Φ. Dictionary Rotation ˆ W = arg min ||Y − WΦX||F W:WT W=I A closed-form solution to this problem can be found by computing the def svd decomposition of the covariance matrix C = ΦXYT = UΣV T and setting: ˆ W = VUT D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 17. Iterative Projections and Rotations algorithm Start from a dictionary Φ(0) returned by the dictionary update step of any DL algorithm. Repeat for t = {0, . . . , T − 1} iterations: T Calculate the Gram matrix: G(t) ← Φ(t) Φ(t) Project Gram matrix onto the structural constraint set: diag(G) ← 1 G ← Limit(G, µ0 ) Factorise Gram matrix and project it onto the spectral constraint set [Q, Λ] ← evd(G) Λ ← Thresh(Λ, N) Φ ← Λ1/2 QT Rotate the dictionary C ← Y(ΦX)T [U, Σ, V] ← svd(C) W ← VUT Φ ← WΦ D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 18. Numerical Experiments: The SMALLBox framework The SMALLBox is a Matlab framework for benchmarking and developing dictionary learning algorithms developed by a team at Queen Mary University of London. Latest version can be downloaded from http://code.soundsoftware.ac.uk/ SMALLBox integrates many third-party toolboxes such as Sparco, SparseLab, CVX, SPAMS, etc. SMALLBox provides a unique interface for different DL algorithms that can be used for benchmark The new distribution of SMALLBox allows to program add-ons to expand the functionalities of the framework without interfering with the core code. IncoherentDL is a SMALLBox add-on and can be used to reproduce some of the results presented in this talk. D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 19. Numerical Experiments: Mutual coherence vs residual norm Test Conditions Tests on a 16kHz guitar audio signal divided in N = 256 long overlapping blocks A fixed number of active atoms was chosen S = 12 (around 5% of the dimension N). A twice-overcomplete dictionary was initialised with either: Randomly selected samples from the training set. An over-complete Gabor frame. DL algorithms were run for 50 iterations. Test Objective The mutual coherence achieved by every learned dictionary is paired with the approximation error defined as: ||Y||F snr(Φ, X) = 20 log10 . ||Y − ΦX||F D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 20. Numerical Experiments: mocod updates, data init. mutual coherence µ reconstruction error 1 25 0.9 20 0.8 15 SNR (dB) 0.7 µ 0.6 10 0.5 5 0.4 0.3 10 −2 0 −2 10 −2 −2 10 0 10 0 0 10 0 10 10 2 η 10 2 2 10 2 10 η ζ 10 ζ 10 4 4 4 4 10 10 10 10 mutual coherence−reconstruction error scatter plot 25 20 ← µmin ← µmax 15 SNR (dB) 10 5 0 0 0.2 0.4 0.6 0.8 1 mutual coherence µ D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 21. Numerical Experiments: mocod updates, Gabor init. mutual coherence µ reconstruction error 1 25 0.9 20 0.8 0.7 15 SNR (dB) µ 0.6 10 0.5 5 0.4 0.3 10 −2 0 −2 −2 −2 10 10 0 10 0 0 10 0 10 10 10 ζ 2 10 2 η ζ 2 2 η 10 10 10 4 4 4 4 10 10 10 10 mutual coherence−reconstruction error scatter plot 25 20 ← µmin ← µmax 15 SNR (dB) 10 5 0 0 0.2 0.4 0.6 0.8 1 mutual coherence µ D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 22. Numerical Experiments: ink-svd and ipr Data Initialisation 25 20 ← µmin ← µmax SNR (dB) 15 10 IPR 5 INK−SVD 0 0.05 0.1 0.5 1 mutual coherence µ Gabor Initialisation 25 20 ← µmin ← µmax SNR (dB) 15 10 IPR 5 INK−SVD 0 0.05 0.1 0.5 1 mutual coherence µ D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 23. Numerical Experiments: Sparse Approximation Test Conditions Matching pursuit algorithm ( mp) run for 1000 iterations on the following signals: Training set. Different guitar recording taken from the rwc database. A piano recording taken from the rwc database. Dictionaries with different mutual coherences were selected as returned from the ipr algorithm with data initialisation. Test Objective The norm of the residual in decibel defined as: 20 log10 ||y − Φx||2 is computed and averaged over the number of signals M and 10 dictionaries resulting from independent trials of the learning algorithm. D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 24. Numerical Experiments: Training set approximation guitar − training signal 0 −20 average residual norm (dB) −40 −60 µ = 0.72 −80 µ = 0.37 µ = 0.19 −100 µ = 0.1 µ = 0.06 −120 0 200 400 600 800 1000 number of iterations D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 25. Numerical Experiments: Guitar approximation guitar − test signal 0 −20 average residual norm (dB) −40 −60 µ = 0.72 −80 µ = 0.37 µ = 0.19 −100 µ = 0.1 µ = 0.06 −120 0 200 400 600 800 1000 number of iterations D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 26. Numerical Experiments: Piano approximation piano 0 −20 average residual norm (dB) −40 −60 µ = 0.72 −80 µ = 0.37 µ = 0.19 −100 µ = 0.1 µ = 0.06 −120 0 200 400 600 800 1000 number of iterations D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 27. Conclusions: Possible Applications Morphological Component Analysis Morphological component analysis is a dictionary learning approach to classification. Different dictionaries are learned on morphologically dissimilar training sets (e.g., edges and textures, percussive and steady state sounds) A test signal is classified according to the support or magnitude of the coefficients of its sparse approximation (i.e., what is the best dictionary to represent it?) ipr could be used to enforce incoherence between the atoms belonging to different morphological components and enhance classification and separation performance. D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 28. Conclusions: Possible Applications Blind Compressed Sensing Blind compressed sensing generalises compressed sensing to the case of an unknown dictionary that generates the signals to be recovered. A set of observations Z is acquired through the known measurement matrix M. Z = MY = MΦX Dictionary learning is used to optimise Ψ and factorize the observed data as Z ≈ ΨX. ˆ The learned dictionary is factorized as the product Ψ ≈ MΦ and ˆ ˆ the signals reconstructed as Y = ΦX. The two factorisations are not unique and strong constraints on Φ are assumed to correctly reconstruct the signals. ipr might be used to constrain the factorisations and lead to a less ambiguous solution. D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 29. Conclusions: Summary The ipr algorithm can be used to learn dictionaries that are both adapted to a training set and mutually incoherent. The ipr algorithm can be used as a de-correlation step in any dictionary learning algorithm. Experimental data show that ipr performed generally better than benchmark techniques on audio signals. Incoherent dictionaries are useful for sparse approximation and could be used in a number of potential applications. D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 30. Conclusions: Summary The ipr algorithm can be used to learn dictionaries that are both adapted to a training set and mutually incoherent. The ipr algorithm can be used as a de-correlation step in any dictionary learning algorithm. Experimental data show that ipr performed generally better than benchmark techniques on audio signals. Incoherent dictionaries are useful for sparse approximation and could be used in a number of potential applications. Thank you for your attention and for any question! D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries