SlideShare ist ein Scribd-Unternehmen logo
1 von 30
Downloaden Sie, um offline zu lesen
Learning incoherent dictionaries for sparse
approximation using iterative projections and
                 rotations

       Daniele Barchiesi and Mark D. Plumbley

                       Centre for Digital Music
       School of Electronic Engineering and Computer Science
                  Queen Mary University of London

              daniele.barchiesi@eecs.qmul.ac.uk
               mark.plumbley@eecs.qmul.ac.uk


                       30th June 2012
Overview


     Background
           Dictionary learning model and algorithms
           Learning incoherent dictionaries
           Previous work
     Learning incoherent dictionaries using iterative projections and
     rotations
           Constructing Grassmannian frames using iterative projections
           The rotation step
           Iterative projections and rotation algorithm
     Numerical experiments
           Incoherence results, comparison with existing methods
           Sparse approximation results
     Conclusions and future research
           Proposed applications
           Summary



                  D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Background: Dictionary Learning

  Problem Definition
  Let {ym ∈ RN }M be a set of M observed signals of dimension N. The
                m=1
                 goal of dictionary learning is to express:

                                            Y ≈ ΦX

      where Y contains the signals along its columns, Φ is a dictionary
   containing unit norm atoms and every column of X contains at most S
                           non-zero coefficients.




                  D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Background: Dictionary Learning

  Problem Definition
  Let {ym ∈ RN }M be a set of M observed signals of dimension N. The
                m=1
                 goal of dictionary learning is to express:

                                              Y ≈ ΦX

      where Y contains the signals along its columns, Φ is a dictionary
   containing unit norm atoms and every column of X contains at most S
                           non-zero coefficients.

  Optimisation

                      ˆ ˆ                                                      2
                     (Φ, X) = arg min                     ||Y − ΦX||2
                                          Φ,X

                                    such that             ||xm ||0 ≤ S             ∀m

  The problem is not convex even if the               0   pseudo-norm is relaxed by the      1   norm

                    D. Barchiesi and M. D. Plumbley       Learning incoherent dictionaries
Background: Dictionary Learning Algorithms

  Optimisation Strategy
      Start from an initial dictionary Φ(0)
      Repeat for t = {1, . . . , T } iterations:
      Sparse coding : given a fixed dictionary Φ(t) , find a sparse
                    approximation X(t) with any suitable algorithm.
      Dictionary update : given X(t) , update the dictionary Φ(t+1) to
                    minimise the DL objective (possibly subject to
                    additional constraints).




                   D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Background: Dictionary Learning Algorithms

  Optimisation Strategy
      Start from an initial dictionary Φ(0)
      Repeat for t = {1, . . . , T } iterations:
      Sparse coding : given a fixed dictionary Φ(t) , find a sparse
                    approximation X(t) with any suitable algorithm.
      Dictionary update : given X(t) , update the dictionary Φ(t+1) to
                    minimise the DL objective (possibly subject to
                    additional constraints).

  Previous Work
  Methods for dictionary learning include:
      Probabilistic models [Lewicki and Sejnowski]
      Method of optimal directions ( mod) [Engan et al.]
      k-svd [Aharon et al.]
      Online learning [Mairal et al.]

                   D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Learning Incoherent Dictionaries

  Mutual Coherence
  The coherence of a dictionary expresses the similarity between atoms or
  groups of atoms in the dictionary. The mutual coherence is defined as:
                                          def
                               µ(Φ) = max φi , φj
                                                    i=j

  Results on sparse recovery link the performance of sparse approximation
  algorithms to the coherence of the dictionary. For over-complete
  approximations, low µ leads to recovery guarantees.




                  D. Barchiesi and M. D. Plumbley         Learning incoherent dictionaries
Learning Incoherent Dictionaries

  Mutual Coherence
  The coherence of a dictionary expresses the similarity between atoms or
  groups of atoms in the dictionary. The mutual coherence is defined as:
                                           def
                                µ(Φ) = max φi , φj
                                                     i=j

  Results on sparse recovery link the performance of sparse approximation
  algorithms to the coherence of the dictionary. For over-complete
  approximations, low µ leads to recovery guarantees.

  Goal
  The objective is to learn dictionaries that are both:
      Well adapted to a set of training data Y
      Mutually incoherent


                   D. Barchiesi and M. D. Plumbley         Learning incoherent dictionaries
Learning Incoherent Dictionaries

  Advantages
  Advantages of incoherent dictionaries include:
      Sub-dictionaries have low condition number and their
      (pseudo)inverse computed by many sparse approximation algorithms
      is well-posed.
      Convergence of greedy algorithms is faster for incoherent dictionaries
      (experimental results).
      Application-oriented intuitions (future work).




                   D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Learning Incoherent Dictionaries

  Advantages
  Advantages of incoherent dictionaries include:
      Sub-dictionaries have low condition number and their
      (pseudo)inverse computed by many sparse approximation algorithms
      is well-posed.
      Convergence of greedy algorithms is faster for incoherent dictionaries
      (experimental results).
      Application-oriented intuitions (future work).

  Previous Work
      Method of coherence-constrained directions ( mocod) [Sapiro et al.]
      Incoherent k-svd ( ink-svd) [Mailh´ et al.]
                                        e
      Parametric dictionary design for sparse coding [Yaghoobi et al.]



                   D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Incoherent Dictionary Learning: Previous Work
  mocod
  Uncostrained, penalised optimisation:
              ˆ ˆ
             (Φ, X) =arg min            ||Y − ΦX||2 + τ
                                                  F                      log(|xkm | + β)+
                          Φ,X
                                                                   m,n
                                                      K
                                                                                 2
                        + ζ ||G − I||2 + η
                                     F                      ||φk ||2 − 1
                                                                   2
                                                      k=1

  where the factor multiplied by τ promotes sparsity and the factors multiplied by
  ζ and η promote incoherence and unit-norm atoms.




                    D. Barchiesi and M. D. Plumbley    Learning incoherent dictionaries
Incoherent Dictionary Learning: Previous Work
  mocod
  Uncostrained, penalised optimisation:
              ˆ ˆ
             (Φ, X) =arg min            ||Y − ΦX||2 + τ
                                                  F                      log(|xkm | + β)+
                          Φ,X
                                                                   m,n
                                                      K
                                                                                 2
                        + ζ ||G − I||2 + η
                                     F                      ||φk ||2 − 1
                                                                   2
                                                      k=1

  where the factor multiplied by τ promotes sparsity and the factors multiplied by
  ζ and η promote incoherence and unit-norm atoms.

  ink-svd
  Greedy algorithm that includes a dictionary de-correlation step after a
  k-svd dictionary update:
       Find pairs of coherent atoms
       De-correlate atoms two-by-two
       Repeat until a target mutual coherence is reached

                    D. Barchiesi and M. D. Plumbley    Learning incoherent dictionaries
ipr Algorithm: constructing Grassmannian frames

  A Grassmannian frame is a dictionary with minimal mutual coherence.
                                  K −N
  For a N × K dictionary, µ ≥ N(K −1) and this bound can be reached
  only for some (N, K ) pairs.




                   D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
ipr Algorithm: constructing Grassmannian frames

  A Grassmannian frame is a dictionary with minimal mutual coherence.
                                  K −N
  For a N × K dictionary, µ ≥ N(K −1) and this bound can be reached
  only for some (N, K ) pairs.

  Iterative Projections Algorithm
      Start from an initial dictionary Φ(0)
                                                     def         T
      Calculate its Gram matrix G(0) = Φ(0) Φ(0)
      Repeat for t = {0, . . . , T − 1} iterations:
      Project Gram matrix onto the structural constraint set
                     def
                Kµ0 = {K : K = KT , diag(K) = 1, max |ki,j | ≤ µ0 }.
                                                                           i>j

      Project Gram matrix onto the spectral constraint set
                       def
                   F =         F : F = FT , eig(F) ≥ 0, rank(F) ≤ N
                                                             T
      Factorise the Gram matrix as Φ(T −1) Φ(T −1) = G(T −1)

                   D. Barchiesi and M. D. Plumbley     Learning incoherent dictionaries
ipr Algorithm: the rotation step

  Idea!
  The factorisation at the end of the iterative projection algorithm is not
  unique, since for any orthonormal matrix W
                              T
                   (WΦ) (WΦ) = ΦT WT WΦ = ΦT Φ

  Therefore, we can optimise an orthonormal matrix for the DL objective!
  This is an (improper) rotation of the dictionary Φ.




                   D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
ipr Algorithm: the rotation step

  Idea!
  The factorisation at the end of the iterative projection algorithm is not
  unique, since for any orthonormal matrix W
                              T
                   (WΦ) (WΦ) = ΦT WT WΦ = ΦT Φ

  Therefore, we can optimise an orthonormal matrix for the DL objective!
  This is an (improper) rotation of the dictionary Φ.

  Dictionary Rotation
                      ˆ
                      W = arg min                    ||Y − WΦX||F
                                  W:WT W=I

  A closed-form solution to this problem can be found by computing the
                                                 def
  svd decomposition of the covariance matrix C = ΦXYT = UΣV T
  and setting:
                                  ˆ
                                 W = VUT

                   D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Iterative Projections and Rotations algorithm

     Start from a dictionary Φ(0) returned by the dictionary update step of any
     DL algorithm.
     Repeat for t = {0, . . . , T − 1} iterations:
                                                           T
     Calculate the Gram matrix: G(t) ← Φ(t) Φ(t)
     Project Gram matrix onto the structural constraint set:

          diag(G) ← 1
          G ← Limit(G, µ0 )
     Factorise Gram matrix and project it onto the spectral constraint set

          [Q, Λ] ← evd(G)
          Λ ← Thresh(Λ, N)
          Φ ← Λ1/2 QT
     Rotate the dictionary

          C ← Y(ΦX)T
          [U, Σ, V] ← svd(C)
          W ← VUT
          Φ ← WΦ
                   D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Numerical Experiments: The SMALLBox framework


  The SMALLBox is a Matlab framework for benchmarking and developing
  dictionary learning algorithms developed by a team at Queen Mary
  University of London.
      Latest version can be downloaded from
      http://code.soundsoftware.ac.uk/
      SMALLBox integrates many third-party toolboxes such as Sparco,
      SparseLab, CVX, SPAMS, etc.
      SMALLBox provides a unique interface for different DL algorithms
      that can be used for benchmark
      The new distribution of SMALLBox allows to program add-ons to
      expand the functionalities of the framework without interfering with
      the core code.
      IncoherentDL is a SMALLBox add-on and can be used to reproduce
      some of the results presented in this talk.



                  D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Numerical Experiments: Mutual coherence vs residual norm

  Test Conditions
       Tests on a 16kHz guitar audio signal divided in N = 256 long
       overlapping blocks
      A fixed number of active atoms was chosen S = 12 (around 5% of
      the dimension N).
      A twice-overcomplete dictionary was initialised with either:
           Randomly selected samples from the training set.
           An over-complete Gabor frame.
      DL algorithms were run for 50 iterations.

  Test Objective
  The mutual coherence achieved by every learned dictionary is paired with
  the approximation error defined as:

                                                            ||Y||F
                     snr(Φ, X) = 20 log10                            .
                                                         ||Y − ΦX||F

                   D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Numerical Experiments: mocod updates, data init.

                               mutual coherence µ                                                                                        reconstruction error




        1                                                                                                     25

      0.9
                                                                                                              20
      0.8
                                                                                                              15




                                                                                                   SNR (dB)
      0.7
  µ




      0.6                                                                                                     10

      0.5
                                                                                                               5
      0.4

       0.3                                                                                  10
                                                                                              −2               0                                                               −2
                                                                                                                                                                              10
        −2                                                                                                    −2
      10                                                                        0                       10                                                                0
              0                                                               10                                         0                                               10
             10                                                    2      η                                             10                                       2
                       2                                       10                                                                    2                          10   η
                  ζ   10                                                                                                     ζ      10
                                                4      4                                                                                             4      4
                                               10     10                                                                                            10   10

                                                                       mutual coherence−reconstruction error scatter plot
                                      25



                                      20            ← µmin                                                                                       ← µmax



                                      15
                           SNR (dB)




                                      10



                                       5



                                       0
                                           0                 0.2                    0.4              0.6                      0.8               1
                                                                                          mutual coherence µ




                                      D. Barchiesi and M. D. Plumbley                                          Learning incoherent dictionaries
Numerical Experiments: mocod updates, Gabor init.
                               mutual coherence µ
                                                                                                                                         reconstruction error



        1
                                                                                                              25
      0.9
                                                                                                              20
      0.8

      0.7                                                                                                     15




                                                                                                   SNR (dB)
  µ




      0.6
                                                                                                              10
      0.5
                                                                                                               5
      0.4

       0.3                                                                                  10
                                                                                              −2
                                                                                                               0                                                               −2
        −2                                                                                                    −2                                                              10
      10                                                                        0                       10                                                                0
              0                                                               10                                         0                                               10
             10                                                                                                         10
                  ζ    2                                       10
                                                                   2
                                                                          η                                                  ζ       2                           2   η
                      10                                                                                                            10                          10
                                                4      4                                                                                             4      4
                                               10     10                                                                                            10   10

                                                                       mutual coherence−reconstruction error scatter plot
                                      25



                                      20            ← µmin                                                                                       ← µmax



                                      15
                           SNR (dB)




                                      10



                                       5



                                       0
                                           0                 0.2                    0.4              0.6                      0.8               1
                                                                                          mutual coherence µ




                                      D. Barchiesi and M. D. Plumbley                                          Learning incoherent dictionaries
Numerical Experiments: ink-svd and ipr
                                                 Data Initialisation
                     25

                     20   ← µmin                                                         ← µmax



          SNR (dB)
                     15

                     10
                                                                                  IPR
                      5                                                           INK−SVD

                      0
                          0.05         0.1                              0.5             1
                                               mutual coherence µ
                                                Gabor Initialisation
                     25

                     20   ← µmin                                                         ← µmax
          SNR (dB)




                     15

                     10
                                                                                  IPR
                      5                                                           INK−SVD
                      0
                          0.05         0.1                              0.5             1
                                               mutual coherence µ

                      D. Barchiesi and M. D. Plumbley      Learning incoherent dictionaries
Numerical Experiments: Sparse Approximation

  Test Conditions
      Matching pursuit algorithm ( mp) run for 1000 iterations on the
      following signals:
           Training set.
           Different guitar recording taken from the rwc database.
           A piano recording taken from the rwc database.
      Dictionaries with different mutual coherences were selected as
      returned from the ipr algorithm with data initialisation.

  Test Objective
  The norm of the residual in decibel defined as:

                                     20 log10 ||y − Φx||2

  is computed and averaged over the number of signals M and 10
  dictionaries resulting from independent trials of the learning algorithm.


                    D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Numerical Experiments: Training set approximation

                                                                 guitar − training signal
                                   0


                                −20
   average residual norm (dB)




                                −40


                                −60

                                             µ = 0.72
                                −80
                                             µ = 0.37
                                             µ = 0.19
                                −100         µ = 0.1
                                             µ = 0.06
                                −120
                                       0       200                 400           600                            800   1000
                                                                  number of iterations


                                           D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Numerical Experiments: Guitar approximation

                                                                    guitar − test signal
                                   0


                                −20
   average residual norm (dB)




                                −40


                                −60

                                             µ = 0.72
                                −80          µ = 0.37
                                             µ = 0.19
                                −100         µ = 0.1
                                             µ = 0.06
                                −120
                                       0       200                 400           600                            800   1000
                                                                  number of iterations


                                           D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Numerical Experiments: Piano approximation

                                                                             piano
                                   0


                                −20
   average residual norm (dB)




                                −40


                                −60
                                             µ = 0.72
                                −80          µ = 0.37
                                             µ = 0.19
                                −100         µ = 0.1
                                             µ = 0.06
                                −120
                                       0       200                 400           600                             800   1000
                                                                  number of iterations


                                           D. Barchiesi and M. D. Plumbley    Learning incoherent dictionaries
Conclusions: Possible Applications


  Morphological Component Analysis
  Morphological component analysis is a dictionary learning approach to
  classification.
      Different dictionaries are learned on morphologically dissimilar
      training sets (e.g., edges and textures, percussive and steady state
      sounds)
      A test signal is classified according to the support or magnitude of
      the coefficients of its sparse approximation (i.e., what is the best
      dictionary to represent it?)
  ipr could be used to enforce incoherence between the atoms belonging
  to different morphological components and enhance classification and
  separation performance.




                  D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Conclusions: Possible Applications


  Blind Compressed Sensing
  Blind compressed sensing generalises compressed sensing to the case of
  an unknown dictionary that generates the signals to be recovered.
      A set of observations Z is acquired through the known measurement
      matrix M. Z = MY = MΦX
      Dictionary learning is used to optimise Ψ and factorize the observed
      data as Z ≈ ΨX.
                                                                  ˆ
      The learned dictionary is factorized as the product Ψ ≈ MΦ and
                                    ˆ    ˆ
      the signals reconstructed as Y = ΦX.
  The two factorisations are not unique and strong constraints on Φ are
  assumed to correctly reconstruct the signals. ipr might be used to
  constrain the factorisations and lead to a less ambiguous solution.




                  D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Conclusions: Summary


     The ipr algorithm can be used to learn dictionaries that are both
     adapted to a training set and mutually incoherent.
     The ipr algorithm can be used as a de-correlation step in any
     dictionary learning algorithm.
     Experimental data show that ipr performed generally better than
     benchmark techniques on audio signals.
     Incoherent dictionaries are useful for sparse approximation and could
     be used in a number of potential applications.




                 D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries
Conclusions: Summary


     The ipr algorithm can be used to learn dictionaries that are both
     adapted to a training set and mutually incoherent.
     The ipr algorithm can be used as a de-correlation step in any
     dictionary learning algorithm.
     Experimental data show that ipr performed generally better than
     benchmark techniques on audio signals.
     Incoherent dictionaries are useful for sparse approximation and could
     be used in a number of potential applications.


        Thank you for your attention
           and for any question!

                 D. Barchiesi and M. D. Plumbley   Learning incoherent dictionaries

Weitere ähnliche Inhalte

Was ist angesagt?

Jarrar.lecture notes.aai.2012s.descriptionlogic
Jarrar.lecture notes.aai.2012s.descriptionlogicJarrar.lecture notes.aai.2012s.descriptionlogic
Jarrar.lecture notes.aai.2012s.descriptionlogicSinaInstitute
 
Harnessing Deep Neural Networks with Logic Rules
Harnessing Deep Neural Networks with Logic RulesHarnessing Deep Neural Networks with Logic Rules
Harnessing Deep Neural Networks with Logic RulesSho Takase
 
TopicModels_BleiPaper_Summary.pptx
TopicModels_BleiPaper_Summary.pptxTopicModels_BleiPaper_Summary.pptx
TopicModels_BleiPaper_Summary.pptxKalpit Desai
 
slides for "Supervised Model Learning with Feature Grouping based on a Discre...
slides for "Supervised Model Learning with Feature Grouping based on a Discre...slides for "Supervised Model Learning with Feature Grouping based on a Discre...
slides for "Supervised Model Learning with Feature Grouping based on a Discre...Kensuke Mitsuzawa
 
Tensor-based Models of Natural Language Semantics
Tensor-based Models of Natural Language SemanticsTensor-based Models of Natural Language Semantics
Tensor-based Models of Natural Language SemanticsDimitrios Kartsaklis
 
Formal language
Formal languageFormal language
Formal languageRajendran
 
Reduction Monads and Their Signatures
Reduction Monads and Their SignaturesReduction Monads and Their Signatures
Reduction Monads and Their SignaturesMarco Maggesi
 
Rethinking Perturbations in Encoder-Decoders for Fast Training
Rethinking Perturbations in Encoder-Decoders for Fast TrainingRethinking Perturbations in Encoder-Decoders for Fast Training
Rethinking Perturbations in Encoder-Decoders for Fast TrainingSho Takase
 
Conditional preference queries and possibilistic logic
Conditional preference queries and possibilistic logicConditional preference queries and possibilistic logic
Conditional preference queries and possibilistic logicFayçal Touazi
 
Learning Morphological Rules for Amharic Verbs Using Inductive Logic Programming
Learning Morphological Rules for Amharic Verbs Using Inductive Logic ProgrammingLearning Morphological Rules for Amharic Verbs Using Inductive Logic Programming
Learning Morphological Rules for Amharic Verbs Using Inductive Logic ProgrammingGuy De Pauw
 
Learning phoneme mappings for transliteration without parallel data
Learning phoneme mappings for transliteration without parallel dataLearning phoneme mappings for transliteration without parallel data
Learning phoneme mappings for transliteration without parallel dataAttaporn Ninsuwan
 
Compositional Distributional Models of Meaning
Compositional Distributional Models of MeaningCompositional Distributional Models of Meaning
Compositional Distributional Models of MeaningDimitrios Kartsaklis
 
A survey on parallel corpora alignment
A survey on parallel corpora alignment A survey on parallel corpora alignment
A survey on parallel corpora alignment andrefsantos
 
Jarrar: First Order Logic- Inference Methods
Jarrar: First Order Logic- Inference MethodsJarrar: First Order Logic- Inference Methods
Jarrar: First Order Logic- Inference MethodsMustafa Jarrar
 
深層意味表現学習 (Deep Semantic Representations)
深層意味表現学習 (Deep Semantic Representations)深層意味表現学習 (Deep Semantic Representations)
深層意味表現学習 (Deep Semantic Representations)Danushka Bollegala
 
Csr2011 june14 12_00_hansen
Csr2011 june14 12_00_hansenCsr2011 june14 12_00_hansen
Csr2011 june14 12_00_hansenCSR2011
 

Was ist angesagt? (20)

Jarrar.lecture notes.aai.2012s.descriptionlogic
Jarrar.lecture notes.aai.2012s.descriptionlogicJarrar.lecture notes.aai.2012s.descriptionlogic
Jarrar.lecture notes.aai.2012s.descriptionlogic
 
Harnessing Deep Neural Networks with Logic Rules
Harnessing Deep Neural Networks with Logic RulesHarnessing Deep Neural Networks with Logic Rules
Harnessing Deep Neural Networks with Logic Rules
 
Topic Models
Topic ModelsTopic Models
Topic Models
 
TopicModels_BleiPaper_Summary.pptx
TopicModels_BleiPaper_Summary.pptxTopicModels_BleiPaper_Summary.pptx
TopicModels_BleiPaper_Summary.pptx
 
slides for "Supervised Model Learning with Feature Grouping based on a Discre...
slides for "Supervised Model Learning with Feature Grouping based on a Discre...slides for "Supervised Model Learning with Feature Grouping based on a Discre...
slides for "Supervised Model Learning with Feature Grouping based on a Discre...
 
Tensor-based Models of Natural Language Semantics
Tensor-based Models of Natural Language SemanticsTensor-based Models of Natural Language Semantics
Tensor-based Models of Natural Language Semantics
 
Formal language
Formal languageFormal language
Formal language
 
Reduction Monads and Their Signatures
Reduction Monads and Their SignaturesReduction Monads and Their Signatures
Reduction Monads and Their Signatures
 
Rethinking Perturbations in Encoder-Decoders for Fast Training
Rethinking Perturbations in Encoder-Decoders for Fast TrainingRethinking Perturbations in Encoder-Decoders for Fast Training
Rethinking Perturbations in Encoder-Decoders for Fast Training
 
Fol
FolFol
Fol
 
Conditional preference queries and possibilistic logic
Conditional preference queries and possibilistic logicConditional preference queries and possibilistic logic
Conditional preference queries and possibilistic logic
 
Learning Morphological Rules for Amharic Verbs Using Inductive Logic Programming
Learning Morphological Rules for Amharic Verbs Using Inductive Logic ProgrammingLearning Morphological Rules for Amharic Verbs Using Inductive Logic Programming
Learning Morphological Rules for Amharic Verbs Using Inductive Logic Programming
 
Learning phoneme mappings for transliteration without parallel data
Learning phoneme mappings for transliteration without parallel dataLearning phoneme mappings for transliteration without parallel data
Learning phoneme mappings for transliteration without parallel data
 
Compositional Distributional Models of Meaning
Compositional Distributional Models of MeaningCompositional Distributional Models of Meaning
Compositional Distributional Models of Meaning
 
A survey on parallel corpora alignment
A survey on parallel corpora alignment A survey on parallel corpora alignment
A survey on parallel corpora alignment
 
W17 5406
W17 5406W17 5406
W17 5406
 
Jarrar: First Order Logic- Inference Methods
Jarrar: First Order Logic- Inference MethodsJarrar: First Order Logic- Inference Methods
Jarrar: First Order Logic- Inference Methods
 
深層意味表現学習 (Deep Semantic Representations)
深層意味表現学習 (Deep Semantic Representations)深層意味表現学習 (Deep Semantic Representations)
深層意味表現学習 (Deep Semantic Representations)
 
Csr2011 june14 12_00_hansen
Csr2011 june14 12_00_hansenCsr2011 june14 12_00_hansen
Csr2011 june14 12_00_hansen
 
Exempler approach
Exempler approachExempler approach
Exempler approach
 

Ähnlich wie Icml12

形式言語理論への 測度論的アプローチ
形式言語理論への 測度論的アプローチ形式言語理論への 測度論的アプローチ
形式言語理論への 測度論的アプローチRyoma Sin'ya
 
Algorithms and Complexity: Cryptography Theory
Algorithms and Complexity: Cryptography TheoryAlgorithms and Complexity: Cryptography Theory
Algorithms and Complexity: Cryptography TheoryAlex Prut
 
Topological Inference via Meshing
Topological Inference via MeshingTopological Inference via Meshing
Topological Inference via MeshingDon Sheehy
 
Cooperating Techniques for Extracting Conceptual Taxonomies from Text
Cooperating Techniques for Extracting Conceptual Taxonomies from TextCooperating Techniques for Extracting Conceptual Taxonomies from Text
Cooperating Techniques for Extracting Conceptual Taxonomies from TextUniversity of Bari (Italy)
 
Research Summary: Hidden Topic Markov Models, Gruber
Research Summary: Hidden Topic Markov Models, GruberResearch Summary: Hidden Topic Markov Models, Gruber
Research Summary: Hidden Topic Markov Models, GruberAlex Klibisz
 
Dirichlet processes and Applications
Dirichlet processes and ApplicationsDirichlet processes and Applications
Dirichlet processes and ApplicationsSaurav Jha
 
Ml mle_bayes
Ml  mle_bayesMl  mle_bayes
Ml mle_bayesPhong Vo
 
IVR - Chapter 7 - Patch models and dictionary learning
IVR - Chapter 7 - Patch models and dictionary learningIVR - Chapter 7 - Patch models and dictionary learning
IVR - Chapter 7 - Patch models and dictionary learningCharles Deledalle
 
Simple effective decipherment via combinatorial optimization
Simple effective decipherment via combinatorial optimizationSimple effective decipherment via combinatorial optimization
Simple effective decipherment via combinatorial optimizationAttaporn Ninsuwan
 
Higher dimensional image analysis using brunn minkowski theorem, convexity an...
Higher dimensional image analysis using brunn minkowski theorem, convexity an...Higher dimensional image analysis using brunn minkowski theorem, convexity an...
Higher dimensional image analysis using brunn minkowski theorem, convexity an...Alexander Decker
 
Functional analysis in mechanics 2e
Functional analysis in mechanics  2eFunctional analysis in mechanics  2e
Functional analysis in mechanics 2eSpringer
 
Functional analysis in mechanics
Functional analysis in mechanicsFunctional analysis in mechanics
Functional analysis in mechanicsSpringer
 
Tro07 sparse-solutions-talk
Tro07 sparse-solutions-talkTro07 sparse-solutions-talk
Tro07 sparse-solutions-talkmpbchina
 
CUMC talk notes v3
CUMC talk notes v3CUMC talk notes v3
CUMC talk notes v3Eeshan Wagh
 
A course on mathematical logic
A course on mathematical logicA course on mathematical logic
A course on mathematical logicSpringer
 
CommunicationComplexity1_jieren
CommunicationComplexity1_jierenCommunicationComplexity1_jieren
CommunicationComplexity1_jierenjie ren
 
Introduction of predicate logics
Introduction of predicate  logicsIntroduction of predicate  logics
Introduction of predicate logicschauhankapil
 

Ähnlich wie Icml12 (20)

形式言語理論への 測度論的アプローチ
形式言語理論への 測度論的アプローチ形式言語理論への 測度論的アプローチ
形式言語理論への 測度論的アプローチ
 
Algorithms and Complexity: Cryptography Theory
Algorithms and Complexity: Cryptography TheoryAlgorithms and Complexity: Cryptography Theory
Algorithms and Complexity: Cryptography Theory
 
H-MLQ
H-MLQH-MLQ
H-MLQ
 
Topological Inference via Meshing
Topological Inference via MeshingTopological Inference via Meshing
Topological Inference via Meshing
 
Canini09a
Canini09aCanini09a
Canini09a
 
Cooperating Techniques for Extracting Conceptual Taxonomies from Text
Cooperating Techniques for Extracting Conceptual Taxonomies from TextCooperating Techniques for Extracting Conceptual Taxonomies from Text
Cooperating Techniques for Extracting Conceptual Taxonomies from Text
 
Research Summary: Hidden Topic Markov Models, Gruber
Research Summary: Hidden Topic Markov Models, GruberResearch Summary: Hidden Topic Markov Models, Gruber
Research Summary: Hidden Topic Markov Models, Gruber
 
Dirichlet processes and Applications
Dirichlet processes and ApplicationsDirichlet processes and Applications
Dirichlet processes and Applications
 
Ml mle_bayes
Ml  mle_bayesMl  mle_bayes
Ml mle_bayes
 
IVR - Chapter 7 - Patch models and dictionary learning
IVR - Chapter 7 - Patch models and dictionary learningIVR - Chapter 7 - Patch models and dictionary learning
IVR - Chapter 7 - Patch models and dictionary learning
 
Simple effective decipherment via combinatorial optimization
Simple effective decipherment via combinatorial optimizationSimple effective decipherment via combinatorial optimization
Simple effective decipherment via combinatorial optimization
 
Higher dimensional image analysis using brunn minkowski theorem, convexity an...
Higher dimensional image analysis using brunn minkowski theorem, convexity an...Higher dimensional image analysis using brunn minkowski theorem, convexity an...
Higher dimensional image analysis using brunn minkowski theorem, convexity an...
 
Functional analysis in mechanics 2e
Functional analysis in mechanics  2eFunctional analysis in mechanics  2e
Functional analysis in mechanics 2e
 
Functional analysis in mechanics
Functional analysis in mechanicsFunctional analysis in mechanics
Functional analysis in mechanics
 
Astaño 4
Astaño 4Astaño 4
Astaño 4
 
Tro07 sparse-solutions-talk
Tro07 sparse-solutions-talkTro07 sparse-solutions-talk
Tro07 sparse-solutions-talk
 
CUMC talk notes v3
CUMC talk notes v3CUMC talk notes v3
CUMC talk notes v3
 
A course on mathematical logic
A course on mathematical logicA course on mathematical logic
A course on mathematical logic
 
CommunicationComplexity1_jieren
CommunicationComplexity1_jierenCommunicationComplexity1_jieren
CommunicationComplexity1_jieren
 
Introduction of predicate logics
Introduction of predicate  logicsIntroduction of predicate  logics
Introduction of predicate logics
 

Kürzlich hochgeladen

(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...AliaaTarek5
 
How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.Curtis Poe
 
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...Wes McKinney
 
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...panagenda
 
How to write a Business Continuity Plan
How to write a Business Continuity PlanHow to write a Business Continuity Plan
How to write a Business Continuity PlanDatabarracks
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsSergiu Bodiu
 
Assure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyesAssure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyesThousandEyes
 
Manual 508 Accessibility Compliance Audit
Manual 508 Accessibility Compliance AuditManual 508 Accessibility Compliance Audit
Manual 508 Accessibility Compliance AuditSkynet Technologies
 
The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsPixlogix Infotech
 
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024BookNet Canada
 
Connecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfConnecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfNeo4j
 
Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...
Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...
Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...Scott Andery
 
UiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPathCommunity
 
Moving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfMoving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfLoriGlavin3
 
[Webinar] SpiraTest - Setting New Standards in Quality Assurance
[Webinar] SpiraTest - Setting New Standards in Quality Assurance[Webinar] SpiraTest - Setting New Standards in Quality Assurance
[Webinar] SpiraTest - Setting New Standards in Quality AssuranceInflectra
 
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024BookNet Canada
 
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24Mark Goldstein
 
Take control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteTake control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteDianaGray10
 
Time Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsTime Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsNathaniel Shimoni
 
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxThe Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxLoriGlavin3
 

Kürzlich hochgeladen (20)

(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
 
How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.
 
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
 
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...
 
How to write a Business Continuity Plan
How to write a Business Continuity PlanHow to write a Business Continuity Plan
How to write a Business Continuity Plan
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platforms
 
Assure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyesAssure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyes
 
Manual 508 Accessibility Compliance Audit
Manual 508 Accessibility Compliance AuditManual 508 Accessibility Compliance Audit
Manual 508 Accessibility Compliance Audit
 
The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and Cons
 
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
 
Connecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfConnecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdf
 
Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...
Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...
Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...
 
UiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to Hero
 
Moving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfMoving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdf
 
[Webinar] SpiraTest - Setting New Standards in Quality Assurance
[Webinar] SpiraTest - Setting New Standards in Quality Assurance[Webinar] SpiraTest - Setting New Standards in Quality Assurance
[Webinar] SpiraTest - Setting New Standards in Quality Assurance
 
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
 
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
 
Take control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteTake control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test Suite
 
Time Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsTime Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directions
 
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxThe Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
 

Icml12

  • 1. Learning incoherent dictionaries for sparse approximation using iterative projections and rotations Daniele Barchiesi and Mark D. Plumbley Centre for Digital Music School of Electronic Engineering and Computer Science Queen Mary University of London daniele.barchiesi@eecs.qmul.ac.uk mark.plumbley@eecs.qmul.ac.uk 30th June 2012
  • 2. Overview Background Dictionary learning model and algorithms Learning incoherent dictionaries Previous work Learning incoherent dictionaries using iterative projections and rotations Constructing Grassmannian frames using iterative projections The rotation step Iterative projections and rotation algorithm Numerical experiments Incoherence results, comparison with existing methods Sparse approximation results Conclusions and future research Proposed applications Summary D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 3. Background: Dictionary Learning Problem Definition Let {ym ∈ RN }M be a set of M observed signals of dimension N. The m=1 goal of dictionary learning is to express: Y ≈ ΦX where Y contains the signals along its columns, Φ is a dictionary containing unit norm atoms and every column of X contains at most S non-zero coefficients. D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 4. Background: Dictionary Learning Problem Definition Let {ym ∈ RN }M be a set of M observed signals of dimension N. The m=1 goal of dictionary learning is to express: Y ≈ ΦX where Y contains the signals along its columns, Φ is a dictionary containing unit norm atoms and every column of X contains at most S non-zero coefficients. Optimisation ˆ ˆ 2 (Φ, X) = arg min ||Y − ΦX||2 Φ,X such that ||xm ||0 ≤ S ∀m The problem is not convex even if the 0 pseudo-norm is relaxed by the 1 norm D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 5. Background: Dictionary Learning Algorithms Optimisation Strategy Start from an initial dictionary Φ(0) Repeat for t = {1, . . . , T } iterations: Sparse coding : given a fixed dictionary Φ(t) , find a sparse approximation X(t) with any suitable algorithm. Dictionary update : given X(t) , update the dictionary Φ(t+1) to minimise the DL objective (possibly subject to additional constraints). D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 6. Background: Dictionary Learning Algorithms Optimisation Strategy Start from an initial dictionary Φ(0) Repeat for t = {1, . . . , T } iterations: Sparse coding : given a fixed dictionary Φ(t) , find a sparse approximation X(t) with any suitable algorithm. Dictionary update : given X(t) , update the dictionary Φ(t+1) to minimise the DL objective (possibly subject to additional constraints). Previous Work Methods for dictionary learning include: Probabilistic models [Lewicki and Sejnowski] Method of optimal directions ( mod) [Engan et al.] k-svd [Aharon et al.] Online learning [Mairal et al.] D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 7. Learning Incoherent Dictionaries Mutual Coherence The coherence of a dictionary expresses the similarity between atoms or groups of atoms in the dictionary. The mutual coherence is defined as: def µ(Φ) = max φi , φj i=j Results on sparse recovery link the performance of sparse approximation algorithms to the coherence of the dictionary. For over-complete approximations, low µ leads to recovery guarantees. D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 8. Learning Incoherent Dictionaries Mutual Coherence The coherence of a dictionary expresses the similarity between atoms or groups of atoms in the dictionary. The mutual coherence is defined as: def µ(Φ) = max φi , φj i=j Results on sparse recovery link the performance of sparse approximation algorithms to the coherence of the dictionary. For over-complete approximations, low µ leads to recovery guarantees. Goal The objective is to learn dictionaries that are both: Well adapted to a set of training data Y Mutually incoherent D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 9. Learning Incoherent Dictionaries Advantages Advantages of incoherent dictionaries include: Sub-dictionaries have low condition number and their (pseudo)inverse computed by many sparse approximation algorithms is well-posed. Convergence of greedy algorithms is faster for incoherent dictionaries (experimental results). Application-oriented intuitions (future work). D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 10. Learning Incoherent Dictionaries Advantages Advantages of incoherent dictionaries include: Sub-dictionaries have low condition number and their (pseudo)inverse computed by many sparse approximation algorithms is well-posed. Convergence of greedy algorithms is faster for incoherent dictionaries (experimental results). Application-oriented intuitions (future work). Previous Work Method of coherence-constrained directions ( mocod) [Sapiro et al.] Incoherent k-svd ( ink-svd) [Mailh´ et al.] e Parametric dictionary design for sparse coding [Yaghoobi et al.] D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 11. Incoherent Dictionary Learning: Previous Work mocod Uncostrained, penalised optimisation: ˆ ˆ (Φ, X) =arg min ||Y − ΦX||2 + τ F log(|xkm | + β)+ Φ,X m,n K 2 + ζ ||G − I||2 + η F ||φk ||2 − 1 2 k=1 where the factor multiplied by τ promotes sparsity and the factors multiplied by ζ and η promote incoherence and unit-norm atoms. D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 12. Incoherent Dictionary Learning: Previous Work mocod Uncostrained, penalised optimisation: ˆ ˆ (Φ, X) =arg min ||Y − ΦX||2 + τ F log(|xkm | + β)+ Φ,X m,n K 2 + ζ ||G − I||2 + η F ||φk ||2 − 1 2 k=1 where the factor multiplied by τ promotes sparsity and the factors multiplied by ζ and η promote incoherence and unit-norm atoms. ink-svd Greedy algorithm that includes a dictionary de-correlation step after a k-svd dictionary update: Find pairs of coherent atoms De-correlate atoms two-by-two Repeat until a target mutual coherence is reached D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 13. ipr Algorithm: constructing Grassmannian frames A Grassmannian frame is a dictionary with minimal mutual coherence. K −N For a N × K dictionary, µ ≥ N(K −1) and this bound can be reached only for some (N, K ) pairs. D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 14. ipr Algorithm: constructing Grassmannian frames A Grassmannian frame is a dictionary with minimal mutual coherence. K −N For a N × K dictionary, µ ≥ N(K −1) and this bound can be reached only for some (N, K ) pairs. Iterative Projections Algorithm Start from an initial dictionary Φ(0) def T Calculate its Gram matrix G(0) = Φ(0) Φ(0) Repeat for t = {0, . . . , T − 1} iterations: Project Gram matrix onto the structural constraint set def Kµ0 = {K : K = KT , diag(K) = 1, max |ki,j | ≤ µ0 }. i>j Project Gram matrix onto the spectral constraint set def F = F : F = FT , eig(F) ≥ 0, rank(F) ≤ N T Factorise the Gram matrix as Φ(T −1) Φ(T −1) = G(T −1) D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 15. ipr Algorithm: the rotation step Idea! The factorisation at the end of the iterative projection algorithm is not unique, since for any orthonormal matrix W T (WΦ) (WΦ) = ΦT WT WΦ = ΦT Φ Therefore, we can optimise an orthonormal matrix for the DL objective! This is an (improper) rotation of the dictionary Φ. D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 16. ipr Algorithm: the rotation step Idea! The factorisation at the end of the iterative projection algorithm is not unique, since for any orthonormal matrix W T (WΦ) (WΦ) = ΦT WT WΦ = ΦT Φ Therefore, we can optimise an orthonormal matrix for the DL objective! This is an (improper) rotation of the dictionary Φ. Dictionary Rotation ˆ W = arg min ||Y − WΦX||F W:WT W=I A closed-form solution to this problem can be found by computing the def svd decomposition of the covariance matrix C = ΦXYT = UΣV T and setting: ˆ W = VUT D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 17. Iterative Projections and Rotations algorithm Start from a dictionary Φ(0) returned by the dictionary update step of any DL algorithm. Repeat for t = {0, . . . , T − 1} iterations: T Calculate the Gram matrix: G(t) ← Φ(t) Φ(t) Project Gram matrix onto the structural constraint set: diag(G) ← 1 G ← Limit(G, µ0 ) Factorise Gram matrix and project it onto the spectral constraint set [Q, Λ] ← evd(G) Λ ← Thresh(Λ, N) Φ ← Λ1/2 QT Rotate the dictionary C ← Y(ΦX)T [U, Σ, V] ← svd(C) W ← VUT Φ ← WΦ D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 18. Numerical Experiments: The SMALLBox framework The SMALLBox is a Matlab framework for benchmarking and developing dictionary learning algorithms developed by a team at Queen Mary University of London. Latest version can be downloaded from http://code.soundsoftware.ac.uk/ SMALLBox integrates many third-party toolboxes such as Sparco, SparseLab, CVX, SPAMS, etc. SMALLBox provides a unique interface for different DL algorithms that can be used for benchmark The new distribution of SMALLBox allows to program add-ons to expand the functionalities of the framework without interfering with the core code. IncoherentDL is a SMALLBox add-on and can be used to reproduce some of the results presented in this talk. D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 19. Numerical Experiments: Mutual coherence vs residual norm Test Conditions Tests on a 16kHz guitar audio signal divided in N = 256 long overlapping blocks A fixed number of active atoms was chosen S = 12 (around 5% of the dimension N). A twice-overcomplete dictionary was initialised with either: Randomly selected samples from the training set. An over-complete Gabor frame. DL algorithms were run for 50 iterations. Test Objective The mutual coherence achieved by every learned dictionary is paired with the approximation error defined as: ||Y||F snr(Φ, X) = 20 log10 . ||Y − ΦX||F D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 20. Numerical Experiments: mocod updates, data init. mutual coherence µ reconstruction error 1 25 0.9 20 0.8 15 SNR (dB) 0.7 µ 0.6 10 0.5 5 0.4 0.3 10 −2 0 −2 10 −2 −2 10 0 10 0 0 10 0 10 10 2 η 10 2 2 10 2 10 η ζ 10 ζ 10 4 4 4 4 10 10 10 10 mutual coherence−reconstruction error scatter plot 25 20 ← µmin ← µmax 15 SNR (dB) 10 5 0 0 0.2 0.4 0.6 0.8 1 mutual coherence µ D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 21. Numerical Experiments: mocod updates, Gabor init. mutual coherence µ reconstruction error 1 25 0.9 20 0.8 0.7 15 SNR (dB) µ 0.6 10 0.5 5 0.4 0.3 10 −2 0 −2 −2 −2 10 10 0 10 0 0 10 0 10 10 10 ζ 2 10 2 η ζ 2 2 η 10 10 10 4 4 4 4 10 10 10 10 mutual coherence−reconstruction error scatter plot 25 20 ← µmin ← µmax 15 SNR (dB) 10 5 0 0 0.2 0.4 0.6 0.8 1 mutual coherence µ D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 22. Numerical Experiments: ink-svd and ipr Data Initialisation 25 20 ← µmin ← µmax SNR (dB) 15 10 IPR 5 INK−SVD 0 0.05 0.1 0.5 1 mutual coherence µ Gabor Initialisation 25 20 ← µmin ← µmax SNR (dB) 15 10 IPR 5 INK−SVD 0 0.05 0.1 0.5 1 mutual coherence µ D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 23. Numerical Experiments: Sparse Approximation Test Conditions Matching pursuit algorithm ( mp) run for 1000 iterations on the following signals: Training set. Different guitar recording taken from the rwc database. A piano recording taken from the rwc database. Dictionaries with different mutual coherences were selected as returned from the ipr algorithm with data initialisation. Test Objective The norm of the residual in decibel defined as: 20 log10 ||y − Φx||2 is computed and averaged over the number of signals M and 10 dictionaries resulting from independent trials of the learning algorithm. D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 24. Numerical Experiments: Training set approximation guitar − training signal 0 −20 average residual norm (dB) −40 −60 µ = 0.72 −80 µ = 0.37 µ = 0.19 −100 µ = 0.1 µ = 0.06 −120 0 200 400 600 800 1000 number of iterations D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 25. Numerical Experiments: Guitar approximation guitar − test signal 0 −20 average residual norm (dB) −40 −60 µ = 0.72 −80 µ = 0.37 µ = 0.19 −100 µ = 0.1 µ = 0.06 −120 0 200 400 600 800 1000 number of iterations D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 26. Numerical Experiments: Piano approximation piano 0 −20 average residual norm (dB) −40 −60 µ = 0.72 −80 µ = 0.37 µ = 0.19 −100 µ = 0.1 µ = 0.06 −120 0 200 400 600 800 1000 number of iterations D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 27. Conclusions: Possible Applications Morphological Component Analysis Morphological component analysis is a dictionary learning approach to classification. Different dictionaries are learned on morphologically dissimilar training sets (e.g., edges and textures, percussive and steady state sounds) A test signal is classified according to the support or magnitude of the coefficients of its sparse approximation (i.e., what is the best dictionary to represent it?) ipr could be used to enforce incoherence between the atoms belonging to different morphological components and enhance classification and separation performance. D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 28. Conclusions: Possible Applications Blind Compressed Sensing Blind compressed sensing generalises compressed sensing to the case of an unknown dictionary that generates the signals to be recovered. A set of observations Z is acquired through the known measurement matrix M. Z = MY = MΦX Dictionary learning is used to optimise Ψ and factorize the observed data as Z ≈ ΨX. ˆ The learned dictionary is factorized as the product Ψ ≈ MΦ and ˆ ˆ the signals reconstructed as Y = ΦX. The two factorisations are not unique and strong constraints on Φ are assumed to correctly reconstruct the signals. ipr might be used to constrain the factorisations and lead to a less ambiguous solution. D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 29. Conclusions: Summary The ipr algorithm can be used to learn dictionaries that are both adapted to a training set and mutually incoherent. The ipr algorithm can be used as a de-correlation step in any dictionary learning algorithm. Experimental data show that ipr performed generally better than benchmark techniques on audio signals. Incoherent dictionaries are useful for sparse approximation and could be used in a number of potential applications. D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries
  • 30. Conclusions: Summary The ipr algorithm can be used to learn dictionaries that are both adapted to a training set and mutually incoherent. The ipr algorithm can be used as a de-correlation step in any dictionary learning algorithm. Experimental data show that ipr performed generally better than benchmark techniques on audio signals. Incoherent dictionaries are useful for sparse approximation and could be used in a number of potential applications. Thank you for your attention and for any question! D. Barchiesi and M. D. Plumbley Learning incoherent dictionaries