learning ancestral atom of structured dictionary via
play

Learning ancestral atom of structured dictionary via sparse coding - PowerPoint PPT Presentation

Learning ancestral atom of structured dictionary via sparse coding Bernoulli Society Satellite Meeting Noboru Murata Waseda University 2 September, 2013 joint work with Toshimitsu Aritake, Hideitsu Hino 1 / 25 sparse coding a methodology


  1. Learning ancestral atom of structured dictionary via sparse coding Bernoulli Society Satellite Meeting Noboru Murata Waseda University 2 September, 2013 joint work with Toshimitsu Aritake, Hideitsu Hino 1 / 25

  2. sparse coding a methodology for representing observations with a sparse combination of basis vectors (atoms) related with various problems: associative memory (Palm, 1980) visual cortex model (Olshausen & Field, 1996) Lasso (least absolute shrinkage and selection operator; Tibshirani, 1996) compressive sensing (Cand` es & Tao, 2006) image restoration/compression (Elad et al., 2005) 2 / 25

  3. basic problem y = ( y 1 , . . . , y n ) T : target signal d = ( d 1 , . . . , d n ) T : atom D = ( d 1 , . . . , d m ) : dictionary (redundant: m > n ) x = ( x 1 , . . . , x m ) T : coefficient vector objective: ∥ y − D x ∥ 2 minimize 2 + η ∥ x ∥ ∗ x where ∥ · ∥ ∗ is a sparse norm, e.g. ℓ 0 and ℓ 1 . 3 / 25

  4. dictionary design dictionary determines overall quality of reconstruction. predefined dictionary: structured wavelets (Daubechies, 1992) curvelets (Cand` es & Donoho, 2001) contourlets (Do & Vetterli, 2005) learned dictionary: unstructured gradient-based method (Olshausen & Field, 1996) Method of Optimal Directions (Engan et al., 1999) K-SVD (Aharon et al., 2006) structured dictionary learning: intermediate Image Signature Dictionary (Aharon & Elad, 2008) Double Sparsity (Rubinstein et al., 2010) 4 / 25

  5. structured dictionary learning ordered dictionary: D = ( d λ ; λ ∈ Λ) typical approaches: meta dictionary: ˜ D = ( ˜ d 1 , . . . , ˜ d M ) d λ = ˜ D α λ where α λ is a meta-coefficient vector additional constraints are imposed on α λ , e.g. sparsity ancestral atom (ancestor): a = ( a 1 , . . . , a N ) T d λ = F λ a where F λ is an extraction operator 5 / 25

  6. dictionary generation structure: designed by a set of extraction operators extraction operator: F p,q d p,q = F p,q a , ( p : scale or downsample level , q : shift ) cut off a piece of ancestor generating operator: D D a = ( d p,q ; ( p, q ) ∈ Λ) = ( F p,q a ; ( p, q ) ∈ Λ) a structured collection of F p,q 6 / 25

  7. example of extraction operators j = ( i − 1) × 2 p + q, { 1 [ F p,q ] ij = 0 otherwise , 7 / 25

  8. example of extraction operators   1 0 0 0 0 0 · · · 0 0 1 0 0 0 0 · · · 0     0 0 1 0 0 0 · · · 0 ∈ ℜ n × N F 0 , 1 =    . . . . . .  ... . . . . . .   . . . . . · · · .   0 0 0 0 1 0 · · · 0  1 0 0 0 0 0 · · · 0  0 0 1 0 0 0 · · · 0     0 0 0 0 1 0 · · · 0 ∈ ℜ n × N F 1 , 1 =    . . . . . .  ... . . . . . .   . . . . . · · · .   0 0 0 0 0 · · · · · · 0 8 / 25

  9. basic idea projection-based algorithm related spaces: D : space of dictionary S : space of structured dictionary A : space of ancestor related maps: dictionary learning: D → D structured dictionary: A → S ⊂ D introduce a fiber bundle structure to D by defining projection from D to S 9 / 25

  10. ancestor aggregation condition: ∑ F T G = p,q F p,q : A → A is bijective ( p,q ) ∈ Λ where F T p,q is the adjoint operator of F p,q mean operator: M : D → A M D = M ( d p,q ; ( p, q ) ∈ Λ) = G − 1 ∑ F T p,q d p,q p,q important relations: π = D ◦ M : D → S (projection) Id = M ◦ D : A → A (identity) 10 / 25

  11. relation of operators 11 / 25

  12. algorithm procedure Ancestor Learning ( a (0) , D , M , U , ε > 0 ) repeat D ( t ) ← D a ( t ) ▷ generate dictionary D ( t +1) ← U D ( t ) ˜ ▷ update dictionary a ( t +1) ← M ˜ D ( t +1) ▷ update ancestor until ∥ a ( t +1) − a ( t ) ∥ < ε return a end procedure 12 / 25

  13. experiment U : OMP + K-SVD ( + renormalization) OMP: greedy algorithm for ℓ 0 norm SC K-SVD: k -means algorithm for dictionary learning compare with ISD (Aharon & Elad, 2008) gradient-based algorithm for estimating ancestor including only shift operation in original version 13 / 25

  14. artificial data ancestor (ground truth) subset of observations 14 / 25

  15. estimated ancestors proposed method ISD noiseless case 15 / 25

  16. estimated ancestors proposed method ISD noisy case 16 / 25

  17. (a-0) examples of atom in level 0 (proposed) (c-0) spectrum of level 0 (proposed) 17 / 25

  18. (a-0) examples of atom in level 0 (proposed) (b-0) examples of atom in level 0 (ISD) (c-0) spectrum of level 0 (proposed) (d-0) spectrum of level 0 (ISD) 18 / 25

  19. (a-1) examples of atom in level 1 (proposed) (b-1) examples of atom in level 1 (ISD) (c-1) spectrum of level 1 (proposed) (d-1) spectrum of level 1 (ISD) 19 / 25

  20. (a-2) examples of atom in level 2 (proposed) (b-2) examples of atom in level 2 (ISD) (c-2) spectrum of level 2 (proposed) (d-2) spectrum of level 2 (ISD) 20 / 25

  21. images (2D atoms) training image test image (peppers) 21 / 25

  22. proposed method ISD estimated ancestor 22 / 25

  23. proposed method ISD spectrum of estimated ancestor 23 / 25

  24. proposed method ISD learned dictionary 24 / 25

  25. concluding remarks we proposed a dictionary generation scheme from an ancestor a condition of structured dictionary identifiabilty a projection-based algorithm to learn the ancestor possible application would be image analysis and compression frequency analysis of signals 25 / 25

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend