stable and efficient representation learning with
play

Stable and Efficient Representation Learning with Nonnegativity - PowerPoint PPT Presentation

Stable and Efficient Representation Learning with Nonnegativity Constraints Tsung-Han Lin and H.T. Kung Unsupervised Representation Learning Layer 3 Representation Layer 2 Representation Encoding Sparse encoder Layer 1 Representation


  1. Stable and Efficient Representation Learning with Nonnegativity Constraints Tsung-Han Lin and H.T. Kung

  2. Unsupervised Representation Learning Layer 3 Representation Layer 2 Representation Encoding Sparse encoder Layer 1 Representation (e.g., l1-regularized sparse coding) Large Dictionary Dictionary

  3. Why Sparse Representations? • Prior knowledge is better encoded into sparse representations – Data is explained by only a few underlying factors – Representations are more linearly separable Feature A Simplifies supervised classifier training: sparse representations work well even when labeled samples are few Feature B

  4. Computing Sparse Representations Sparse approximation: + = 0.5 × 0.3 ×

  5. Computing Sparse Representations Sparse approximation: • L1 relaxation approach: good classification accuracy, but computation is expensive • Greedy approach (e.g., orthogonal matching pursuit): fast, but yields suboptimal classification accuracy CIFAR-10 classification with single-layer architecture L1-regularized OMP Classification 78.7 76.0 accuracy (%) [Coates 2011]

  6. Major Findings • Weak stability is the key to OMP’s suboptimal performance • By allowing only additive features (via nonnegativity constraints), classification with OMP delivers higher accuracy by large margins • Competitive classification accuracy with deep neural networks

  7. Stability of Representations Data Input Encoder ? + n

  8. Orthogonal Matching Pursuit (OMP) Select k atoms from a dictionary D that minimize | x-Dz | d 2 Select the atom that has the largest d 3 correlation with the residual x d 1 Support set d 1 k

  9. Orthogonal Matching Pursuit (OMP) Select k atoms from a dictionary D that minimize | x-Dz | d 2 Select the atom that has the largest d 3 correlation with the residual r (1) x x d 1 Estimate the coefficients of the Dz (1) selected atoms by least squares Update the residual using current estimate Support set d 1 k

  10. Orthogonal Matching Pursuit (OMP) Select k atoms from a dictionary D that minimize | x-Dz | d 2 Select the atom that has the largest d 3 correlation with the residual r (1) x d 1 Estimate the coefficients of the Dz (1) selected atoms by least squares Update the residual using current estimate Support set d 1 d 3 k

  11. 1. Larger region for “+d 1 ” noise tolerance d 1 d 1 residual residual n n d 2 d 2 δ “-d 2 ” 2. Terminate without overfitting OMP Nonnegative OMP Use only additive features by constraining the atoms and coefficients to be nonnegative

  12. Allowing Only Additive Features = + Cancellation

  13. Allowing Only Additive Features = + Enforce nonnegativity to eliminate cancellation On input: On dictionary: • Any nonnegative sparse 3 coding algorithms “+” • We use spherical K-means 0 channel 3 0 -2 On representation: 0 -1 “−” Encode with nonnegative • 2 channel OMP (NOMP) Sign splitting 1

  14. Evaluate the Stability of Representations Feature dictionary learned from image datasets Representation A Grating A Rotate by some Encode by Measure change by small angle δ OMP/NOMP their correlation Grating B Representation B Correlation between representation A and B Encoder Rotation angle δ 0 0.01π 0.02π 0.03π 0.04π OMP 1 0.71 0.54 0.43 0.34 NOMP 1 0.92 0.80 0.68 0.57

  15. Classification: NOMP vs OMP NOMP has ~3% Classification accuracy on CIFAR-10 improvement over OMP

  16. NOMP Outperforms When Fewer Labeled Samples Are Available Classification accuracy on CIFAR-10 with fewer labeled training samples

  17. STL-10: 10 classes, 100 labeled samples/class, 96x96 images airplane, bird, 67.9% 64.5% car, cat, deer, dog, horse, monkey, ship, Hierarchical matching This work truck pursuit (2012) CIFAR-100: 100 classes, 500 labeled samples/class, 32x32 images aquatic mammals, fish, flowers, food containers, fruit and vegetables, household electrical devices, 60.1% 61.4% household furniture, insects, large carnivores, large man-made outdoor things, large natural outdoor scenes, Maxout network (2013) This work large omnivores and herbivores, medium-sized mammals, non-insect invertebrates, people, reptiles, small mammals, trees, vehicles

  18. Conclusion • Greedy sparse encoder is useful, giving a scalable unsupervised representation learning pipeline that attains state-of-the-art classification performance • Proper choice of encoder is critical: the stability of encoder is a key to the quality of representations

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend