lista theoretical linear convergence practical weights
play

LISTA: Theoretical Linear Convergence, Practical Weights and - PowerPoint PPT Presentation

LISTA: Theoretical Linear Convergence, Practical Weights and Thresholds Xiaohan Chen , Jialin Liu , Zhangyang Wang , and Wotao Yin TAMU, CSE UCLA, Math NeurIPS18 Overview Recover sparse x from b := Ax + white


  1. LISTA: Theoretical Linear Convergence, Practical Weights and Thresholds Xiaohan Chen ⋆ , Jialin Liu † , Zhangyang Wang ⋆ , and Wotao Yin † ⋆ TAMU, CSE † UCLA, Math NeurIPS’18

  2. Overview Recover sparse x ∗ from b := Ax ∗ + white noise Our methods improve on LISTA (Gregor&LeCun’10) and related work by � learning fewer parameters (faster training) � adding support detection (faster recovery) � proving linear convergence and robustness (theoretical guarantee)

  3. Review: ISTA and LISTA ISTA (iterative soft thresholding) x ( k +1) = SoftThreshold θ � x ( k ) + αA T ( b − Ax ( k ) ) � . α, θ are chosen by hand or cross validation.

  4. Review: ISTA and LISTA ISTA (iterative soft thresholding) x ( k +1) = SoftThreshold θ � x ( k ) + αA T ( b − Ax ( k ) ) � . α, θ are chosen by hand or cross validation. LISTA (Learned ISTA) x ( k +1) = SoftThreshold θ k � 2 x ( k ) � W k 1 b + W k . θ k , W k 1 , W k 2 are chosen by stochastic optimization � E x ⋆ ,b � x K ( b ) − x ⋆ � 2 � minimize { θ k ,W k 1 ,W k 2 } using synthesized ( x ⋆ , b ) obeying b = Ax ⋆ + white noise.

  5. Review: ISTA and LISTA ISTA (iterative soft thresholding) x ( k +1) = SoftThreshold θ � x ( k ) + αA T ( b − Ax ( k ) ) � . α, θ are chosen by hand or cross validation. LISTA (Learned ISTA) x ( k +1) = SoftThreshold θ k � 2 x ( k ) � W k 1 b + W k . θ k , W k 1 , W k 2 are chosen by stochastic optimization � E x ⋆ ,b � x K ( b ) − x ⋆ � 2 � minimize { θ k ,W k 1 ,W k 2 } using synthesized ( x ⋆ , b ) obeying b = Ax ⋆ + white noise. Compare: ISTA is slow, no training. LISTA is fast, difficult-to-train.

  6. Proposed — coupled LISTA LISTA-CP : couple W k 1 and W k 2 via W k 1 A + W k 2 = I. We show: x ( k ) → x ⋆ implies this relation to hold asymptotically. 3 2.5 2 1.5 1 0.5 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16

  7. Proposed — support selection LISTA-CPSS : support selection � Only the large coordinates pass activations to the next iteration. � Ideas from Linearized Bregman iteration (kicking) 1 and Fixed-Point Continuation method (FPC) 2 . 1 Stanley Osher et al. ’2011 2 Elaine Hale, Wotao Yin, Yin Zhang ’2008

  8. Robust global linear convergence Theorem Fix A , sparsity level s , and noise level σ . There exist { θ k , W k 1 } such that LISTA-CP obeys � x ( k ) − x ⋆ � 2 ≤ sC 1 e − C 2 k + C 3 σ, k = 1 , 2 , . . . where C 1 , C 2 , C 3 > 0 are constants. LISTA-CPSS improves the constants C 2 , C 3 .

  9. Weight coupling test 0 -10 -20 -30 -40 ISTA AMP LISTA-CP FISTA LISTA -50 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 � CP can stabilize intermediate results. � CP will not hurt final recovery performance.

  10. Support selection test (no noise) 0 -10 -20 -30 ← LISTA and LISTA-CP -40 -50 ISTA LAMP ← LISTA-SS FISTA LISTA-CP -60 AMP LISTA-SS ← LISTA-CPSS LISTA LISTA-CPSS -70 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16

  11. Thank you! 10:45 AM – 12:45 PM Room 210 & 230 AB #163 Welcome to our poster for more details!

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend