From Supervised to Unsupervised Computational Sensing
Ali Mousavi
Aug 12th 2019
brain vision summit
Brain
1
From Supervised to Unsupervised Computational Sensing Ali Mousavi - - PowerPoint PPT Presentation
From Supervised to Unsupervised Computational Sensing Ali Mousavi Aug 12 th 2019 brain Brain vision summit 1 Collaborators Rich Baraniuk Arian Maleki Rice University Columbia University Chris Metzler Reinhard Heckel Gautam Dasarathy 2
1
2
Reinhard Heckel Rice University
Chris Metzler Stanford University Gautam Dasarathy Arizona StateUniversity
Subject
Simpler Hardware Measurements
Computation
Software
Subject Expensive Hardware
3
4
Subject Simpler Hardware Measurements Computational Software Recovered Subject
5
Subject
Simpler Hardware Measurements
Computation
Software
×
= =
× ×
=
Underdetermined Overdetermined Determined
6
Subject
Simpler Hardware Measurements
Computation
Software
7
8
9
Until Convergence
10
11
12
13
14
[Donoho, Maleki, Montanari 2009] y = Φx
Φ|y
C
η ( Φ| y )
xo
15
[Donoho, Maleki, Montanari 2009] y = Φx
Φ|y
C
η ( Φ| y )
xo
Residual
16
[Donoho, Maleki, Montanari 2009] y = Φx
Φ|y
C
η ( Φ| y )
xo
Residual
Gradient Step
17
[Donoho, Maleki, Montanari 2009] y = Φx
Φ|y
C
η ( Φ| y )
xo
Residual
Gradient Step Projection Operator
18
−1 −0.5 0.5 1 −0.5 0.5
Soft Thresholding
[Donoho, Maleki, Montanari 2009] y = Φx
Φ|y
C
η ( Φ| y )
xo
Residual
Gradient Step Projection Operator
19
[Donoho, Maleki, Montanari 2009]
y = Φx
Φ|y
C
η ( Φ| y )
xo
M ⌧ N
20
[Metzler, Maleki, Baraniuk 2015]
y = Φx
Φ|y
D ( Φ
|
y )
C
xo
M ⌧ N
21
Initial Estimate Calculate the Residual Update the Estimate
Until Convergence
Initial Estimate Updated Residual Updated Estimate Updated Residual Updated Estimate [Gregor and LeCun, 2010]
22
Φ Φ
Φ| Φ|
xl+1 = Dl(xl + Φ|zl) zl = y − Φxl + 1 δ zl−1 ⌦ divDl(xl−1 + Φ|zl−1) ↵
[Metzler, Mousavi, Baraniuk, NIPS 2017] 23
End-to-End Training Layer-by-Layer Training
L1 L2 L3 L4 L5
L1 L1 L1 L1 L2 L2 L2 L3 L3 L4 L1 L2 L3 L4 L5
Denoiser-by-Denoiser Training
L1 L2 L3 L4 L5 D1, D2, . . . , Dq
24
End-to-End Training Layer-by-Layer Training
L1 L2 L3 L4 L5
L1 L1 L1 L1 L2 L2 L2 L3 L3 L4 L1 L2 L3 L4 L5
Denoiser-by-Denoiser Training
L1 L2 L3 L4 L5 D1, D2, . . . , Dq
Layer-by-layer training of LDAMP is MMSE optimal.
[Metzler, Mousavi, Baraniuk, NIPS 2017]
Denoiser-by-denoiser training of LDAMP is MMSE optimal.
[Metzler, Mousavi, Baraniuk, NIPS 2017] 25
End-to-End Training Layer-by-Layer Training
L1 L2 L3 L4 L5
L1 L1 L1 L1 L2 L2 L2 L3 L3 L4 L1 L2 L3 L4 L5
Denoiser-by-Denoiser Training
L1 L2 L3 L4 L5 D1, D2, . . . , Dq
Layer-by-layer training of LDAMP is MMSE optimal.
[Metzler, Mousavi, Baraniuk, NIPS 2017]
Denoiser-by-denoiser training of LDAMP is MMSE optimal.
[Metzler, Mousavi, Baraniuk, NIPS 2017]
Average PSNR (dB) of one hundred 40x40 images Recovered from i.i.d Gaussian Measurements
more generalizable.
degrades the performance.
26
Original Image TVAL3 (26.4 dB, 6.85 sec) BM3D-AMP (27.2 dB, 75.04 sec) LDAMP (28.1 dB, 1.22 sec)
512x512 images, 20x undersampling, noiseless measurements
27
M N
x ky Φxk2 2 + λ ⇥ f(x)
28
29
M N
30
M N
x ky Φxk2 2 + λ ⇥ f(x)
Supervised
31
M N
x ky Φxk2 2 + λ ⇥ f(x)
Unsupervised
32
33
34
MSE SURE
35
Original Noisy Image BM3D (26.0 dB, 4.01 sec.) DnCNN SURE (26.5 dB, 0.04 sec.)DnCNN MSE (26.7 dB, 0.04 sec.)
36
Image: Measurements: Measurement Operator: Noise: Setting:
=
M N
37
SURE MSE zk = y − Φxk + 1 mzk−1divDk−1
θk−1(xk−1 + Φ∗zk−1)
σk = kzkk2 pm
xk+1 = Dk
θk(xk + Φ∗zk)
38
L1 L1 L1 L1 L2 L2 L2 L3 L3 L4 L1 L2 L3 L4 L5
Layer by Layer Training
[Donoho et al. 2009, 2011] [Bayati and Montanari, 2011]
Original Image BM3D-AMP (31.3 dB, 13.2 sec.) LDAMP MSE (34.6 dB, 0.4 sec.) LDAMP SURE (31.9 dB, 0.4 sec.)
5x undersampling
39
for signal acquisition.
paradigms.
Sampling Modeling Reconstruction Nyquist Rate (~1900) Compressive Sensing (~2007) Our Work 40