How to use Gaussian mixture models on patches for solving image - - PowerPoint PPT Presentation

how to use gaussian mixture models on patches for solving
SMART_READER_LITE
LIVE PREVIEW

How to use Gaussian mixture models on patches for solving image - - PowerPoint PPT Presentation

How to use Gaussian mixture models on patches for solving image inverse problems Workshop MixStatSeq Antoine Houdard LTCI, Tlcom ParisTech MAP5, Universit Paris Descartes antoine.houdard@telecom-paristech.fr houdard.wp.imt.fr Joint


slide-1
SLIDE 1

How to use Gaussian mixture models on patches for solving image inverse problems

Workshop MixStatSeq Antoine Houdard

LTCI, Télécom ParisTech MAP5, Université Paris Descartes antoine.houdard@telecom-paristech.fr houdard.wp.imt.fr Joint work with C. Bouveyron & J. Delon

1 / 29

slide-2
SLIDE 2

Image restoration : solving an inverse problem

Image restoration problem :

find the clean image u from the observed degraded image v s.t. v = Φu + ǫ, with Φ degradation operator and ǫ additive noise.

Gaussian white noise case :

Here we deal with the simpler problem Φ = I and ǫ ∼ N(0, σ2I)

2 / 29

slide-3
SLIDE 3

Patch-based image denoising

most of the denoising methods rely on the description of the image by

patches (NL-means, NL-Bayes, S-PLE, LDMM, PLE, BM3D, DA3D)

« Les patchs sont aux images ce que les phonèmes sont à la chaîne parlée. » Pattern Theory, Desolneux & Mumford

3 / 29

slide-4
SLIDE 4

Patch-based image denoising

the statistical framework

We consider each clean patch xi as a realization of a random vector Xi

with some prior distribution PX

the Gaussian white noise model for patches yields

with Ni ∼ N(0, Ip).

Hypothesis : Ni and Xi are independent and the Ni’s are i.i.d. so we can write the posterior distribution with Bayes’ theorem

PX|Y (x|y) = PY |X(y|x)PX(x) PY (y) .

4 / 29

slide-5
SLIDE 5

Patch-based image denoising

denoising strategies

Denoising strategies x = E[X|Y = y] the minimum mean square error (MMSE) estimator x = Dy + α s.t. D and α minimize E[DY + α − X2] which is the linear MMSE also called Wiener estimator x = arg maxx∈Rp p(x|y) the maximum a posteriori (MAP)

5 / 29

slide-6
SLIDE 6

Patch-based image denoising

choice and inference of the model

In the literature

local Gaussian models [NL-bayes] Gaussian mixture models (GMM) [PLE, S-PLE, EPLL]

Advantages of Gaussian models and GMM

able to encode information of the patches make computation of estimators easy 6 / 29

slide-7
SLIDE 7

Patch-based image denoising

Gaussian and GMM models

The covariance matrix in Gaussian models and GMM is able to encode geometric structure in patches : Left : Covariance matrix Σ. Right : patches generated from the Gaussian model N(0, Σ).

7 / 29

slide-8
SLIDE 8

Patch-based image denoising

Gaussian and GMM models

The covariance matrix in Gaussian models and GMM is able to encode geometric structure in patches : Left : Covariance matrix Σ. Right : patches generated from the Gaussian model N(0, Σ).

7 / 29

slide-9
SLIDE 9

Restore with the right model

covariance matrix clean patch noisy patch denoised

8 / 29

slide-10
SLIDE 10

Patch-based image denoising

summary of the framework

9 / 29

slide-11
SLIDE 11

The curse of dimensionality

Parameters estimation for Gaussian models or GMMs suffers from the curse

  • f dimensionality

This term curse was first used by R. Bellman in the introduction of his book “Dynamic programming” in 1957 : All [problems due to high dimension] may be subsumed under the heading “the curse of dimensionality”. Since this is a curse, [...], there is no need to feel discouraged about the possibility of obtaining significant results despite it.

10 / 29

slide-12
SLIDE 12

The curse of dimensionality

High-dimensional spaces are empty

In high-dimensional space no one can hear you scream !

11 / 29

slide-13
SLIDE 13

The curse of dimensionality

High-dimensional spaces are empty

Neighborhoods are no more local ! Data are isolated

11 / 29

slide-14
SLIDE 14

The curse of dimensionality

In patches space

We consider patches of size p = 10 × 10 → High dimension. → the estimation of sample covariance matrices is difficult : ill conditioned, singular...

12 / 29

slide-15
SLIDE 15

The curse of dimensionality

In patches space

We consider patches of size p = 10 × 10 → High dimension. → the estimation of sample covariance matrices is difficult : ill conditioned, singular... In the literature, this issue is worked around by

the use of small patches in NL-Bayes (3 × 3 or 5 × 5) a model of mixture with fixed lower dimensions covariances in S-PLE

We propose a fully statistical model, that estimates a lower dimension for each group.

12 / 29

slide-16
SLIDE 16

Reminder : Noise model and notations

We denote

{y1, . . . , yn} ∈ Rp the (observed) noisy patches of the image ; {x1, . . . , xn} ∈ Rp the corresponding (unobserved) clean patches.

We suppose they are realizations of random variables Y and X that follow the classical degradation model :

= +

13 / 29

slide-17
SLIDE 17

Reminder : Noise model and notations

We denote

{y1, . . . , yn} ∈ Rp the (observed) noisy patches of the image ; {x1, . . . , xn} ∈ Rp the corresponding (unobserved) clean patches.

We suppose they are realizations of random variables Y and X that follow the classical degradation model :

= +

We design for X the High-Dimensional Mixture Model for Image Denoising (HDMI)

13 / 29

slide-18
SLIDE 18

The HDMI model

Model on the actual patches X. Let Z be the latent random variable

indicating the group from which the patch X has been generated. We assume that X lives in a low-dimensional subspace which is specific to its latent group : X|Z=k = UkT + µk, where Uk is a p × dk orthonormal transformation matrix and T ∈ Rdk such that T | Z = k ∼ N(0, Λk), with Λk = diag(λk

1, . . ., λk dk).

Model on the noisy patches. This implies that Y follow

p(y) =

K

  • k=1

πkg (y; µk, Σk) where πk is the mixture proportion for the kth component and Σk = UkΛkU T

k + σ2Ip.

14 / 29

slide-19
SLIDE 19

The HDMI model

The projection of the covariance matrix ∆k = QkΣkQt

k has the specific

structure : ∆k =             ak1 ... akd σ2 ... σ2                dk        (p − dk) where akj = λk

j + σ2 and akj > σ2, for j = 1, . . . , dk.

15 / 29

slide-20
SLIDE 20

The HDMI model

Y N X T Z µk, dk ak1, ..., akdk Qk π σ2

Figure – Graphical representation of the HDMI model.

16 / 29

slide-21
SLIDE 21

Denoising with the HDMI model

The HDMI model being known, each patch is denoised with the MMSE estimator

  • xi = E[X|Y = yi],

which can be computed as follow : Proposition. E[X|Y = yi] =

K

  • k=1

ψk(yi)tik, with tik the posterior probability for the patch yi to belong in the kth group and ψk(yi) = µk + Uk    

ak1−σ2 ak1

...

akdk −σ2 akdk

    U T

k (yi − µk),

17 / 29

slide-22
SLIDE 22

Model inference

EM algorithm : maximize w.r.t. θ the conditional expectation of the complete log-likelihood : Ψ(θ, θ∗)

def

=

K

  • k=1

n

  • i=1

tik log (πkg (yi; θk)) , where tik = E [z = k|yi, θ∗] and θ∗ a given set of parameters.

E-step estimation of tik knowing the current parameters M-step compute maximum likelihood estimators (MLE) for parameters :

  • πk = nk

n ,

  • µk = 1

nk

  • i

tikyi,

  • Sk = 1

nk

  • i

tik(yi − µk)(yi − µk)T , with nk =

i tik. Then

Qk is formed by the dk first eigenvectors of Sk and akj is the jth eigenvalue of Sk.

18 / 29

slide-23
SLIDE 23

Model inference

The hyper-parameters

The hyper-parameters K and d1, . . . , dK cannot be determined by maximizing the log-likelihood since they control the model complexity. We propose to set K at a given value (in the experiments we use K = 40 and K = 90) and to choose the intrinsic dimensions dk :

using an heuristic that links dk with the noise variance σ when known ; using a model selection tool in order to select the best σ when unknown. 19 / 29

slide-24
SLIDE 24

Estimation of intrinsic dimensions

when σ is known

With dk begin fixed, the MLE for the noise variance in the kth group is

  • σ2

|k =

1 p − dk

p

  • j=dk+1
  • akj.

When the noise variance σ is known, this gives us the following heuristic :

  • Heuristic. Given a value of σ2 and for k = 1, ..., K, we estimate the

dimension dk by

  • dk = argmind
  • 1

p − d

p

  • j=d+1
  • akj − σ2
  • .

20 / 29

slide-25
SLIDE 25

Estimation of intrinsic dimensions

when σ is unknown

Each value of σ yields a different model, we propose to select the one with the better BIC (Bayesian Information Criterion) BIC(M) = ℓ(ˆ θ) − ξ(M) 2 log(n), where ξ(M) is the complexity of the model. why BIC is well-adapted for the selection of σ ?

if σ is too small, the likelihood is good but the complexity explodes ; if σ is too high, the complexity is low but the likelihood is bad. 21 / 29

slide-26
SLIDE 26

Estimation of intrinsic dimensions

when σ is unknown

∆k =             ak1 ... akd σ2 ... σ2                dk        (p − dk) why BIC is well-adapted for the selection of σ ?

if σ is too small, the likelihood is good but the complexity explodes ; if σ is too high, the complexity is low but the likelihood is bad. 21 / 29

slide-27
SLIDE 27

Experiment : selection of σ with BIC

22 / 29

slide-28
SLIDE 28

Numerical experiments

Visualization of the intrinsic dimensions

We display for each pixel the dimension of the most probable group of the patch around it. clean noisy clustering dimensions map Simpson Barbara

23 / 29

slide-29
SLIDE 29

Regularizing effect of the dimension reduction

24 / 29

slide-30
SLIDE 30

Numerical Experiments

Clean image

25 / 29

slide-31
SLIDE 31

Numerical Experiments

Noisy image σ = 50

25 / 29

slide-32
SLIDE 32

Numerical Experiments

Denoised with BM3D, Foi et al. 2007, psnr = 27.17dB

25 / 29

slide-33
SLIDE 33

Numerical Experiments

Denoised with FFDNet, Zhang et al. 2018, psnr = 27.58dB

25 / 29

slide-34
SLIDE 34

Numerical Experiments

Denoised with HDMIsup K = 90, psnr = 27.28dB

25 / 29

slide-35
SLIDE 35

Numerical Experiments

Clean image

26 / 29

slide-36
SLIDE 36

Numerical Experiments

Noisy image σ = 50

26 / 29

slide-37
SLIDE 37

Numerical Experiments

Denoised with BM3D, Foi et al. 2007, psnr = 26.55.dB

26 / 29

slide-38
SLIDE 38

Numerical Experiments

Denoised with FFDNet, Zhang et al. 2018, psnr = 27.45dB

26 / 29

slide-39
SLIDE 39

Numerical Experiments

Denoised with HDMIsup K = 90, psnr = 27.05dB

26 / 29

slide-40
SLIDE 40

Numerical Experiments

PSNR HDMI vs FFDNet

26 / 29

slide-41
SLIDE 41

Numerical Experiments

Best of both worlds, psnr = 27.86dB

26 / 29

slide-42
SLIDE 42

Other inverse problem : missing pixels

70% missing pixels

EM is well-adapted for missing data → the model can be easily adapted for missing pixel restoration

27 / 29

slide-43
SLIDE 43

Other inverse problem : missing pixels

restored with HDMI

EM is well-adapted for missing data → the model can be easily adapted for missing pixel restoration

27 / 29

slide-44
SLIDE 44

Conclusion and further work

High dimensional mixtures models for patches

can model the full process of the generation of the noisy patches ; for denoising : can be used unsupervised (σ unknown) and reach

state-of-the-art performances ;

not restricted to denoising : interpolation, inpainting, image synthesis ; complementary to DL approaches : yield simple image models, easy to

interpret ; Some issues and further work

high computation time → learn the model on a subsample of the patches in the case of high σ some miss-classification can yield artifacts →

explore other initialization ?

low-frequency noise in flat areas → explore aggregation methods

(weighted, EPLL) ? Preprint available at : up5.fr/HDMI

  • r

houdard.wp.imt.fr/hdmi/

28 / 29

slide-45
SLIDE 45

Thank you for your attention ! Any question ?

Preprint available at : up5.fr/HDMI

29 / 29