pde s on the space of patches for image denoising and
play

PDEs on the Space of Patches for Image Denoising and Registration - PowerPoint PPT Presentation

PDEs on the Space of Patches for Image Denoising and Registration David Tschumperl Luc Brun - Patch-based Image Representation, Manifolds and Sparsity , Rennes/France, April 2009. GREYC IMAGE (CNRS UMR 6072), Caen/France


  1. PDE’s on the Space of Patches for Image Denoising and Registration David Tschumperlé ⋆ Luc Brun ⋆ - Patch-based Image Representation, Manifolds and Sparsity , Rennes/France, April 2009. ⋆ GREYC IMAGE (CNRS UMR 6072), Caen/France

  2. Presentation Layout • Definition of a Patch Space Γ . • Patch-based Tikhonov Regularization. • Patch-based Anisotropic Diffusion PDE’s. • Patch-based Lucas-Kanade registration. • Conclusions & Perspectives.

  3. Presentation Layout ⇒ Definition of a Patch Space Γ . • Patch-based Tikhonov Regularization. • Patch-based Anisotropic Diffusion PDE’s. • Patch-based Lucas-Kanade registration. • Conclusions & Perspectives.

  4. Located Patch of an Image • Considering a 2D image I : Ω ⊂ R 2 → R n ( n = 3 , for color images). • An image patch P I ( x,y ) is a discretized p × p neighborhood of I , which can be ordered as a np 2 -dimensional vector : P I � � ( x,y ) = I 1( x − q,y − q ) , . . . , I 1( x + q,y + q ) , I 2( x − q,y − q ) , . . . , I n ( x + q,y + q ) • We define a located patch as the ( np 2 + 2) -D vector ( x, y, λ P I ( x,y ) ) ( λ > 0 balances importance of spatial/intensity features).

  5. Space Γ of Located Patches • Γ = Ω × R np 2 defines a ( np 2 + 2) -dimensional space of located patches.

  6. Space Γ of Located Patches • Γ = Ω × R np 2 defines a ( np 2 + 2) -dimensional space of located patches. • The Euclidean distance between two points p 1 , p 2 ∈ Γ measures a spatial & intensity dissimilarity between corresponding located patches : ( x 1 − x 2 ) 2 + ( y 1 − y 2 ) 2 + λ 2 SSD ( P 1 , P 2 ) � d ( p 1 , p 2 ) = (SSD = Sum of Squared Differences)

  7. Mapping an Image I on the Patch Space Γ I : Γ → R np 2 +1 , a mapping of the image I on Γ : • We define ˜ � ( P I p = ( x, y, P I ( x,y ) , 1) ( x,y ) ) if ˜ ∀ p ∈ Γ , I ( p ) = � 0 elsewhere

  8. Mapping an Image I on the Patch Space Γ I : Γ → R np 2 +1 , a mapping of the image I on Γ : • We define ˜ � ( P I p = ( x, y, P I ( x,y ) , 1) if ( x,y ) ) ˜ ∀ p ∈ Γ , I ( p ) = � 0 elsewhere • The last value of ˜ I ( p ) models the meaningfulness of a located patch p . All patches coming from the original image I have the same unit weight. ⇒ ˜ I is a patch-based representation of I in Γ , as an implicit surface.

  9. Inverse Mapping to the Image Domain Ω • Question : Is it possible to retrieve I from ˜ I ?

  10. Inverse Mapping to the Image Domain Ω • Question : Is it possible to retrieve I from ˜ I ? YES ! ⇒ (1) Find the most significant patches p = ( x, y, P ) ∈ Γ for each location ( x, y ) ∈ Ω : ˜ sig ( x,y ) = argmax q ∈ R np 2 ˜ I P I np 2 +1 ( x, y, q )

  11. Inverse Mapping to the Image Domain Ω ⇒ (2) Get the central pixel of these patches, and normalize it by its meaningfulness : 2 ( x, y, P ˜ ˜ I sig ( x,y ) ) I ip 2 + p 2+1 ˆ ∀ ( x, y ) ∈ Ω , I i ( x,y ) = I np 2 +1 ( x, y, P ˜ ˜ I sig ( x,y ) ) (Other solutions may be considered, for instance : averaging spatially-overlapping meaningful patches).

  12. From Non-Local to Local processing • Mapping I in Γ transforms a non-local processing problem into a local one. • Local or semi-local measures of ˜ I in Γ (gradients,curvatures,...) will be related to non-local features of the original image I (patch dissimilarity, variance,...).

  13. Main Idea of this Talk ⇒ Apply local algorithms on ˜ I in order to build their patch-based counterparts . ⇒ Find correspondences between non-local and local algorithms.

  14. What Local Algorithms to Apply in Γ ? ⇒ PDE’s and variational methods are good candidates. - They are purely local or semi-local. - They are adaptive to local image informations (non-linear). - They are often expressed independently on the data dimension. - They give interesting solutions for a wide range of different (local) problems.

  15. What Local Algorithms to Apply in Γ ? ⇒ PDE’s and variational methods are good candidates. - They are purely local or semi-local. - They are adaptive to local image informations (non-linear). - They are often expressed independently on the data dimension. - They give interesting solutions for a wide range of different (local) problems. ⇒ In this talk : • Diffusion PDE’s for image denoising. • PDE’s for image registration , coming from a variational formulation.

  16. Presentation Layout • Definition of a Patch Space Γ . ⇒ Patch-based Tikhonov Regularization. • Patch-based Anisotropic Diffusion PDE’s. • Patch-based Lucas-Kanade registration. • Conclusions & Perspectives.

  17. Tikhonov Regularization in Γ • We minimize the classical Tikhonov regularization functional for ˜ I in Γ : � I ( p ) � 2 d p E (˜ �∇ ˜ I ) = Γ �� np 2 +1 where �∇ ˜ �∇ ˜ I i ( p ) � 2 I ( p ) � = i =1

  18. Tikhonov Regularization in Γ • We minimize the classical Tikhonov regularization functional for ˜ I in Γ : � I ( p ) � 2 d p E (˜ �∇ ˜ I ) = Γ �� np 2 +1 where �∇ ˜ �∇ ˜ I i ( p ) � 2 I ( p ) � = i =1 • The Euler-Lagrange equations of E give the desired minimizing flow for ˜ I : ˜ I [ t =0] = ˜  I noisy   ∂ ˜ ∂t = ∆˜  I i I i  ⇒ Heat flow in the high-dimensional space of patches Γ .

  19. Solution to the Tikhonov Regularization in Γ • This high-dimensional heat flow has an explicit solution (at time t ) : √ e − � p � 2 1 I [ t ] = ˜ I noisy ∗ G σ ˜ ∀ p ∈ Γ , G σ ( p ) = σ = 2 t. with 2 σ 2 and np 2+2 (2 πσ 2 ) 2

  20. Solution to the Tikhonov Regularization in Γ • This high-dimensional heat flow has an explicit solution (at time t ) : √ e − � p � 2 1 I [ t ] = ˜ I noisy ∗ G σ ˜ ∀ p ∈ Γ , G σ ( p ) = σ = 2 t. with 2 σ 2 and np 2+2 (2 πσ 2 ) 2 I noisy vanishes almost everywhere (except on the original • Simplification : As ˜ located patches of I ), the convolution simplifies to : � I [ t ] ˜ ˜ I noisy ( x,y, P ) = ) G σ ( p − x,q − y, P I noisy −P ) dp dq ( p,q, P I noisy Ω ( p,q ) ( p,q ) ⇒ Computing the solution does not require to build an explicit representation of the patch-based representation ˜ I .

  21. Inverse mapping of the Tikhonov Regularization in Γ • Finding the most significant patches in Γ : the flow preserves the locations of the I [ t ] on Ω is then : local maxima. The inverse mapping of ˜ Ω I noisy � ( p,q ) w ( x,y,p,q ) dp dq I [ t ] ∀ ( x, y ) ∈ Ω , ( x,y ) = � Ω w ( x,y,p,q ) dp dq �P I noisy −P I noisy � 2 2 πσ 2 e − ( x − p )2+( y − q )2 ( x,y ) ( p,q ) e − 1 1 with w ( x,y,p,q ) = × 2 σ 2 2 σ 2 np 2 (2 πσ 2 ) 2

  22. Inverse mapping of the Tikhonov Regularization in Γ • Finding the most significant patches in Γ : the flow preserves the locations of the I [ t ] on Ω is then : local maxima. The inverse mapping of ˜ Ω I noisy � ( p,q ) w ( x,y,p,q ) dp dq I [ t ] ∀ ( x, y ) ∈ Ω , ( x,y ) = � Ω w ( x,y,p,q ) dp dq �P I noisy −P I noisy � 2 2 πσ 2 e − ( x − p )2+( y − q )2 ( x,y ) ( p,q ) e − 1 1 with w ( x,y,p,q ) = × 2 σ 2 2 σ 2 np 2 (2 πσ 2 ) 2 ⇒ Variant of the NL-means algorithm (Buades-Morel:05) with an additional weight depending on the spatial distance between patches in Ω . ⇒ NL-means is an isotropic diffusion process in the space of patches Γ .

  23. Tikhonov Regularization in the Patch Space Γ

  24. (Useless) Results (Tikhonov Regularization in Γ ) Noisy color image

  25. (Useless) Results (Tikhonov Regularization in Γ ) Tikhonov regularization in the image domain Ω (= isotropic smoothing)

  26. (Useless) Results (Tikhonov Regularization in Γ ) Tikhonov regularization in the 5 × 5 patch space Γ ( ≈ Non Local-means algorithm)

  27. Presentation Layout • Definition of a Patch Space Γ . • Patch-based Tikhonov Regularization. ⇒ Patch-based Anisotropic Diffusion PDE’s. • Patch-based Lucas-Kanade registration. • Conclusions & Perspectives.

  28. Behavior of Isotropic Diffusion in Γ • Isotropic diffusion in Γ (NL-means) does not take care of the geometry of the patch mapping ˜ I : The smoothing is done homogeneously in all directions.

  29. What We Want to Do : Anisotropic Diffusion • Anisotropic diffusion would adapt the smoothing kernel to the local geometry of the patch mapping ˜ I . ⇒ This anisotropic behavior can be described with diffusion tensors .

  30. Introducing Diffusion Tensors • A second-order tensor is a symmetric and semi-positive definite p × p matrix. ( p is the dimension of the considered space). • It has p positive eigenvalues λ i and p orthogonal eigenvectors u [ i ] : λ i u [ i ] u [ i ] T � T = i

  31. Introducing Diffusion Tensors • A second-order tensor is a symmetric and semi-positive definite p × p matrix. ( p is the dimension of the considered space). • It has p positive eigenvalues λ i and p orthogonal eigenvectors u [ i ] : λ i u [ i ] u [ i ] T � T = i ( np 2 + 2) × ( np 2 + 2) Tensor 2 × 2 Tensor (e.g. in Ω ) 3 × 3 Tensor • Diffusion tensors describe how much pixel values locally diffuse along given orthogonal orientations, i.e. the “geometry” of the performed smoothing.

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend