Recall 1 Wavelet coefficients of images are Laplacian distributed! - - PowerPoint PPT Presentation

recall 1
SMART_READER_LITE
LIVE PREVIEW

Recall 1 Wavelet coefficients of images are Laplacian distributed! - - PowerPoint PPT Presentation

Recall 1 Wavelet coefficients of images are Laplacian distributed! The various wavelet coefficients are not statistically independent. Large wavelet coefficients tend to occur near each within the same sub-band. And at the same relative


slide-1
SLIDE 1

Recall 1

  • Wavelet coefficients of images are Laplacian

distributed!

  • The various wavelet coefficients are not statistically

independent.

Ajit Rajwade 1

Large wavelet coefficients tend to occur near each within the same sub-band. And at the same relative spatial locations in sub-bands at adjacent scales or

  • rientations
slide-2
SLIDE 2

Wavelet coefficient dependency

Ajit Rajwade 2

Image source: Buccigrossi et al, Image Compression via Joint Statistical Characterization in the Wavelet Domain, IEEE Transactions on Image Processing, 1997

slide-3
SLIDE 3

Wavelet coefficient dependency

  • The conditional density of the child wavelet coefficient (c)

given the parent (p) (figure 6A,B two slides before) reveals: 1. E(c|p) = 0 for all values of p. 2. They are not independent statistically – because the variance of c depends on the value of p. 3. The right side of the conditional density of the log of the squared coefficient is unimodal and concentrated on a unit slope line. 4. Left side shows c being constant (not dependent on p).

  • This pattern is also observed for siblings (adjacent spatial

locations), cousins (same spatial location, adjacent

  • rientations).
  • This pattern is robust across a wide range of images.

Ajit Rajwade 3

slide-4
SLIDE 4

Wavelet coefficient dependency

  • So how do we model this mathematically?
  • Here is one statistical model:

Ajit Rajwade 4

Neighbors of coefficient c (some cousins & siblings, parent)

2 2 2

  

k k k p

w c

Can be obtained by least squares method

slide-5
SLIDE 5

Ajit Rajwade 5

2 2 2

  

k k k p

w c

Can be obtained by least squares method

1 ) 1 ( ) 1 ( 1 2 2 1 1 2 2 2 2

) ˆ ˆ ( ˆ ˆ ˆ , ˆ , ˆ ˆ 1 ) ( c : have we , in Bringing neighbors) K assuming ( , , c ) (1 t coefficien for index ,

       

                         

T T n K K n K K n ik k k i

P P P c w R P R w P w P w R P R w R c wP n i i p w c    

Least squares estimate of w

slide-6
SLIDE 6

Wavelet coefficient dependency

Ajit Rajwade 6

slide-7
SLIDE 7

Application to denoising

  • We have already seen the formula:
  • The same formula is applicable here with the

following modification:

Ajit Rajwade 7

i T i T i

y U y U ) ( ˆ 1 ) ( ˆ

2 2 2

           n U n x y     

2 2 2

   

k k k p

w 

i T k k k k k k i

y U p w p w ) ( ˆ

2 2 2 2 2

 

        

slide-8
SLIDE 8

Application to denoising

  • But this is a chicken and egg problem – because we

do not know the values of {wk} or α beforehand!

  • But we can estimate these values by minimizing:
  • The values of {pk} are obtained from a denoising

algorithm that ignores wavelet coefficient dependency, e.g. using and then used for estimating {wk} or α.

Ajit Rajwade 8 i T i

y U ) ( ˆ

2 2 2

       

slide-9
SLIDE 9

Sample results

Ajit Rajwade 9

(Left) Original and (Right) noisy image (Left) Marginal MAP with independent Gaussian prior and (Right) new model MMSE estimator using GGD prior

  • n wavelet coefficients
slide-10
SLIDE 10

Matrix Completion

CS 754 Lecture Notes

Ajit Rajwade 10

slide-11
SLIDE 11

Matrix Completion in Practice: Scenario 1

  • Consider a survey of m people where each is

asked q questions.

  • It may not be possible to ask each person all q

questions.

  • Consider a matrix of size m by q (each row is the

set of questions asked to any given person).

  • This matrix is only partially filled (many missing

entries).

  • Is it possible to infer the full matrix given just

the recorded entries?

slide-12
SLIDE 12

Matrix Completion in Practice: Scenario 2

  • Some online shopping sites such as Amazon, Flipkart, Ebay, Netflix
  • etc. have recommender systems.
  • These websites collect product ratings from users (especially

Netflix).

  • Based on user ratings, these websites try to recommend other

products/movies to the user that he/she will like with a high probability.

  • Consider a matrix with the number of rows equal to the number of

users, and number of columns equal to the number of movies/products.

  • This matrix will be HIGHLY incomplete (no user has the patience to

rate too many movies!!) – maybe only 5% of the entries will be filled up.

  • Can the recommender system infer user preferences from just the

defined entries?

slide-13
SLIDE 13

Matrix Completion in Practice: Scenario 2

  • Read about the Netflix Prize to design a better

recommender system: http://en.wikipedia.org/wiki/Netflix_Prize

slide-14
SLIDE 14

Matrix Completion in Practice: Scenario 3

  • Consider an image or a video with several pixel values

missing.

  • This is not uncommon in range imagery or remote

sensing applications!

  • Consider a matrix whose each column is a (vectorized)

patch of m pixels. Let the number of columns be K.

  • This m by K matrix will have many missing entries.
  • Is it possible to infer the complete matrix given just

the defined pixel values?

  • If the answer were yes, note the implications for image

compression!

slide-15
SLIDE 15

Matrix Completion in Practice: Scenario 4

  • Consider a long video sequence of F frames.
  • Suppose I mark out M salient (interesting points) {Pi},

1<=i<=m, in the first frame.

  • And track those points in all subsequent frames.
  • Consider a matrix M of size m x 2F where row j contains the

X and Y coordinates of points on the motion trajectory of initial point Pj (in each of the F frames).

  • Unfortunately, many salient points may not be trackable

due to occlusion or errors from the tracking algorithms.

  • So M is highly incomplete.
  • Is it possible to infer the true matrix from only the

available measurements?

slide-16
SLIDE 16

A property of these matrices

  • Scenario 1: Many people will tend to give very

similar or identical answers to many survey questions.

  • Scenario 2: Many people will have similar

preferences for movies (only a few factors affect user choices).

  • Scenario 3: Non-local self-similarity!
  • This makes the matrices in all these scenarios

(approximately) low in rank!

slide-17
SLIDE 17

A property of these matrices

  • Scenario 4: The true matrix underlying M in question

has been PROVED to be of low rank (in fact, rank 3) under orthographic projection (ref: Tomasi and Kanade,

“Shape and Motion from Image Streams Under Orthography: a Factorization Method”, IJCV 1992) and a few other more

complex camera models (up to rank 9).

  • In case of orthographic projection, M can be expressed

as a product of two matrices – a rotation matrix of size 2F x 3, and a shape matrix of size 3 x P. Hence it has rank 3.

  • M is useful for many computer vision problems such as

structure from motion, motion segmentation and multi-frame point correspondences.