Statistical Image Models Eero Simoncelli Howard Hughes Medical - - PowerPoint PPT Presentation

statistical image models
SMART_READER_LITE
LIVE PREVIEW

Statistical Image Models Eero Simoncelli Howard Hughes Medical - - PowerPoint PPT Presentation

Statistical Image Models Eero Simoncelli Howard Hughes Medical Institute, Center for Neural Science, and Courant Institute of Mathematical Sciences New York University Photographic Images Diverse specialized structures:


slide-1
SLIDE 1

Statistical Image Models

Eero Simoncelli

Howard Hughes Medical Institute, Center for Neural Science, and Courant Institute of Mathematical Sciences New York University

slide-2
SLIDE 2

Photographic Images

Diverse specialized structures:

  • edges/lines/contours
  • shadows/highlights
  • smooth regions
  • textured regions
slide-3
SLIDE 3

Photographic Images

Diverse specialized structures:

  • edges/lines/contours
  • shadows/highlights
  • smooth regions
  • textured regions

Occupy a small region of the full space

slide-4
SLIDE 4

space of all images t y p i c a l i m a g e s

One could describe this set as a deterministic manifold....

slide-5
SLIDE 5
slide-6
SLIDE 6
slide-7
SLIDE 7
  • Step edges are rare (lighting, junctions, texture, noise)
slide-8
SLIDE 8
  • Step edges are rare (lighting, junctions, texture, noise)
  • One scale’s texture is another scale’s edge
slide-9
SLIDE 9
  • Step edges are rare (lighting, junctions, texture, noise)
  • One scale’s texture is another scale’s edge
  • Need seamless transitions from isolated features to

dense textures

slide-10
SLIDE 10

One could describe this set as a deterministic manifold....

space of all images t y p i c a l i m a g e s

slide-11
SLIDE 11

One could describe this set as a deterministic manifold.... But seems more natural to use probability

space of all images t y p i c a l i m a g e s

slide-12
SLIDE 12

One could describe this set as a deterministic manifold.... But seems more natural to use probability

space of all images t y p i c a l i m a g e s

P(x)

slide-13
SLIDE 13

“Applications”

  • Engineering: compression, denoising, restoration,

enhancement/modification, synthesis, manipulation

[Hubel ‘95]

slide-14
SLIDE 14

“Applications”

  • Engineering: compression, denoising, restoration,

enhancement/modification, synthesis, manipulation

  • Science: optimality principles for neurobiology (evolution,

development, learning, adaptation)

[Hubel ‘95]

slide-15
SLIDE 15

Density models

nonparametric parametric/ constrained

slide-16
SLIDE 16

Density models

nonparametric parametric/ constrained build a histogram from lots of

  • bservations...
slide-17
SLIDE 17

Density models

nonparametric parametric/ constrained build a histogram from lots of

  • bservations...

use “natural constraints” (geometry/photometry

  • f image formation,

computation, maxEnt)

slide-18
SLIDE 18

Density models

nonparametric parametric/ constrained build a histogram from lots of

  • bservations...

use “natural constraints” (geometry/photometry

  • f image formation,

computation, maxEnt) historical trend (technology driven)

slide-19
SLIDE 19

50 100 150 200 250

histogram Original image

Range: [0, 237] Dims: [256, 256]

slide-20
SLIDE 20

50 100 150 200 250

histogram Original image

Range: [0, 237] Dims: [256, 256]

50 100 150 200 250

histogram Equalized image

Range: [1.99, 238] Dims: [256, 256]

slide-21
SLIDE 21

50 100 150 200 250

histogram Original image

Range: [0, 237] Dims: [256, 256]

50 100 150 200 250

histogram Equalized image

Range: [1.99, 238] Dims: [256, 256]

slide-22
SLIDE 22

General methodology

Observe “interesting” Joint Statistics Transform to Optimal Representation

slide-23
SLIDE 23

General methodology

Observe “interesting” Joint Statistics Transform to Optimal Representation

slide-24
SLIDE 24

General methodology

Observe “interesting” Joint Statistics Transform to Optimal Representation

“Onion peeling”

slide-25
SLIDE 25

Evolution of image models

  • I. (1950’s): Fourier + Gaussian
  • II. (mid 80’s - late 90’s): Wavelets + kurtotic marginals
  • III. (mid 90’s - present): Wavelets + local context
  • local amplitude (contrast)
  • local orientation
  • IV. (last 5 years): Hierarchical models
slide-26
SLIDE 26

I(x,y) I(x+1,y) I(x,y) I(x+2,y) I(x,y) I(x+4,y)

10 1 Spatial separation (pixels)

Correlation

a. b.

Pixel correlation

slide-27
SLIDE 27

I(x,y) I(x+4,y)

10 20 30 40 1 Spatial separation (pixels)

Correlation

b.

I(x,y) I(x+1,y) I(x,y) I(x+2,y) I(x,y) I(x+4,y)

10 1 Spatial separation (pixels)

Correlation

a. b.

Pixel correlation

slide-28
SLIDE 28

Translation invariance

Assuming translation invariance,

slide-29
SLIDE 29

Translation invariance

Assuming translation invariance, => covariance matrix is Toeplitz (convolutional)

slide-30
SLIDE 30

Translation invariance

Assuming translation invariance, => covariance matrix is Toeplitz (convolutional) => eigenvectors are sinusoids

slide-31
SLIDE 31

Translation invariance

Assuming translation invariance, => covariance matrix is Toeplitz (convolutional) => eigenvectors are sinusoids => can diagonalize (decorrelate) with F.T.

slide-32
SLIDE 32

Translation invariance

Assuming translation invariance, => covariance matrix is Toeplitz (convolutional) => eigenvectors are sinusoids => can diagonalize (decorrelate) with F.T. Power spectrum captures full covariance structure

slide-33
SLIDE 33

Spectral power

Structural:

F(sω) = spF(ω) F(ω) ∝ 1 ωp

[Ritterman 52; DeRiugin 56; Field 87; Tolhurst 92; Ruderman/Bialek 94; ...]

Assume scale-invariance: then:

slide-34
SLIDE 34

Spectral power

Structural:

F(sω) = spF(ω) F(ω) ∝ 1 ωp

[Ritterman 52; DeRiugin 56; Field 87; Tolhurst 92; Ruderman/Bialek 94; ...]

Assume scale-invariance: then:

1 2 3 1 2 3 4 5 6

Log10 spatialfrequency (cycles/image) Log10 power

Empirical:

slide-35
SLIDE 35

Principal Components Analysis (PCA) + whitening

  • 20

20

  • 20

20 4

  • 4

4

  • 4

20

  • 20

20

  • 20

a. b. c.

slide-36
SLIDE 36

PCA basis for image blocks

slide-37
SLIDE 37

PCA basis for image blocks

PCA is not unique

slide-38
SLIDE 38

Maximum entropy (maxEnt)

E (f(x)) = c f(x) = x2 f(x) = |x| The density with maximal entropy satisfying pME(x) ∝ exp (−λf(x)) is of the form Examples: where λ depends on c

slide-39
SLIDE 39

Model I (Fourier/Gaussian)

Basis set: Image: Coefficient density:

slide-40
SLIDE 40

F -1

1/f2 P(c)

Gaussian model is weak

P(x)

F−1

ω−2

slide-41
SLIDE 41

F -1

1/f2 P(c)

Gaussian model is weak

a. b.

ω2 F F−1

P(x)

F−1

ω−2

slide-42
SLIDE 42

F -1

1/f2 P(c)

Gaussian model is weak

a. b.

ω2 F F−1

  • 20

20

  • 20

20 4

  • 4

4

  • 4

20

  • 20

20

  • 20

a. b. c.

P(x)

F−1

ω−2

slide-43
SLIDE 43

Bandpass Filter Responses

500 500 10

  • 4

10

  • 2

10

Filter Response Probability

Response histogram Gaussian density

[Burt&Adelson 82; Field 87; Mallat 89; Daugman 89, ...]

slide-44
SLIDE 44

“Independent” Components Analysis (ICA)

For Linearly Transformed Factorial (LTF) sources: guaranteed independence

(with some minor caveats)

  • 4

4

  • 4

4 20

  • 20

20

  • 20

20

  • 20

20

  • 20
  • 4

4

  • 4

4

a. b. c. d.

[Comon 94; Cardoso 96; Bell/Sejnowski 97; ...]

slide-45
SLIDE 45

ICA on image blocks

[Olshausen/Field ’96; Bell/Sejnowski ’97] [example obtained with FastICA, Hyvarinen]

slide-46
SLIDE 46

Marginal densities

P(x) ∝ exp −|x/s|p

[Mallat 89; Simoncelli&Adelson 96; Moulin&Liu 99; ...]

Well-fit by a generalized Gaussian:

Wavelet coefficient value log(Probability) p = 0.46 H/H = 0.0031 Wavelet coefficient value log(Probability) p = 0.58 H/H = 0.0011 Wavelet coefficient value log(Probability) p = 0.48 H/H = 0.0014

slide-47
SLIDE 47

Kurtosis vs. bandwidth

0.5 1 1.5 2 2.5 3 4 6 8 10 12 14 16 Filter Bandwidth (octaves) Sample Kurtosis

[after Field 87]

Note: Bandwidth matters much more than orientation

[see Bethge 06]

slide-48
SLIDE 48

Octave-bandwidth representations

Filter: Spatial Frequency Selectivity:

slide-49
SLIDE 49

Model II (LTF)

Basis set: Image: Coefficient density:

slide-50
SLIDE 50

LTF also a weak model...

Sample Gaussianized

Sample ICA-transformed and Gaussianized

slide-51
SLIDE 51

Trouble in paradise

slide-52
SLIDE 52

Trouble in paradise

  • Biology: Visual system uses a cascade
  • Where’s the retina? The LGN?
  • What happens after V1? Why don’t responses get

sparser? [Baddeley etal 97; Chechik etal 06]

slide-53
SLIDE 53

Trouble in paradise

  • Biology: Visual system uses a cascade
  • Where’s the retina? The LGN?
  • What happens after V1? Why don’t responses get

sparser? [Baddeley etal 97; Chechik etal 06]

  • Statistics: Images don’t obey ICA source model
  • Any bandpass filter gives sparse marginals [Baddeley 96]

=> Shallow optimum [Bethge 06; Lyu & Simoncelli 08]

  • The responses of ICA filters are highly dependent

[Wegmann & Zetzsche 90, Simoncelli 97]

slide-54
SLIDE 54

Conditional densities

  • 40
40 50 0.2 0.6 1
  • 40
40 0.2 0.6 1
  • 40

40

  • 40

40

[Simoncelli 97; Schwartz&Simoncelli 01]

slide-55
SLIDE 55

[Schwartz&Simoncelli 01]

slide-56
SLIDE 56
  • Large-magnitude subband coefficients are found at

neighboring positions, orientations, and scales.

slide-57
SLIDE 57

Method 1: Conditional Gaussian

[Simoncelli 97; Buccigrossi&Simoncelli 99; see also ARCH models in econometrics!]

Modeling heteroscedasticity

(i.e., variable variance)

P (xn|{xk}) ∼ N

  • 0;
  • k

wnk |xk|2 + σ2

slide-58
SLIDE 58

Joint densities

adjacent near far

  • ther scale
  • ther ori

100 100 150 100 50 50 100 150 100 100 150 100 50 50 100 150 100 100 150 100 50 50 100 150 500 500 150 100 50 50 100 150 100 100 150 100 50 50 100 150 100 100 150 100 50 50 100 150 100 100 150 100 50 50 100 150 100 100 150 100 50 50 100 150 500 500 150 100 50 50 100 150 100 100 150 100 50 50 100 150

[Simoncelli, ‘97; Wainwright&Simoncelli, ‘99]

  • Nearby: densities are approximately circular/elliptical
  • Distant: densities are approximately factorial
slide-59
SLIDE 59

ICA-transformed joint densities

d=2 d=16 d=32

kurtosis

  • rientation

data (ICA’d): factorialized: sphericalized:

4 6 8 10 12 !/4 !/2 3!/4 ! 4 6 8 10 12 !/4 !/2 3!/4 ! 4 6 8 10 12 !/4 !/2 3!/4 !

slide-60
SLIDE 60

ICA-transformed joint densities

d=2 d=16 d=32

kurtosis

  • rientation

data (ICA’d): factorialized: sphericalized:

4 6 8 10 12 !/4 !/2 3!/4 ! 4 6 8 10 12 !/4 !/2 3!/4 ! 4 6 8 10 12 !/4 !/2 3!/4 !

  • Local densities are elliptical (but non-Gaussian)
  • Distant densities are factorial

[Wegmann&Zetzsche ‘90; Simoncelli ’97; + many recent models]

slide-61
SLIDE 61

3 6 9 12 15 18 20 0.05 0.1 0.15 0.2

kurtosis

blk size = 3x3

blk spherical factorial

Spherical vs LTF

3x3 7x7 15x15

3 6 9 12 15 18 20 0.05 0.1 0.15 0.2

kurtosis

blk size = 7x7

blk spherical factorial 3 6 9 12 15 18 20 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4

kurtosis

blk size = 11x11

blk spherical factorial

data (ICA’d): factorialized: sphericalized:

  • Histograms, kurtosis of projections of image blocks onto random

unit-norm basis functions.

  • These imply data are closer to spherical than factorial

[Lyu & Simoncelli 08]

slide-62
SLIDE 62
  • Zetzsche & Krieger, 1999;
  • Huang & Mumford, 1999;
  • Wainwright & Simoncelli, 2000;
  • Hyvärinen and Hoyer, 2000;
  • Parra et al., 2001;
  • Srivastava et al., 2002;
  • Sendur & Selesnick, 2002;
  • Teh et al., 2003;
  • Gehler and Welling, 2006
  • Lyu & Simoncelli, 2008
  • etc.

non-Gaussian elliptical observations and models of natural images:

slide-63
SLIDE 63
  • is Gaussian,
  • and are independent
  • is elliptically symmetric, with covariance
  • marginals of are leptokurtotic

[Wainwright&Simoncelli 99]

Modeling heteroscedasticity

Method 2: Hidden scaling variable for each patch

Gaussian scale mixture (GSM)

[Andrews & Mallows 74]:

  • u
  • x = √z

u z

  • x
  • u

z > 0

  • x

∝ Cu

slide-64
SLIDE 64
  • Empirically, z is approximately lognormal

[Portilla etal, icip-01]

  • Alternatively, can use Jeffrey’s noninformative prior

[Figueiredo&Nowak, ‘01; Portilla etal, ‘03]

GSM - prior on z

pz(z) = exp (−(log z − µl)2/(2σ2

l ))

z(2πσ2

l )1/2

pz(z) ∝ 1/z

slide-65
SLIDE 65

GSM simulation

!!" " !" #"

"

#"

!

!!" " !" #"

"

#"

!

Image data GSM simulation

[Wainwright & Simoncelli, NIPS*99]

slide-66
SLIDE 66

Model III (GSM)

Basis set: Image: Coefficient density:

slide-67
SLIDE 67

Original coefficients Normalized by √z marginal

500 500 10 8 6 4 2 Log probability 5 5 10 9 8 7 6 5 4 Log probability

joint

100 50 50 100 100 50 50 100 2 4 6 8 2 4 6 8

subband

[Schwartz&Simoncelli 01] [Ruderman&Bialek 94]

slide-68
SLIDE 68

First Order Ideal Conditional Model 1 2 3 4 5 6 1 2 3 4 5 6 Empirical Conditional Entropy Model Encoding cost (bits/coeff) Gaussian Model Generalized Laplacian 3 4 5 2.5 3 3.5 4 4.5 5 5.5 Empirical First Order Entropy (bits/coeff) Model Encoding Cost (bits/coeff)

[Buccigrossi & Simoncelli 99]

slide-69
SLIDE 69

Bayesian denoising

  • Additive Gaussian noise:
  • Bayes’ least squares solution is conditional mean:

y = x + w

P(y|x) ∝ exp[−(y − x)2/2σ2

w]

ˆ x(y) = I E(x|y)

=

  • dxP(y|x)P(x)x/P(y)
slide-70
SLIDE 70
  • I. Classical

If signal is Gaussian, BLS estimator is linear:

denoised (ˆ x) noisy (y)

ˆ x(y) = σ2

x

σ2

x + σ2 n

· y

=> suppress fine scales, retain coarse scales

slide-71
SLIDE 71

Non-Gaussian coefficients

[Burt&Adelson ‘81; Field ‘87; Mallat ‘89; Daugman ‘89; etc]

!!"" " !"" #"

!$

#"

!%

#"

"

&'()*+,-*./01.* 2+0343'(')5

  • *./01.*,6'.)07+48

94:..'41,;*1.')5,,

slide-72
SLIDE 72
  • II. BLS for non-Gaussian prior
  • Assume marginal distribution [Mallat ‘89]:
  • Then Bayes estimator is generally nonlinear:

P(x) ∝ exp −|x/s|p

p = 2.0 p = 1.0 p = 0.5

[Simoncelli & Adelson, ‘96]

slide-73
SLIDE 73

MAP shrinkage

p=2.0 p=1.0 p=0.5

[Simoncelli 99]

slide-74
SLIDE 74

Denoising: Joint

I E(x| y) =

dz P(z|

y) I E(x| y, z) =

dz P(z|

y)

 zCu(zCu + Cw)−1

y

 

ctr

where P(z| y) = P( y|z) P(z) P y , P( y|z) = exp(− yT(zCu + Cw)−1 y/2)

  • (2π)N|zCu + Cw|

Numerical computation of solution is reasonably efficient if

  • ne jointly diagonalizes Cu and Cw ...

[Portilla, Strela, Wainwright, Simoncelli, ’03]

IPAM, 9/04 20

slide-75
SLIDE 75

Example estimators

Estimators for the scalar and single-neighbor cases

NOISY COEFF. ESTIMATED COEFF. !w

!!" " !" !#" " #" !!" " !" $%&'()*%+,,- $%&'()./0+$1 +'1&2/1+3)*%+,,-

[Portilla etal 03]

slide-76
SLIDE 76

Comparison to other methods

Results averaged over 3 images

!" #" $" %" &" !$'& !$ !#'& !# !!'& !! !"'& " "'& ()*+,-*.)/-0123 /456,(74-.)/-0123 )89!:1:;<=>? .=9@A?-9?=@BC8D !" #" $" %" &" !$'& !$ !#'& !# !!'& !! !"'& " "'& ()*+,-*.)/-0123 ,456748+91:;<= :>6965#8*>?6<

[Portilla etal 03]

slide-77
SLIDE 77

Original Noisy (22.1 dB) Matlab’s wiener2 (28 dB) BLS-GSM (30.5 dB)

slide-78
SLIDE 78

Original Noisy (8.1 dB) UndWvlt Thresh (19.0 dB) BLS-GSM (21.2 dB)

slide-79
SLIDE 79

Real sensor noise

400 ISO denoised

slide-80
SLIDE 80

GSM summary

  • GSM captures local variance
  • Underlying Gaussian leads to simple computation
  • Excellent denoising results
  • What’s missing?
  • Global model of z variables [Wainwright etal 99;

Romberg etal ‘99; Hyvarinen/Hoyer ‘02; Karklin/ Lewicki ‘02; Lyu/Simoncelli 08]

  • Explicit geometry: phase and orientation
slide-81
SLIDE 81

Global models for z

  • Non-overlapping neighborhoods, tree-structured z

[Wainwright etal 99; Romberg etal ’99]

  • Field of GSMs: z is an exponentiated GMRF, is

a GMRF, subband is the product

[Lyu&Simoncelli 08]

z

Coarse scale Fine scale

  • u
  • u
slide-82
SLIDE 82

Lena Boats

! "#"! $! !# %! "## !& !' !$ !" # "

σ

∆()*+,

! "#"! $! !# %! "## !& !' !$ !" # "

σ

∆()*+,

FoE kSVD GSM BM3D FoGSM

[Lyu&Simoncelli, PAMI 08]

State-of-the-art denoising

slide-83
SLIDE 83

2-band steerable pyramid: Image decomposition in terms of multi-scale gradient measurements

Measuring Orientation

[Simoncelli et.al., 1992; Simoncelli & Freeman 1995]

slide-84
SLIDE 84

Multi-scale gradient basis

slide-85
SLIDE 85

Multi-scale gradient basis

  • Multi-scale bases: efficient representation
slide-86
SLIDE 86

Multi-scale gradient basis

  • Multi-scale bases: efficient representation
  • Derivatives: good for analysis
  • Local Taylor expansion of image structures
  • Explicit geometry (orientation)
slide-87
SLIDE 87

Multi-scale gradient basis

  • Multi-scale bases: efficient representation
  • Derivatives: good for analysis
  • Local Taylor expansion of image structures
  • Explicit geometry (orientation)
  • Combination:
  • Explicit incorporation of geometry in basis
  • Bridge between PDE / harmonic analysis

approaches

slide-88
SLIDE 88
  • rientation
  • rientation

magnitude

[Hammond&Simoncelli 06; cf. Oppenheim and Lim 81]

slide-89
SLIDE 89

Importance of local orientation

Randomized orientation Randomized magnitude

[Hammond&Simoncelli 05]

slide-90
SLIDE 90

Reconstruction from orientation

  • Reconstruction by projections onto convex sets
  • Resilient to quantization

Quantized to 2 bits

[Hammond&Simoncelli 06]

Original

slide-91
SLIDE 91

Image patches related by rotation

two-band steerable pyramid coefficients

[Hammond&Simoncelli 06]

slide-92
SLIDE 92

raw patches rotated patches

  • -- Raw Patches

Rotated Patches

PCA of normalized gradient patches

[Hammond&Simoncelli 06]

slide-93
SLIDE 93

Orientation-Adaptive GSM model

patch rotation operator hidden magnitude/orientation variables Model a vectorized patch of wavelet coefficients as:

[Hammond&Simoncelli 06]

slide-94
SLIDE 94

Orientation-Adaptive GSM model

patch rotation operator hidden magnitude/orientation variables Model a vectorized patch of wavelet coefficients as: Conditioned on ; is zero mean gaussian with covariance

[Hammond&Simoncelli 06]

slide-95
SLIDE 95

[Hammond&Simoncelli 06]

Estimation of C(θ) from noisy data

noisy patch unknown, approximate by measured from noisy data. Assuming independent and noise rotationally invariant (assuming w.l.o.g. E[z] =1 )

slide-96
SLIDE 96

Bayesian MMSE Estimator

[Hammond&Simoncelli 06]

slide-97
SLIDE 97

Bayesian MMSE Estimator

condition on and integrate

  • ver hidden variables

[Hammond&Simoncelli 06]

slide-98
SLIDE 98

Bayesian MMSE Estimator

condition on and integrate

  • ver hidden variables

[Hammond&Simoncelli 06]

slide-99
SLIDE 99

Bayesian MMSE Estimator

condition on and integrate

  • ver hidden variables

Wiener estimate

[Hammond&Simoncelli 06]

slide-100
SLIDE 100

Bayesian MMSE Estimator

condition on and integrate

  • ver hidden variables

has covariance separable prior for hidden variables Wiener estimate

[Hammond&Simoncelli 06]

slide-101
SLIDE 101

Bayesian MMSE Estimator

condition on and integrate

  • ver hidden variables

has covariance separable prior for hidden variables Wiener estimate

[Hammond&Simoncelli 06]

slide-102
SLIDE 102
  • agsm

13.1 dB gsm2 12.4 dB σ = 40 noisy 2.81 dB

slide-103
SLIDE 103

Locally adaptive covariance

  • Karklin & Lewicki 08: Each patch is Gaussian,

with covariance constructed from a weighted outer- product of fixed vectors:

  • Guerrero-Colon, Simoncelli & Portilla 08: Each

patch is a mixture of GSMs (MGSMs):

p( x) =

  • k

Pk

  • p(zk) G(

x; zkCk) dzk p( y) =

  • n

exp(−|yn|) Bn =

  • k

wnk bk bk

T

log C( y) =

  • n

ynBn p( x) = G ( x; C( y))

slide-104
SLIDE 104

MGSMs generative model

  • x

Patch chosen from with probabilities Parameters:

  • Covariances
  • Scale densities
  • Component probabilities
  • Number of components

Ck {√z1 u1, √z2 u2, ...√zK uK} {P1, P2, ..., PK} pk(zk) Pk K Parameters can be fit to data of one or more images by maximizing likelihood (EM-like)

[Guerrero-Colon, Simoncelli, Portilla 08]

slide-105
SLIDE 105

MGSM “segmentation”

First six eigenvectors

  • f GSM

covariance matrices

image 1 2 4

[Guerrero-Colon, Simoncelli, Portilla 08]

slide-106
SLIDE 106

MGSM “segmentation”

Eigenvectors of GSM components represent invariant subspaces: “generalized complex cells”

slide-107
SLIDE 107

Potential of local homogeneous models?

  • marginal statistics [var,skew,kurtosis]
  • local raw correlations
  • local variance correlations
  • local phase correlations

Consider an implicit model: maxEnt subject to constraints on subband coefficients:

[Portilla & Simoncelli 00;

  • cf. Zhu, Wu & Mumford 97]
slide-108
SLIDE 108

Visual texture

slide-109
SLIDE 109

Visual texture

Homogeneous, with repeated structures

slide-110
SLIDE 110

Visual texture

Homogeneous, with repeated structures “You know it when you see it”

slide-111
SLIDE 111

All Images

Texture Images Equivalence class (visually indistinguishable)

slide-112
SLIDE 112

Iterative synthesis algorithm

Synthesis Analysis

Transform

Measure Statistics

Example Texture Random Seed Synthesized Texture

Transform

Measure Statistics

Adjust Inverse Transform

[Portilla&Simoncelli 00; cf. Heeger&Bergen ‘95]

slide-113
SLIDE 113

Examples: Artificial

slide-114
SLIDE 114

Photographic, quasi-periodic

slide-115
SLIDE 115

Photographic, aperiodic

slide-116
SLIDE 116

Photographic, structured

slide-117
SLIDE 117

Photographic, color

slide-118
SLIDE 118

Non-textures?

slide-119
SLIDE 119

Texture mixtures

slide-120
SLIDE 120

Texture mixtures

Convex combinations in parameter space

slide-121
SLIDE 121

Texture mixtures

Convex combinations in parameter space => Parameter space includes non-textures

slide-122
SLIDE 122

Summary

  • Fusion of empirical data with structural principles
  • Statistical models have led to state-of-the-art image

processing, and are relevant for biological vision

  • Local adaptation to {variance, orientation,

phase, ...} gives improvement, but makes learning harder

  • Cascaded representations emerge naturally
  • There’s still much room for improvement!
slide-123
SLIDE 123

Cast

  • Local GSM model: Martin Wainwright, Javier Portilla
  • GSM Denoising: Javier Portilla, Martin Wainwright,

Vasily Strela

  • Variance-adaptive compression: Robert Buccigrossi
  • Local orientation and OAGSM: David Hammond
  • Field of GSMs: Siwei Lyu
  • Mixture of GSMs: Jose-Antonio Guerrero-Colón,

Javier Portilla

  • Texture representation/synthesis: Javier Portilla