Function interpolation and compressed sensing Ben Adcock Department - - PowerPoint PPT Presentation

function interpolation and compressed sensing
SMART_READER_LITE
LIVE PREVIEW

Function interpolation and compressed sensing Ben Adcock Department - - PowerPoint PPT Presentation

Weighted 1 minimization Introduction Infinite-dimensional framework References Function interpolation and compressed sensing Ben Adcock Department of Mathematics Simon Fraser University 1 / 28 Weighted 1 minimization Introduction


slide-1
SLIDE 1

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Function interpolation and compressed sensing

Ben Adcock

Department of Mathematics Simon Fraser University

1 / 28

slide-2
SLIDE 2

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Outline

Introduction Infinite-dimensional framework New recovery guarantees for weighted ℓ1 minimization References

2 / 28

slide-3
SLIDE 3

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Outline

Introduction Infinite-dimensional framework New recovery guarantees for weighted ℓ1 minimization References

3 / 28

slide-4
SLIDE 4

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

High-dimensional approximation

Let

  • D ⊆ Rd be a domain, d ≫ 1
  • f : D → C be a (smooth) function
  • {ti}m

i=1 be a set of sample points

Goal: Approximate f from {f (ti)}m

i=1.

Applications: Uncertainty Quantification (UQ), scattered data approximation, numerical PDEs,.... Main issue: curse of dimensionality (exponential blow-up with d).

4 / 28

slide-5
SLIDE 5

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Quantifying uncertainty via polynomial chaos expansions

Uncertainty Quantification: Understand how output f (the quantity of interest) of a physical system behave as functions of the inputs t (the parameters). Polynomial Chaos Expansions: (Xiu & Karniadakis, 2002). Expand f (t) using multivariate orthogonal polynomials f (t) ≈

M

  • i=1

xiφi(t). Non-intrusive methods: Recover {xi}M

i=1 from samples {f (ti)}m i=1.

5 / 28

slide-6
SLIDE 6

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Stochastic Collocation

Two widely-used approaches: Structured meshes and interpolation (M = m): E.g. Sparse grids.

  • Efficient interpolation schemes in moderate dimensions
  • But may be too structured for very high dimensions, or miss certain

features (e.g. anisotropic behaviour). Unstructured meshes and regression (m > M): Random sampling combined with least-squares fitting.

  • For the right distributions, can obtain stable approximation with

d-independent scaling of m and M.

  • But still inefficient, especially in high dimensions.

Question

Can compressed sensing techniques be useful here?

6 / 28

slide-7
SLIDE 7

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Stochastic Collocation

Two widely-used approaches: Structured meshes and interpolation (M = m): E.g. Sparse grids.

  • Efficient interpolation schemes in moderate dimensions
  • But may be too structured for very high dimensions, or miss certain

features (e.g. anisotropic behaviour). Unstructured meshes and regression (m > M): Random sampling combined with least-squares fitting.

  • For the right distributions, can obtain stable approximation with

d-independent scaling of m and M.

  • But still inefficient, especially in high dimensions.

Question

Can compressed sensing techniques be useful here?

6 / 28

slide-8
SLIDE 8

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Compressed sensing in UQ

Theoretical work:

  • Rauhut & Ward (2011), 1D Legendre polynomials
  • Yan, Guo & Xiu (2012), dD Legendre polynomials
  • Tang & Iaccarino (2014), randomized quadratures
  • Hampton & Doostan (2014), coherence-optimized sampling
  • Xu & Zhou (2014), deterministic sampling
  • Rauhut & Ward (2014), weighted ℓ1 minimization
  • Chkifa, Dexter, Tran & Webster (2015), weighted ℓ1 minimization

Applications to UQ:

  • Doostan & Owhadi (2011), Mathelin & Gallivan (2012), Lei, Yang,

Zheng, Lin & Baker (2014), Rauhut & Schwab (2015), Yang, Lei, Baker & Lin (2015), Jakeman, Eldred & Sargsyan (2015), Karagiannis, Konomi & Lin (2015), Guo, Narayan, Xiu & Zhou (2015) and others.

7 / 28

slide-9
SLIDE 9

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Are polynomial coefficient sparse?

Low dimensions: polynomial coefficients exhibit decay, not sparsity:

Polynomial coefficients Wavelet coefficients

20 40 60 80 100 120 0.02 0.04 0.06 0.08 20 40 60 80 100 120 0.2 0.4 0.6 0.8 1.0 1.2

Decay Sparsity

Nonlinear approximation error ≈ Linear approximation error We may as well use interpolation/least squares.

8 / 28

slide-10
SLIDE 10

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Are polynomial coefficient sparse?

Higher dimensions: polynomial coefficients are increasingly sparse (Doostan et al., Schwab et al., Webster et al.,....).

5000 10000 15000 0.2 0.4 0.6 0.8 1.0

Polynomial coefficients, d = 10

Nonlinear approximation error ≪ Linear approximation error

9 / 28

slide-11
SLIDE 11

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Sparsity and lower sets

In high dimensions, polynomial coefficients concentrate on lower sets:

Definition (Lower set)

A set ∆ ⊆ Nd is lower if, for any i = (i1, . . . , id) and j = (j1, . . . , jd) with jk ≤ ik, ∀k, we have i ∈ ∆ ⇒ j ∈ ∆. Note: The number of lower sets of size s is O

  • s log(s)d−1

.

10 / 28

slide-12
SLIDE 12

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Outline

Introduction Infinite-dimensional framework New recovery guarantees for weighted ℓ1 minimization References

11 / 28

slide-13
SLIDE 13

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Setup

Let

  • ν be a measure on D with
  • D dν = 1,
  • T = {ti}m

i=1 ⊆ D, m ∈ N be drawn independently from ν,

  • {φj}j∈N be an orthonormal system in L2

ν(D) ∩ L∞(D) (typically,

tensor algebraic polynomials). Suppose that f =

  • j∈N

xjφj, xj = f , φjL2

ν,

where {xj}j∈N are the coefficients of f in the system {φj}j∈N.

12 / 28

slide-14
SLIDE 14

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Current approaches – discretize first

Most existing approaches follow a ‘discretize first’ approach. Choose M ≥ m and solve the finite-dimensional problem min

z∈CM z1,w subject to Az − y2 ≤ δ,

(⋆) for some δ ≥ 0, where z1,w = M

i=1 wi|zi|, {wi}M i=1 are weights and

A = {φj(ti)}m,M

i=1,j=1 ,

y = {f (ti)}m

i=1.

If ˆ x ∈ CM is a minimizer, set f ≈ ˜ f = M

i=1 ˆ

xiφi.

13 / 28

slide-15
SLIDE 15

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Current approaches – discretize first

Most existing approaches follow a ‘discretize first’ approach. Choose M ≥ m and solve the finite-dimensional problem min

z∈CM z1,w subject to Az − y2 ≤ δ,

(⋆) for some δ ≥ 0, where z1,w = M

i=1 wi|zi|, {wi}M i=1 are weights and

A = {φj(ti)}m,M

i=1,j=1 ,

y = {f (ti)}m

i=1.

If ˆ x ∈ CM is a minimizer, set f ≈ ˜ f = M

i=1 ˆ

xiφi.

13 / 28

slide-16
SLIDE 16

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

The choice of δ

The parameter δ is chosen so that the best approximation M

i=1 xiφi to f

from span{φi}M

i=1 is feasible for (⋆).

In other words, we require δ ≥

  • f −

M

  • i=1

xiφi

  • L∞

=

  • i>M

xiφi

  • L∞

. Equivalently, we treat the expansion tail as noise in the data.

Problems

  • This tail error is unknown in general.
  • A good estimation is necessary in order to get good accuracy.
  • Empirical estimation via cross validation is tricky and wasteful.
  • Solutions of (⋆) do not interpolate the data.

14 / 28

slide-17
SLIDE 17

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

The choice of δ

The parameter δ is chosen so that the best approximation M

i=1 xiφi to f

from span{φi}M

i=1 is feasible for (⋆).

In other words, we require δ ≥

  • f −

M

  • i=1

xiφi

  • L∞

=

  • i>M

xiφi

  • L∞

. Equivalently, we treat the expansion tail as noise in the data.

Problems

  • This tail error is unknown in general.
  • A good estimation is necessary in order to get good accuracy.
  • Empirical estimation via cross validation is tricky and wasteful.
  • Solutions of (⋆) do not interpolate the data.

14 / 28

slide-18
SLIDE 18

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

New approach

We propose the infinite-dimensional ℓ1 minimization inf

z∈ℓ1

w(N) z1,w subject to Uz = y,

where y = {f (ti)}m

i=1, {wi}i∈N are weights and

U = {φj(ti)}m,∞

i=1,j=1 ∈ Cm×∞,

is an infinitely fat matrix.

Advantages

  • Solutions are interpolatory.
  • No need to know the expansion tail.
  • Agnostic to the ordering of the functions {φi}i∈N.

Note: a similar setup can also handle noisy data.

15 / 28

slide-19
SLIDE 19

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

New approach

We propose the infinite-dimensional ℓ1 minimization inf

z∈ℓ1

w(N) z1,w subject to Uz = y,

where y = {f (ti)}m

i=1, {wi}i∈N are weights and

U = {φj(ti)}m,∞

i=1,j=1 ∈ Cm×∞,

is an infinitely fat matrix.

Advantages

  • Solutions are interpolatory.
  • No need to know the expansion tail.
  • Agnostic to the ordering of the functions {φi}i∈N.

Note: a similar setup can also handle noisy data.

15 / 28

slide-20
SLIDE 20

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Discretization

We cannot numerically solve the problem inf

z∈ℓ1

w(N) z1,w subject to Uz = y.

(1) Discretization strategy: Introduce a parameter K ∈ N and solve the finite-dimensional problem min

z∈PK (ℓ1

w(N)) z1,w subject to UPKz = y,

(2) where PK is defined by PKz = {z1, . . . , zK, 0, 0, . . .}.

  • Note: UPK is equivalent to a fat m × K matrix.

Main Idea

Choose K suitably large, and independent of f , so that solutions of (2) are close to solutions of (1).

16 / 28

slide-21
SLIDE 21

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

How to choose K

Let TK(x) be the additional error introduced by this discretization.

Theorem (BA)

Let x ∈ ℓ1

˜ w(N), where ˜

wi ≥ √ iw 2

i , ∀i. Suppose that K is sufficiently

large so that σr = σr(PKU∗) > 0, where r = rank(U). Then TK(x) ≤ x − PKx1,w + 1/σrx − PKx1, ˜

w.

The truncation condition σr ≈ 1 depends only on T and {φi}i∈N and is independent of the function f to recover. Example: Let D = (−1, 1)d with tensor Jacobi polynomials or the Fourier basis and equispaced data. Then K = O

  • m1+ǫ

, ǫ > 0, suffices.

Rule-of-thumb

Letting K ≈ 4m works fine in most settings.

17 / 28

slide-22
SLIDE 22

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Outline

Introduction Infinite-dimensional framework New recovery guarantees for weighted ℓ1 minimization References

18 / 28

slide-23
SLIDE 23

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Background

Unweighted ℓ1 minimization:

  • Recovery guarantees: Rauhut & Ward (2011), Yan, Guo & Xiu (2012).
  • Applications to UQ: Doostan & Owhadi (2011), Mathelin & Gallivan

(2012), Hampton & Doostan (2014), Tang & Iaccarino (2014), Guo, Narayan, Xiu & Zhou (2015).

Weighted ℓ1 minimization: Observed empirically to give superior results.

100 200 300 400 10

−4

10

−2

10 α = 0.0 α = 0.5 α = 1.0 α = 1.5 α = 2.0 100 200 300 400 10

−2

10

−1

α = 0.0 α = 0.5 α = 1.0 α = 1.5 α = 2.0 100 200 300 400 10

−4

10

−2

10 α = 0.0 α = 0.5 α = 1.0 α = 1.5 α = 2.0

f (t) = e2t1 cos(3t2) f (t) = sin(et1t2t3/2) f (t) = e− t1+t2+t3+t4

6

Plot of error versus m with algebraic weights: wi = (i1 · · · id)α, α ≥ 0.

19 / 28

slide-24
SLIDE 24

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Standard weighting strategies

Non-adapted weights: Slowly-growing (e.g. algebraic) weights used to alleviate aliasing/overfitting.

  • Rauhut & Ward (2014), Rauhut & Schwab (2015), BA (2015).

Adapted weights: Weights chosen according to support estimates.

  • A priori estimates: Peng, Hampton & Doostan (2014).
  • Iterative re-weighting: Yang & Karniadakis (2014).
  • See also: Bah & Ward (2015).

Goal

Find recovery guarantees that explain the effectiveness of both strategies.

20 / 28

slide-25
SLIDE 25

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Existing recovery guarantees

Rauhut & Ward (2014):

  • Weights: wi ≥ φiL∞
  • Weighted sparsity: s = |∆|w =

i∈∆ w 2 i , where ∆ = supp(x).

  • Recovery guarantee: m s × log factors.

Problem: This is not sharp. Let wi = iα and suppose that f is such that xj = 0, 1 ≤ j ≤ k, xj ≈ 0, j > k. This is reasonable for oscillatory functions, for example. Then m k2α+1 × log factors. This estimate deteriorates with increasing α.

  • Note: The same argument generalizes to any dimension when the

coefficients lie on a hyperbolic cross, BA (2015).

21 / 28

slide-26
SLIDE 26

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Existing recovery guarantees

Rauhut & Ward (2014):

  • Weights: wi ≥ φiL∞
  • Weighted sparsity: s = |∆|w =

i∈∆ w 2 i , where ∆ = supp(x).

  • Recovery guarantee: m s × log factors.

Problem: This is not sharp. Let wi = iα and suppose that f is such that xj = 0, 1 ≤ j ≤ k, xj ≈ 0, j > k. This is reasonable for oscillatory functions, for example. Then m k2α+1 × log factors. This estimate deteriorates with increasing α.

  • Note: The same argument generalizes to any dimension when the

coefficients lie on a hyperbolic cross, BA (2015).

21 / 28

slide-27
SLIDE 27

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Example

Take f (t) = cos(45 √ 2t + 1/3) and consider Chebyshev polynomials with random samples drawn from the Chebyshev measure.

20 40 60 80 100 −15 −10 −5

i log10 |xi|

50 100 150 200 250 10

−4

10

−2

10 α = 0.00 α = 0.50 α = 1.00 α = 1.50 α = 2.00

Coefficients xj Error versus m

22 / 28

slide-28
SLIDE 28

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

A new recovery guarantee

Theorem (BA)

Let w = {wi}i∈N be weights, x ∈ ℓ1

w(N) and ∆ ⊆ {1, . . . , K} be such

that mini∈{1,...,K}\∆{wi} ≥ 1. Let t1, . . . , tm be drawn independently from ν. Then x − ˆ x2 x − P∆x1,w + TK(x), with probability at least 1 − ǫ, provided m

  • |∆|u +

max

i∈{1,...,K}\∆{u2 i /w 2 i } max{|∆|w, 1}

  • · L,

(⋆) where ui = max{φiL∞, 1} and L = log(ǫ−1) · log(2N

  • max{|∆|w, 1}).

Remarks:

  • The weights ui are intrinsic to the problem.
  • This is a nonuniform guarantee – (⋆) relies heavily on this approach.
  • As is typical, the error bound is weaker (ℓ2/ℓ1

w).

23 / 28

slide-29
SLIDE 29

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Consequence I: Sharpness for linear models

Consider the main estimate: m

  • |∆|u +

max

i∈{1,...,K}\∆{u2 i /w 2 i } max{|∆|w, 1}

  • · L.

Sharpness for linear models: Let ∆ = {1, . . . , k}. Suppose that ui = O (iγ) and wi = O (iα) for α > γ ≥ 0. Then m k2γ+1 × log factors.

  • This is independent of the weights and optimal, up to log factors.
  • Extends to any dimension for coefficients lying on a hyperbolic cross.

24 / 28

slide-30
SLIDE 30

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Consequence II: Optimal non-adapted weights

For non-adapted weights, the estimate m

  • |∆|u +

max

i∈{1,...,K}\∆{u2 i /w 2 i } max{|∆|w, 1}

  • · L.

is minimized by setting wi = ui. Example 1: Legendre polynomials, uniform measure.

  • wi = 1: m 3d · s · L, where s = |∆|.
  • wi = ui: m s2 · L provided ∆ is a lower set.
  • Note that s2 is sharp and avoids the curse of dimensionality.

Example 2: Chebyshev polynomials, Chebyshev measure.

  • wi = 1: m 2d · s · L.
  • wi = ui: m slog(3)/ log(2) · L provided ∆ is a lower set.

25 / 28

slide-31
SLIDE 31

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Consequence II: Optimal non-adapted weights

For non-adapted weights, the estimate m

  • |∆|u +

max

i∈{1,...,K}\∆{u2 i /w 2 i } max{|∆|w, 1}

  • · L.

is minimized by setting wi = ui. Example 1: Legendre polynomials, uniform measure.

  • wi = 1: m 3d · s · L, where s = |∆|.
  • wi = ui: m s2 · L provided ∆ is a lower set.
  • Note that s2 is sharp and avoids the curse of dimensionality.

Example 2: Chebyshev polynomials, Chebyshev measure.

  • wi = 1: m 2d · s · L.
  • wi = ui: m slog(3)/ log(2) · L provided ∆ is a lower set.

25 / 28

slide-32
SLIDE 32

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Consequence II: Optimal non-adapted weights

For non-adapted weights, the estimate m

  • |∆|u +

max

i∈{1,...,K}\∆{u2 i /w 2 i } max{|∆|w, 1}

  • · L.

is minimized by setting wi = ui. Example 1: Legendre polynomials, uniform measure.

  • wi = 1: m 3d · s · L, where s = |∆|.
  • wi = ui: m s2 · L provided ∆ is a lower set.
  • Note that s2 is sharp and avoids the curse of dimensionality.

Example 2: Chebyshev polynomials, Chebyshev measure.

  • wi = 1: m 2d · s · L.
  • wi = ui: m slog(3)/ log(2) · L provided ∆ is a lower set.

25 / 28

slide-33
SLIDE 33

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Consequence III: The benefits of adapted weights

Corollary (BA)

Assume ui = 1 for simplicity. Let x be s-sparse with support ∆. Let Γ ⊆ {1, . . . , K} and suppose that wi = σ < 1, i ∈ Γ, and wi = 1, i / ∈ Γ. Then we require m (2(1 − ρα) + (1 + γ)ρ) · s · L, measurements, where α = |∆ ∩ Γ|/|Γ|, |Γ|/|∆| = ρ.

  • Recall that m 2 · s · L in the unweighted case.
  • Hence we see an improvement whenever α > 1

2(1 + γ).

  • That is, we estimate ≈ 50% of the support correctly, for small γ.

Related work:

  • Friedlander, Mansour, Saab & Yilmaz (2012), Yu & Baek (2013) (random

Gaussian measurements).

26 / 28

slide-34
SLIDE 34

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Outline

Introduction Infinite-dimensional framework New recovery guarantees for weighted ℓ1 minimization References

27 / 28

slide-35
SLIDE 35

Introduction Infinite-dimensional framework Weighted ℓ1 minimization References

Thanks!

For more info, see:

  • B. Adcock, Infinite-dimensional weighted ℓ1 minimization and function

approximation from pointwise data, arXiv:1503.02352 (2015).

  • B. Adcock, Infinite-dimensional compressed sensing and function

interpolation, arXiv:1509.06073 (2015).

28 / 28