A regularized least-squares method for sparse low-rank approximation - - PowerPoint PPT Presentation

a regularized least squares method for sparse low rank
SMART_READER_LITE
LIVE PREVIEW

A regularized least-squares method for sparse low-rank approximation - - PowerPoint PPT Presentation

Workshop Numerical methods for high-dimensional problems April 18, 2014 A regularized least-squares method for sparse low-rank approximation of multivariate functions Mathilde Chevreuil joint work with Prashant Rai, Loic Giraldi, Anthony


slide-1
SLIDE 1

Workshop “Numerical methods for high-dimensional problems” April 18, 2014

A regularized least-squares method for sparse low-rank approximation of multivariate functions

Mathilde Chevreuil

joint work with Prashant Rai, Loic Giraldi, Anthony Nouy

GeM – Institut de Recherche en G´ enie Civil et M´ ecanique LUNAM Universit´ e UMR CNRS 6183 / Universit´ e de Nantes / Centrale Nantes

slide-2
SLIDE 2

Motivations

Chorus ANR project

  • Aero-thermal regulation in an aircraft cabin

39 random parameters Data basis of 2000 evaluations of the model

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 2

slide-3
SLIDE 3

Motivations

  • In telecommunication: electromagnetic field and the Specific Absorption Rate (SAR)

induced in the body Over 4 random parameters FDTD method: 2 days/run. Few evaluations of the model available

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 3

slide-4
SLIDE 4

Motivations

  • In telecommunication: electromagnetic field and the Specific Absorption Rate (SAR)

induced in the body Over 4 random parameters FDTD method: 2 days/run. Few evaluations of the model available Aim Construct a surrogate model of the true model from a small collection of evaluations

  • f the true model that allows fast evaluations of output quantities of interest, observables
  • r objective function.

Propagation: estimation of quantiles, sensitivity analysis ... Optimization or identification Probabilistic inverse problem

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 3

slide-5
SLIDE 5

Uncertainty quantification using functional approaches

  • Stochastic/parametric models

Uncertainties represented by “simple” random variables ξ = (ξ1, · · · , ξd) : Θ → Ξ defined

  • n a probability space (Θ, B, P).

Model ξ u(ξ)

Ideal approach Compute an accurate and explicit representation of u(ξ): u(ξ) ≈

  • α∈IP

uαφα(ξ), ξ ∈ Ξ where the φα(ξ) constitute a suitable basis of multiparametric functions

Polynomial chaos [Ghanem and Spanos 1991, Xiu and Karniadakis 2002, Soize and Ghanem 2004] Piecewise polynomials, wavelets [Deb 2001, Le Maˆ ıtre 2004, Wan 2005]

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 4

slide-6
SLIDE 6

Motivations

  • Aproximation spaces

SP = span{φα(ξ) = φ(1)

α1(ξ1) . . . φ(d) αd (ξd); α ∈ IP}

with a pre-defined index set IP, e.g.

  • α ∈ Nd; |α|∞ ≤ r
  • α ∈ Nd; |α|1 ≤ r
  • α ∈ Nd; |α|q ≤ r
  • , 0 < q < 1

Issue

  • Approximation of a high dimensional function u(ξ), ξ ∈ Ξ ⊂ Rd

#(IP) ≈ 10, 1010, 10100, 101000, ...

  • Use of deterministic solvers in a black box manner

Numerous evaluations of possibly fine deterministic models

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 5

slide-7
SLIDE 7

Motivations

Objective Compute an approximation of u ∈ SP u(ξ) ≈

  • α∈IP

uαφα(ξ) using few samples {u(y q)}Q

q=1

where {y q}Q

q=1 is a collection of sample points and the u(y q) are solutions of the

deterministic problem Exploit structures of u(ξ) u(ξ) can be sparse on particular basis functions u(ξ) can have suitable low rank representations Can we exploit sparsity within low rank structure of u?

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 6

slide-8
SLIDE 8

Outline

1

Motivations and framework

2

Sparse low rank approximation

3

Tensor formats and algorithms Canonical decomposition Tensor Train format

4

Conclusion

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 7

slide-9
SLIDE 9

Outline

1

Motivations and framework

2

Sparse low rank approximation

3

Tensor formats and algorithms Canonical decomposition Tensor Train format

4

Conclusion

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 7

slide-10
SLIDE 10

Low rank approximation

Approximation of function u using tensor approximation methods

  • Exploit the tensor structure of function space

SP = S1

P1 ⊗ . . . ⊗ Sd Pd ;

Sk

Pk = span

  • φ(k)

i

Pk

i=1

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 8

slide-11
SLIDE 11

Low rank approximation

Approximation of function u using tensor approximation methods

  • Exploit the tensor structure of function space

SP = S1

P1 ⊗ . . . ⊗ Sd Pd ;

Sk

Pk = span

  • φ(k)

i

Pk

i=1

  • Low rank tensor subsets M

M = {v = FM(p1, p2, . . . , pn)} with dim(M) = O(d)

[Nouy 2010, Khoromskij and Schwab 2010, Ballani 2010, Beylkin et al 2011, Matthies and Zander 2012, Doostan et al 2012, ...]

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 8

slide-12
SLIDE 12

Low rank approximation

Approximation of function u using tensor approximation methods

  • Exploit the tensor structure of function space

SP = S1

P1 ⊗ . . . ⊗ Sd Pd ;

Sk

Pk = span

  • φ(k)

i

Pk

i=1

  • Low rank tensor subsets M

M = {v = FM(p1, p2, . . . , pn)} with dim(M) = O(d)

[Nouy 2010, Khoromskij and Schwab 2010, Ballani 2010, Beylkin et al 2011, Matthies and Zander 2012, Doostan et al 2012, ...]

  • Sparse low rank tensor subsets Mm-sparse, ideally

Mm-sparse = {v = FM(p1, p2, . . . , pn); pi0 ≤ mi; 1 ≤ i ≤ n} with dim(Mm-sparse) ≪ dim(M) .

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 8

slide-13
SLIDE 13

Low rank approximation

Least-squares in low rank subsets

  • Approximation of v(ξ) ∈ M defined by

min

v∈M u − v2 Q

with u − v2

Q = Q

  • k=1

|u(y k) − v(y k)|2

[Beylkin et al 2011, Doostan et al 2012]

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 9

slide-14
SLIDE 14

Low rank approximation

Least-squares in low rank subsets

  • Approximation of v(ξ) ∈ M defined by

min

v∈M u − v2 Q

with u − v2

Q = Q

  • k=1

|u(y k) − v(y k)|2

[Beylkin et al 2011, Doostan et al 2012]

  • Approximation of v(ξ) ∈ Mm−sparse defined by

min

v∈M u − v2 Q s.t. pi0 ≤ mi

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 9

slide-15
SLIDE 15

Low rank approximation

Least-squares in low rank subsets

  • Approximation of v(ξ) ∈ M defined by

min

v∈M u − v2 Q

with u − v2

Q = Q

  • k=1

|u(y k) − v(y k)|2

[Beylkin et al 2011, Doostan et al 2012]

  • Approximation of v(ξ) ∈ Mm−sparse defined by

min

v∈M u − v2 Q s.t. pi0 ≤ mi → min v∈M u − v2 Q + n

  • i=1

λipi1 (Lasso)

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 9

slide-16
SLIDE 16

Low rank approximation

Least-squares in low rank subsets

  • Approximation of v(ξ) ∈ M defined by

min

v∈M u − v2 Q

with u − v2

Q = Q

  • k=1

|u(y k) − v(y k)|2

[Beylkin et al 2011, Doostan et al 2012]

  • Approximation of v(ξ) ∈ Mm−sparse defined by

min

v∈M u − v2 Q s.t. pi0 ≤ mi → min v∈M u − v2 Q + n

  • i=1

λipi1 (Lasso) Alternating least-squares with sparse regularization For 1 ≤ i ≤ n and for fixed pj with j = i min

pi u − FM(p1, . . . , pi, . . . , pn)2 Q+λipi1

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 9

slide-17
SLIDE 17

Outline

1

Motivations and framework

2

Sparse low rank approximation

3

Tensor formats and algorithms Canonical decomposition Tensor Train format

4

Conclusion

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 10

slide-18
SLIDE 18

Approximation in canonical tensor subset

Rank-one canonical tensor subset R1 =

  • w = w (1) ⊗ . . . ⊗ w (d) ; w (k) ∈ Sk

Pk s.t. w (k)(ξk) = φ(k)(ξk)Tw(k)

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 11

slide-19
SLIDE 19

Approximation in canonical tensor subset

Rank-one canonical tensor subset R1 =

  • w = φ, w(1) ⊗ . . . ⊗ w(d); w(k) ∈ RPk
  • where φ =
  • φ(1) ⊗ . . . ⊗ φ(d)

(ξ) and with dim(R1 ) = d

k=1 Pk

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 11

slide-20
SLIDE 20

Approximation in canonical tensor subset

Rank-one canonical tensor subset Rγ

1 =

  • w = φ, w(1) ⊗ . . . ⊗ w(d); w(k) ∈ RPk , w(k)1 ≤ γk
  • where φ =
  • φ(1) ⊗ . . . ⊗ φ(d)

(ξ) and with dim(Rγ

1 ) = d k=1 Pk

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 11

slide-21
SLIDE 21

Approximation in canonical tensor subset

Rank-one canonical tensor subset Rγ

1 =

  • w = φ, w(1) ⊗ . . . ⊗ w(d); w(k) ∈ RPk , w(k)1 ≤ γk
  • where φ =
  • φ(1) ⊗ . . . ⊗ φ(d)

(ξ) and with dim(Rγ

1 ) = d k=1 Pk

Rank-m tensor subsets Rγ1,...,γm

m

={v =

m

  • i=1

wi ; wi ∈ Rγi

1 }

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 11

slide-22
SLIDE 22

Approximation in canonical tensor subset

Rank-one canonical tensor subset Rγ

1 =

  • w = φ, w(1) ⊗ . . . ⊗ w(d); w(k) ∈ RPk , w(k)1 ≤ γk
  • where φ =
  • φ(1) ⊗ . . . ⊗ φ(d)

(ξ) and with dim(Rγ

1 ) = d k=1 Pk

Rank-m tensor subsets Rγ1,...,γm

m

={v =

m

  • i=1

wi ; wi ∈ Rγi

1 }

=

  • v = φ,

m

  • i=1

w(1)

i

⊗ . . . ⊗ w(d)

i

; w(k)

i

1 ≤ γi

k

  • Motivations and framework

Sparse LR approx. Tensor formats & alg. Conclusion 11

slide-23
SLIDE 23

Algorithm for adaptive sparse tensor approximation

  • Algorithms

Progressive construction based on corrections in Rγ

1

greedy construction of a basis {wi}m

i=1 selected in a tensor subset Rγi 1

Compute um = m

i=1 αiwi using regularized least-squares

[MC, R. Lebrun, A. Nouy, P. Ray, A least-squares method for sparse low rank approximation of multivariate functions, arXiv:1305.0030, 2013]

Direct approximation in Rγ1,...,γm

m

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 12

slide-24
SLIDE 24

Algorithm for adaptive sparse tensor approximation

Algorithm for progressive construction Let u0 = 0. For m ≥ 1, Compute a sparse rank-one correction wm ∈ Rγ

1 by solving

min

w∈Rγ

1

u − um−1 − w2

Q

Computed using alternating minimization on the parameters of Rγ

1 .

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 13

slide-25
SLIDE 25

Algorithm for adaptive sparse tensor approximation

Algorithm for progressive construction Let u0 = 0. For m ≥ 1, Compute a sparse rank-one correction wm ∈ Rγ

1 by solving

min

w(1)∈RP1 ,...,w(d)∈RPd u − um−1 − φ, w(1) ⊗ . . . ⊗ w(d)2 Q + d

  • k=1

λkw(k)1 Computed using Alternating regularized Least Squares

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 13

slide-26
SLIDE 26

Algorithm for adaptive sparse tensor approximation

Algorithm for progressive construction Let u0 = 0. For m ≥ 1, Compute a sparse rank-one correction wm ∈ Rγ

1 by solving

min

w(1)∈RP1 ,...,w(d)∈RPd u − um−1 − φ, w(1) ⊗ . . . ⊗ w(d)2 Q + d

  • k=1

λkw(k)1 Computed using Alternating regularized Least Squares: for 1 ≤ j ≤ d

min

w(j)∈Rnj z − Φ(j)w(j)2 2 + λjw(j)1

where (Φ(j))qi = φ(j)

i (yq j )

  • k=j

w(k)(yq

k )

Lasso problem computed with LARS algorithm Optimal solution selected using the fast LOO CV error estimate

[Blatman, Sudret 2011]

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 13

slide-27
SLIDE 27

Algorithm for adaptive sparse tensor approximation

Algorithm for progressive construction Let u0 = 0. For m ≥ 1, Compute a sparse rank-one correction wm ∈ Rγ

1 by solving

min

w(1)∈RP1 ,...,w(d)∈RPd u − um−1 − φ, w(1) ⊗ . . . ⊗ w(d)2 Q + d

  • k=1

λkw(k)1 Computed using Alternating regularized Least Squares: for 1 ≤ j ≤ d

min

w(j)∈Rnj z − Φ(j)w(j)2 2 + λjw(j)1

where (Φ(j))qi = φ(j)

i (yq j )

  • k=j

w(k)(yq

k )

Lasso problem computed with LARS algorithm Optimal solution selected using the fast LOO CV error estimate

[Blatman, Sudret 2011]

Set Um = span{wi}m

i=1 (reduced approximation space)

Compute um = m

i=1 αiwi ∈ Rγ1,...,γm m

using sparse regularization min

α=(α1,...,αm)∈Rm u − m

  • i=1

αiwi2

Q + λ′α1

Best rank selected using cross validation method

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 13

slide-28
SLIDE 28

Illustration: checker-board function

  • Rank-2 function: u(ξ1, ξ2) = 2

i=1 w (1) i

(ξ1)w (2)

i

(ξ2)

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 0.2 0.4 0.6 0.8 1 1.2

ξ1 w1 (1)(ξ1)

(a) w (1)

1 (ξ1)

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 0.2 0.4 0.6 0.8 1 1.2

ξ1 w2 (1)(ξ1)

(b) w (1)

2 (ξ1)

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 0.2 0.4 0.6 0.8 1 1.2

ξ2 w1 (2)(ξ2)

(c) w (2)

1 (ξ2)

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 0.2 0.4 0.6 0.8 1 1.2

ξ2 w2 (2)(ξ2)

(d) w (2)

2 (ξ2)

with dimension: d = 2 ξi ∈ U(0, 1). Ξ = (0, 1)2.

1/6 2/6 3/6 4/6 5/6 1 1/6 2/6 3/6 4/6 5/6 1

  • Approximation of u in S1

P1 ⊗ S2 P2

Piecewise polynomials of degree p defined on a uniform partition of Ξk of s intervals: Sk

Pk = Pp,s

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 14

slide-29
SLIDE 29

Illustration: checker-board function

  • Performance of the method for sparse low rank approximation

Q = 200 samples Optimal rank mop selected using 3-fold cross validation Relative error ε estimated with Monte Carlo integration Comparison of different regularizations within Alternated Least Squares

OLS ℓ2 ℓ1 Approximation space ε mop ε mop ε mop Rm(P2,3 ⊗ P2,3), P = 92 0.527 2 0.508 2 0.507 2 Rm(P2,6 ⊗ P2,6), P = 182 0.664 2 0.061 8 2.41 10−13 2 Rm(P2,12 ⊗ P2,12), P = 362

  • 0.566

4 1.50 10−12 3 Rm(P10,6 ⊗ P10,6), P = 662

  • 0.855

10 7.88 10−13 2

With few samples: ℓ1-regularization detects sparsity and gives accurate results Rank 2 is retrieved

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 15

slide-30
SLIDE 30

Illustration: Friedman function

  • Friedman function

f (ξ) = 10sin(πξ1ξ2) + 20(ξ3 − 0.5)2 + 10ξ4 + 5ξ5 Dimensions: d = 5 ξi, i = 1, . . . , 5 are uniform random variables over [0, 1].

  • Approximation in SP = 5

k=1 Sk Pk

Polynomials of degree p: Sk

Pk = Pp

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 16

slide-31
SLIDE 31

Illustration: Friedman function

  • Friedman function

f (ξ) = 10sin(πξ1ξ2) + 20(ξ3 − 0.5)2 + 10ξ4 + 5ξ5 Dimensions: d = 5 ξi, i = 1, . . . , 5 are uniform random variables over [0, 1].

  • Approximation in SP = 5

k=1 Sk Pk

Polynomials of degree p: Sk

Pk = Pp

What is the sufficient number of samples Q∗ just needed given an a priori underlying approximation space ? Q∗ = f (p, d, m)

[G. Migliorati, F. Nobile, E. von Schwerin, and R. Tempone, 2011]: sufficient condition for a stable

approximation of a multivariate function using OLS: Q∗ ∼ (#(IP))2

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 16

slide-32
SLIDE 32

Illustration: Friedman function

  • Number of samples needed (no regularization): rank-1

Number of samples: Q = cd(p + 1) c ∈ R∗

+

1 2 3 4 5 6 7 8 9 10 0.08 0.1 0.12 0.14 0.16 0.18 0.2 0.22

Polynomial degree Error

c=1 c=3 c=5 c=10 c=20

Number of samples: Q = cd(p + 1)2 c ∈ R∗

+

1 2 3 4 5 6 7 8 9 10 0.05 0.1 0.15 0.2 0.25 0.3

Polynomial degree Error

c=0.5 c=1 c=1.5 c=2 c=3

Q∗ = d(p + 1)2

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 17

slide-33
SLIDE 33

Illustration: Friedman function

  • Number of samples needed (no regularization): rank-4

Number of samples: Q = cmd(p + 1) c ∈ R∗

+

1 2 3 4 5 6 7 8 9 10 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45

Polynomial degree Error

c=1 c=3 c=5 c=10 c=20

Number of samples: Q = cmd(p + 1)2 c ∈ R∗

+

1 2 3 4 5 6 7 8 9 10 0.05 0.1 0.15 0.2 0.25 0.3 0.35

Polynomial degree Error

c=0.5 c=1 c=1.5 c=2 c=3

Q∗ = md(p + 1)2

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 17

slide-34
SLIDE 34

Illustration: vibration analysis

  • Discrete problem

u ∈ CN, (−ω2M + iωC + K)u = f

where K = E ˜ K and C = iωηE ˜ K with E =

  • 0.975 + 0.025ξ1
  • n horizontal plate,

0.975 + 0.025ξ2

  • n vertical plate,

η =

  • 0.0075 + 0.0025ξ3
  • n horizontal plate,

0.0075 + 0.0025ξ4

  • n vertical plate,

where the ξk ∼ U(−1, 1), k = 1, · · · , 4. Ξ = (−1, 1)8.

  • Approximation of a Variable of Interest I(u) in SP = 5

k=1 Sk Pk

I(u)(ξ) = log uc , Polynomials of degree p: Sk

Pk = Pp

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 18

slide-35
SLIDE 35

Illustration: vibration analysis

Evolution of error w.r.t. p Q = 80 Q = 200 Sparsity ratio w.r.t. p Q = 80 Q = 200

dashed lines: OLS, solid lines: with ℓ1 regularization

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 19

slide-36
SLIDE 36

Illustration: vibration analysis

Evolution of error w.r.t. p

dashed lines: OLS, solid lines: with ℓ1 regularization

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 20

slide-37
SLIDE 37

Outline

1

Motivations and framework

2

Sparse low rank approximation

3

Tensor formats and algorithms Canonical decomposition Tensor Train format

4

Conclusion

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 21

slide-38
SLIDE 38

Tensor Train format

ξ1 . . . ξd ξ1 ξ2 . . . ξd ξ2 . . . ξd−2 ξd−1ξd ξd−1 ξd

[Oseldets 2009,...]

v =

r1

  • i1=1

v (1)

1,i1 ⊗ v (2,...,d) i1

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 22

slide-39
SLIDE 39

Tensor Train format

ξ1 . . . ξd ξ1 ξ2 . . . ξd ξ2 . . . ξd−2 ξd−1ξd ξd−1 ξd

[Oseldets 2009,...]

v =

r1

  • i1=1

v (1)

1,i1 ⊗ v (2,...,d) i1

v =

r1

  • i1=1

v (1)

1,i1 ⊗ r2

  • i2=1

v (2)

i1i2 ⊗ . . . ⊗ rd−1

  • id−1=1

v (d−1)

id−2id−1 ⊗ v (d) id−1,1

Tensor Train subsets TT(1,r1,...,rd−1,1) = TTr The set of tensors TTr(S) is defined by TTr =

  • v =
  • i∈I
  • k

v (k)

ik−1ik ; v (k) ik−1ik ∈ Sk Pk

  • .

where I = {i = (i0, i1, . . . , id−1, id); ik ∈ {1, . . . , rk}} with r0 = rd = 1 Parameterization TTr =

  • v = Fr(v1, . . . , vd); vk ∈ (RPk )rk−1×rk

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 22

slide-40
SLIDE 40

Algorithms

Alternating least-squares in TTr

  • For a given rank vector r

min

v∈TTr u − v2 Q + d

  • k=1

λkvec(vk)1

  • Question of selection of rank vector r

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 23

slide-41
SLIDE 41

Algorithm for adaptive sparse tensor approximation: DMRG

Re-parameterization

  • Consider the tensor w (k) ∈ (Sk

Pk )rk−1 ⊗ (Sk+1 Pk+1)rk+1: w (k) = r∗

k

ik =1 v (k,∗) ik

⊗ v (k+1,∗)

ik

⇒ v = F k

r (v, w (k)) = r1

  • i1=1

. . .

rk−1

  • ik−1=1

rk+1

  • ik+1=1

. . .

rd−1

  • id−1=1

v (1)

1i1 ⊗ . . . ⊗ w (k) ik−1ik+1 ⊗ . . . ⊗ v (d) id−11

  • Compute sparse low-rank w (k) with adaptive rank

Modified alternating least-squares algorithm For k ∈ {1, · · · , d − 1} Compute w (k) ∈ (Sk

Pk )rk−1 ⊗ (Sk+1 Pk+1)rk+1 by solving

min

w(k)∈R(rk−1Pk )×(rk+1Pk+1)

  • u − Fk(v, w(k))
  • 2

Q

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 24

slide-42
SLIDE 42

Algorithm for adaptive sparse tensor approximation: DMRG

Re-parameterization

  • Consider the tensor w (k) ∈ (Sk

Pk )rk−1 ⊗ (Sk+1 Pk+1)rk+1: w (k) = r∗

k

ik =1 v (k,∗) ik

⊗ v (k+1,∗)

ik

⇒ v = F k

r (v, w (k)) = r1

  • i1=1

. . .

rk−1

  • ik−1=1

rk+1

  • ik+1=1

. . .

rd−1

  • id−1=1

v (1)

1i1 ⊗ . . . ⊗ w (k) ik−1ik+1 ⊗ . . . ⊗ v (d) id−11

  • Compute sparse low-rank w (k) with adaptive rank

Modified alternating least-squares algorithm For k ∈ {1, · · · , d − 1} Compute sparse w (k) ∈ (Sk

Pk )rk−1 ⊗ (Sk+1 Pk+1)rk+1 by solving

min

w(k)∈R(rk−1Pk )×(rk+1Pk+1)

  • u − Fk(v, w(k))
  • 2

Q + λkvec(w(k))1

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 24

slide-43
SLIDE 43

Algorithm for adaptive sparse tensor approximation: DMRG

Re-parameterization

  • Consider the tensor w (k) ∈ (Sk

Pk )rk−1 ⊗ (Sk+1 Pk+1)rk+1: w (k) = r∗

k

ik =1 v (k,∗) ik

⊗ v (k+1,∗)

ik

⇒ v = F k

r (v, w (k)) = r1

  • i1=1

. . .

rk−1

  • ik−1=1

rk+1

  • ik+1=1

. . .

rd−1

  • id−1=1

v (1)

1i1 ⊗ . . . ⊗ w (k) ik−1ik+1 ⊗ . . . ⊗ v (d) id−11

  • Compute sparse low-rank w (k) with adaptive rank

Modified alternating least-squares algorithm For k ∈ {1, · · · , d − 1} Compute sparse w (k) ∈ (Sk

Pk )rk−1 ⊗ (Sk+1 Pk+1)rk+1 by solving

min

w(k)∈R(rk−1Pk )×(rk+1Pk+1)

  • u − Fk(v, w(k))
  • 2

Q + λkvec(w(k))1

Compute best low-rank approximation in (Sk

Pk )rk−1 ⊗ (Sk+1 Pk+1)rk+1 using SVD →

adaptive rank r ∗

k

v =

r1

  • i1=1

. . .

r∗

k

  • ik =1

. . .

rd−1

  • id−1=1

v (1)

1i1 ⊗ . . . ⊗ v (k,∗) ik−1ik ⊗ v (k+1,∗) ik ik+1

⊗ . . . ⊗ v (d)

id−11

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 24

slide-44
SLIDE 44

Illustration: sine of a sum

  • Sine function:

u(ξ) = sin(ξ1 + ξ2 + . . . + ξ6) with ξi ∈ U(−1, 1). Ξ = (−1, 1)6.

  • Evolution of error with respect to sample size Q
  • Approx. in SP = 6

k=1 Sk Pk ;Sk Pk = P2

10

2

10

3

10

−3

10

−2

10

−1

10

Sample Size Error

Tensor Train (MALS) Least Square+l1 Greedy R1 Rm

  • Approx. in SP = 6

k=1 Sk Pk ;Sk Pk = P4

100 200 300 400 500 600 700 10

−4

10

−3

10

−2

10

−1

10 Sample Size Error

Tensor Train (MALS) Least Square+l1 Greedy R1 Rm Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 25

slide-45
SLIDE 45

Illustration: borehole function

  • The Borehole function models water flow through a borehole:

f (ξ) = 2πTu(Hu − Hl) ln(r/rw)

  • 1 +

2LTu ln(r/rw )r2

w Kw + Tu

Tl

  • Dimension: d = 8

rw radius of borehole (m) N(µ = 0.10, σ = 0.0161812) r radius of influence (m) LN(µ = 7.71, σ = 1.0056) Tu transmissivity of upper aquifer (m2/yr) U[63070, 115600] Hu potentiometric head of upper aquifer (m) U[990, 1110] Tl transmissivity of lower aquifer (m2/yr) U[63.1, 116] Hl potentiometric head of lower aquifer (m) U[700, 820] L length of borehole (m) U[1120, 1680] Kw hydraulic conductivity of borehole (m/yr) U[9855, 12045]

  • Approximation in SP = 8

k=1 Sk Pk

Polynomials of degree p: Sk

Pk = Pp

p = 2: P = 6561 p = 3: P = 65536

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 26

slide-46
SLIDE 46

Illustration: borehole function

  • Behavior of the algorithm

Evolution of error w.r.t. Q

10

2

10

3

10

−4

10

−3

10

−2

10

−1

Sample Size (Q) Error

p=2 p=3

TT ranks (Q=200, p=3)

10 20 30 40 50 5 10 15 20 DMRG iteration Rank

r1 r2 r3 r4 r5 r6 r7

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 27

slide-47
SLIDE 47

Illustration: Canister

  • Stochastic PDE

             ∂u ∂t − ∇(κ∇u) + c(D · ∇u) = σu

  • n

Ω1 ∪ Ω2 u = ξ1 on Γ1 × Ωt u = 0 on Γ2 × Ωt u,n = 0 on (∂Ω\(Γ1 ∪ Γ2)) × Ωt with

ξ1 u(t = 0) U[0.8, 1.2] on Ω ξ2 σ U[8, 12] on Ω2 ξ3 σ U[0.8, 1] on Ω1 ξ4 c U[1, 5] ξ5 κ U[0.02, 0.03]

  • Approximation of a Variable of Interest I(u) in SP = 5

k=1 Sk Pk

I(u) =

  • T
  • Ω3

u(x, t)dxdt Polynomials of degree p: Sk

Pk = Pp

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 28

slide-48
SLIDE 48

Illustration: Canister

  • Evolution of error with respect to sample size Q
  • Approx. in SP = 5

k=1 Sk Pk ;Sk Pk = P2

10

1

10

2

10

3

10

−2

10

−1

Sample Size Error Tensor Train (MALS) Least Square+l1 Greedy R1

  • Approx. in SP = 5

k=1 Sk Pk ;Sk Pk = P3

10

1

10

2

10

3

10

−2

10

−1

Sample Size Error Tensor Train (MALS) Least Square+l1 Greedy R1 Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 29

slide-49
SLIDE 49

Illustration: Canister

  • Order of separation of variables

ξ1 . . . ξ5 ξ1 ξ2 . . . ξ5 ξ2 ξ3ξ4ξ5 ξ3 ξ4ξ5 ξ4 ξ5

10

1

10

2

10

3

0.02 0.04 0.06 0.08 0.1 Sample Size Error Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 30

slide-50
SLIDE 50

Illustration: Canister

  • Order of separation of variables

ξ1 . . . ξ5 ξ1 ξ2 . . . ξ5 ξ2 ξ3ξ4ξ5 ξ3 ξ4ξ5 ξ4 ξ5

10

1

10

2

10

3

0.02 0.04 0.06 0.08 0.1 Sample Size Error Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 30

slide-51
SLIDE 51

Illustration: Canister

  • Order of separation of variables

ξ1 . . . ξ5 ξ1 ξ2 . . . ξ5 ξ2 ξ3ξ4ξ5 ξ3 ξ4ξ5 ξ4 ξ5

10

1

10

2

10

3

0.02 0.04 0.06 0.08 0.1 Sample Size Error

ξ1 . . . ξ5 ξ1 ξ2 . . . ξ5 ξ5 ξ2ξ3ξ4 ξ2 ξ3ξ4 ξ3 ξ4

10

1

10

2

10

3

0.02 0.04 0.06 0.08 0.1 Sample Size Error Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 30

slide-52
SLIDE 52

Outline

1

Motivations and framework

2

Sparse low rank approximation

3

Tensor formats and algorithms Canonical decomposition Tensor Train format

4

Conclusion

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 31

slide-53
SLIDE 53

Conclusion

Least-squares method for sparse low rank approximation of high dimensional functions A non intrusive method Detects and exploits low-rank and sparsity Adaptive rank Outlook More analyses on the suffisant number of samples to find an approximation in a tensor subset Include adaptivity with respect to polynomial degree for underlying approximation spaces Strategies for optimal separation of variables (choice of tree)

This research was supported by EADS Innovation Works and by the French National Research Agency (ANR) (CHORUS project)

Motivations and framework Sparse LR approx. Tensor formats & alg. Conclusion 32