Learning from random moments Rmi Gribonval - Inria Rennes - Bretagne - - PowerPoint PPT Presentation

learning from random moments
SMART_READER_LITE
LIVE PREVIEW

Learning from random moments Rmi Gribonval - Inria Rennes - Bretagne - - PowerPoint PPT Presentation

Learning from random moments Rmi Gribonval - Inria Rennes - Bretagne Atlantique remi.gribonval@inria.fr Joint work with: G. Blanchard (U. Potsdam) N. Keriven, Y Traonmilin (Inria Rennes) 1 Main Contributors & Collaborators Anthony


slide-1
SLIDE 1

Rémi Gribonval - Inria Rennes - Bretagne Atlantique

remi.gribonval@inria.fr Joint work with: G. Blanchard (U. Potsdam)

  • N. Keriven, Y Traonmilin (Inria Rennes)

Learning from random moments

1

slide-2
SLIDE 2
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Main Contributors & Collaborators

Anthony Bourrier Nicolas Keriven Yann Traonmilin Nicolas Tremblay Gilles Puy Mike Davies Patrick Perez Gilles Blanchard

2

slide-3
SLIDE 3
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Foreword

Signal processing & machine learning

inverse problems & generalized method of moments embeddings with random projections & random features /kernels image super-resolution, source localization & k-means

Continuous vs discrete ?

wavelets (1990s): from continuous to discrete compressive sensing (2000s): in the discrete world current trends : back to continuous !

  • ff-the-grid compressive sensing, FRI, high-resolution methods

compressive statistical learning from random moments

3

slide-4
SLIDE 4

Learning from random moments: the concept Compressive Statistical Learning (guarantees) Recent developments & perspectives

  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

4

slide-5
SLIDE 5
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Large-scale learning

x1 x2

xn

X

5

slide-6
SLIDE 6
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

High feature dimension d Large collection size n = “volume”

Large-scale learning

x1 x2

xn

X

5

slide-7
SLIDE 7
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

High feature dimension d Large collection size n = “volume”

Large-scale learning

x1 x2

xn

X

Challenge: compress before learning ?

X

5

slide-8
SLIDE 8
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

dimension reduction subsampling sketching

Compressive learning: three routes

Y = MX

x1 x2

xn

X

random projections - Johnson Lindenstrauss lemma see e.g. [Calderbank & al 2009, Reboredo & al 2013]

6

slide-9
SLIDE 9
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

dimension reduction subsampling sketching

x1 x2

xn

Compressive learning: three routes

x1 x2

xn

X

Nyström method & coresets see e.g. [Williams&Seeger 2000, Agarwal & al 2003, Felman 2010]

7

slide-10
SLIDE 10
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

dimension reduction subsampling random moments

Compressive learning: three routes

x1 x2

xn

X

Inspiration: compressive sensing [Foucart & Rauhut 2013] sketching/hashing [Thaper & al 2002, Cormode & al 2005] Connections with: generalized method of moments [Hall 2005] kernel mean embeddings[Smola & al 2007, Sriperimbudur & al 2010]

z ∈ Rm

EΦ1(X)

EΦm(X)

8

slide-11
SLIDE 11
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Example: Compressive K-means

X

Training set

n = 70000; d = 784; k = 10

9

slide-12
SLIDE 12
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

  • 1
  • 0.5

0.5 1

  • Dim. 5
  • 1
  • 0.5

0.5 1

  • Dim. 6
  • 1
  • 0.5

0.5 1

  • Dim. 5
  • 1
  • 0.5

0.5 1

  • Dim. 6

Example: Compressive K-means

X

Training set

Spectral features

n = 70000; d = k = 10

9

slide-13
SLIDE 13
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

  • 1
  • 0.5

0.5 1

  • Dim. 5
  • 1
  • 0.5

0.5 1

  • Dim. 6
  • 1
  • 0.5

0.5 1

  • Dim. 5
  • 1
  • 0.5

0.5 1

  • Dim. 6

Example: Compressive K-means

z ∈ Rm m & kd = 100 X

memory size independent of n

Training set Sketch vector

Spectral features

n = 70000; d = k = 10 Sketch(X)

= 1 n

n

X

i=1

Φ(xi)

streaming / distributed computation 9

slide-14
SLIDE 14
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Example: Compressive K-means

z ∈ Rm m & kd = 100

memory size independent of n

Sketch vector

n = 70000; d = k = 10

= 1 n

n

X

i=1

Φ(xi)

Privacy-aware streaming / distributed computation 9

slide-15
SLIDE 15
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

  • 1
  • 0.5

0.5 1

  • Dim. 5
  • 1
  • 0.5

0.5 1

  • Dim. 6
  • 1
  • 0.5

0.5 1

  • Dim. 5
  • 1
  • 0.5

0.5 1

  • Dim. 6

Example: Compressive K-means

Learn centroids from sketch = moment fitting

z ∈ Rm m & kd = 100

memory size independent of n

Sketch vector

n = 70000; d = k = 10

= 1 n

n

X

i=1

Φ(xi)

Privacy-aware streaming / distributed computation 9

slide-16
SLIDE 16
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

  • 1
  • 0.5

0.5 1

  • Dim. 5
  • 1
  • 0.5

0.5 1

  • Dim. 6
  • 1
  • 0.5

0.5 1

  • Dim. 5
  • 1
  • 0.5

0.5 1

  • Dim. 6
  • 1
  • 0.5

0.5 1

  • Dim. 5
  • 1
  • 0.5

0.5 1

  • Dim. 6
  • 1
  • 0.5

0.5 1

  • Dim. 5
  • 1
  • 0.5

0.5 1

  • Dim. 6

Example: Compressive K-means

Learn centroids from sketch = moment fitting

z ∈ Rm m & kd = 100

memory size independent of n

Sketch vector

n = 70000; d = k = 10

= 1 n

n

X

i=1

Φ(xi)

Privacy-aware streaming / distributed computation 9

slide-17
SLIDE 17
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

  • 1
  • 0.5

0.5 1

  • Dim. 5
  • 1
  • 0.5

0.5 1

  • Dim. 6
  • 1
  • 0.5

0.5 1

  • Dim. 5
  • 1
  • 0.5

0.5 1

  • Dim. 6
  • 1
  • 0.5

0.5 1

  • Dim. 5
  • 1
  • 0.5

0.5 1

  • Dim. 6
  • 1
  • 0.5

0.5 1

  • Dim. 5
  • 1
  • 0.5

0.5 1

  • Dim. 6

Example: Compressive K-means

Learn centroids from sketch = moment fitting

z ∈ Rm m & kd = 100

memory size independent of n

Sketch vector

n = 70000; d = k = 10

= 1 n

n

X

i=1

Φ(xi)

Privacy-aware streaming / distributed computation 9

Φ(x) := {eıωT

j x}m

j=1

Using: random Fourier features

Vector-valued function

slide-18
SLIDE 18
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Sketching & Neural networks

Sketching for k-means

empirical characteristic function

z` = 1 n

n

X

i=1

ejw>

` xi

10

slide-19
SLIDE 19
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Sketching & Neural networks

Sketching for k-means

empirical characteristic function

X z` = 1 n

n

X

i=1

ejw>

` xi

10

slide-20
SLIDE 20
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

wT

` X

Sketching & Neural networks

Sketching for k-means

empirical characteristic function

X wT

`

z` = 1 n

n

X

i=1

ejw>

` xi

10

slide-21
SLIDE 21
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Sketching & Neural networks

Sketching for k-means

empirical characteristic function

X

W

WX m z` = 1 n

n

X

i=1

ejw>

` xi

10

slide-22
SLIDE 22
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Sketching & Neural networks

Sketching for k-means

empirical characteristic function

X

W

WX h(WX) h(·) = ej(·) m z` = 1 n

n

X

i=1

ejw>

` xi

10

slide-23
SLIDE 23
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Sketching & Neural networks

Sketching for k-means

empirical characteristic function

X

W

WX h(WX) h(·) = ej(·) z

average

m m z` = 1 n

n

X

i=1

ejw>

` xi

10

slide-24
SLIDE 24
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

~ One-layer random neural net

DNN ~ hierarchical sketching ?

Sketching & Neural networks

Sketching for k-means

empirical characteristic function

X

W

WX h(WX) h(·) = ej(·) z

average

see also [Bruna & al 2013, Giryes & al 2015]

m m z` = 1 n

n

X

i=1

ejw>

` xi

10

slide-25
SLIDE 25
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Sketching & Privacy

Sketching

empirical characteristic function

X

W

WX h(WX) h(·) = ej(·) z

average

Privacy-reserving

sketch and forget

see also [Bruna & al 2013, Giryes & al 2015]

~ One-layer random neural net

DNN ~ hierarchical sketching ?

z` = 1 n

n

X

i=1

ejw>

` xi

11

slide-26
SLIDE 26
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Sketching & Online Learning

Sketching

empirical characteristic function

X

W

WX h(WX) h(·) = ej(·) z

average

streaming

Streaming algorithms

One pass; online update

z` = 1 n

n

X

i=1

ejw>

` xi

12

slide-27
SLIDE 27
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Sketching & Distributed Computing

Sketching

empirical characteristic function

X

W

WX h(WX) h(·) = ej(·) z

average

… … … …

DIS TRI BU TED

Distributed computing

Decentralized (HADOOP) / parallel (GPU)

z` = 1 n

n

X

i=1

ejw>

` xi

13

slide-28
SLIDE 28

Learning from random moments: the concept Compressive Statistical Learning (guarantees) Recent developments & perspectives

  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

14

slide-29
SLIDE 29
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Statistical learning 101

Statistical risk Target Empirical version PAC / excess risk control /generalization error

can be achieved if uniform convergence, i.e. whp

R(p, ✓) = Ex∼p`(x, ✓)

θ? ∈ arg min

R(p?, θ)

R(p?, ˆ θn) ≤ R(p?, θ?) + ηn

ˆ θn ∈ arg min

θ

R(ˆ pn, θ)

xi ∼ p?, i.i.d.

sup

|R(ˆ pn, θ) − R(p?, θ)| ≤ ηn/2

15

ˆ pn := 1

n n

X

i=1

δxi

slide-30
SLIDE 30
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Compressive Statistical Learning

Step 1: given learning task

Design sketching function

Step 2: compress & learn

Summarize training collection with sketch Learn from sketch with some algorithm

z = 1

n n

X

i=1

Φ(xi)

R(p?, ˆ θ(z)) ≤ R(p?, θ?) + ηn

Φ(x) ∈ Rm

16

z 7! ˆ θ(z)

➡ controlled excess risk (PAC)?

slide-31
SLIDE 31
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Task: select a k-dimensional subspace

Loss function:

Step 1: hoice of sketching function

naive: full covariance matrix, refined:= using compressive matrix sensing

Learn from sketch: low-rank matrix recovery

✓ statistical guarantees

Worked example 1: Compressive PCA

x ∈ Rd

m = O(kd) m = O(d2)

17

`(x, ✓) = kx Pθxk2

θ

Φ(x) = xxT Φ(x) = {(ωT

j x)2}m j=1

Σ = EXXT

z ≈ vec(Σ) z ≈ A(vec(Σ))

ˆ Σk(z) := arg min

rankΣ≤k

Σ⌫0

kz A(vec(Σ))k2

ˆ θ(z) := span( ˆ Σk(z))

R(ˆ θ) ≤ (1 + C)R(θ∗) + O(1/√n)

slide-32
SLIDE 32
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Task: select a k-dimensional subspace

Loss function:

Step 1: hoice of sketching function

naive: full covariance matrix, refined:= using compressive matrix sensing

Learn from sketch: low-rank matrix recovery

✓ statistical guarantees

Worked example 1: Compressive PCA

x ∈ Rd

m = O(kd) m = O(d2)

17

`(x, ✓) = kx Pθxk2

θ

Φ(x) = xxT Φ(x) = {(ωT

j x)2}m j=1

Σ = EXXT

z ≈ vec(Σ) z ≈ A(vec(Σ))

ˆ Σk(z) := arg min

rankΣ≤k

Σ⌫0

kz A(vec(Σ))k2

ˆ θ(z) := span( ˆ Σk(z))

R(ˆ θ) ≤ (1 + C)R(θ∗) + O(1/√n)

slide-33
SLIDE 33
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Task: select a k-dimensional subspace

Loss function:

Step 1: hoice of sketching function

naive: full covariance matrix, refined:= using compressive matrix sensing

Learn from sketch: low-rank matrix recovery

✓ statistical guarantees

Worked example 1: Compressive PCA

x ∈ Rd

m = O(kd) m = O(d2)

17

`(x, ✓) = kx Pθxk2

θ

Φ(x) = xxT Φ(x) = {(ωT

j x)2}m j=1

Σ = EXXT

z ≈ vec(Σ) z ≈ A(vec(Σ))

ˆ Σk(z) := arg min

rankΣ≤k

Σ⌫0

kz A(vec(Σ))k2

ˆ θ(z) := span( ˆ Σk(z))

R(ˆ θ) ≤ (1 + C)R(θ∗) + O(1/√n)

slide-34
SLIDE 34
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Task: select a k-dimensional subspace

Loss function:

Step 1: hoice of sketching function

naive: full covariance matrix, refined:= using compressive matrix sensing

Learn from sketch: low-rank matrix recovery

✓ statistical guarantees

✓ # of parameters to learn

Worked example 1: Compressive PCA

x ∈ Rd

m = O(kd) m = O(d2)

17

`(x, ✓) = kx Pθxk2

θ

Φ(x) = xxT Φ(x) = {(ωT

j x)2}m j=1

Σ = EXXT

z ≈ vec(Σ) z ≈ A(vec(Σ))

ˆ Σk(z) := arg min

rankΣ≤k

Σ⌫0

kz A(vec(Σ))k2

ˆ θ(z) := span( ˆ Σk(z))

R(ˆ θ) ≤ (1 + C)R(θ∗) + O(1/√n)

slide-35
SLIDE 35
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Worked example 2: Compressive K-Means

Task: find k centroids

Loss function:

Standard approach: “K-means algorithm”

aka Lloyd-Max algorithm [Steinhaus 1956, Lloyd 1957 (publ. 1982)]

several passes on the training set

Naive sketching =histograms

bins of size within domain of radius R

exponential sketch size compressive sensing suggests can we avoid discretization (bypass curse of dimensionality) ?

✏ N = O((R/✏)d) m = N m = O(k log N) = O(kd log(R/✏))

18

`(x, ✓) = min

i

kx ✓ik2 θ = {θ1, . . . , θk}, θi ∈ Rd

slide-36
SLIDE 36
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Worked example 2: Compressive K-Means

Task: find k centroids

Loss function:

Standard approach: “K-means algorithm”

aka Lloyd-Max algorithm [Steinhaus 1956, Lloyd 1957 (publ. 1982)]

several passes on the training set

Naive sketching =histograms

bins of size within domain of radius R

exponential sketch size compressive sensing suggests can we avoid discretization (bypass curse of dimensionality) ?

✏ N = O((R/✏)d) m = N m = O(k log N) = O(kd log(R/✏))

18

`(x, ✓) = min

i

kx ✓ik2 θ = {θ1, . . . , θk}, θi ∈ Rd

slide-37
SLIDE 37
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Worked example 2: Compressive K-Means

Task: find k centroids

Loss function:

Standard approach: “K-means algorithm”

aka Lloyd-Max algorithm [Steinhaus 1956, Lloyd 1957 (publ. 1982)]

several passes on the training set

Naive sketching =histograms

bins of size within domain of radius R

exponential sketch size compressive sensing suggests can we avoid discretization (bypass curse of dimensionality) ?

✏ N = O((R/✏)d) m = N m = O(k log N) = O(kd log(R/✏))

18

`(x, ✓) = min

i

kx ✓ik2 θ = {θ1, . . . , θk}, θi ∈ Rd

slide-38
SLIDE 38
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Worked example 2: Compressive K-Means

Task: find k centroids

Loss function:

Standard approach: “K-means algorithm”

aka Lloyd-Max algorithm [Steinhaus 1956, Lloyd 1957 (publ. 1982)]

several passes on the training set

Naive sketching =histograms

bins of size within domain of radius R

exponential sketch size compressive sensing suggests can we avoid discretization (bypass curse of dimensionality) ?

✏ N = O((R/✏)d) m = N m = O(k log N) = O(kd log(R/✏))

18

`(x, ✓) = min

i

kx ✓ik2 θ = {θ1, . . . , θk}, θi ∈ Rd

slide-39
SLIDE 39
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Choice of the sketching function ?

Observation: distribution p(x) is spatially localized Intuition (from compressive sensing) need “incoherent” sampling choose Fourier measurements = empirical characteristic function

Sketching function Random Fourier Features [Rahimi & Recht 2007] Sketch vector z = Random Fourier Moments

Compressive K-means: How to Sketch ?

ω` ∈ Rd

1 ≤ ` ≤ m

19

Φ(x) =

1 √m

⇣ ej!>

` x⌘m

`=1

z` ≈ EX∼pejw>

` X

slide-40
SLIDE 40
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Learning centroids from a sketch ?

Learning principle = moment fitting

Parametric optimization problem ✓Statistical guarantees (assume -separated centroids)

20

ˆ θ(z) = arg min

θj∈Rd min αj kz k

X

j=1

αjΦ(θj)k2

m ≈ O(k2d log(R/✏))

compare FoCM 2017 version:

m ≈ O(k2d2 log(R/✏))

slide-41
SLIDE 41
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Learning centroids from a sketch ?

Learning principle = moment fitting

Parametric optimization problem ✓Statistical guarantees (assume -separated centroids)

Empirical learning algorithms

Inspiration: sparse recovery algorithms

Discretization + convex relaxation [Bunea & al 2010] Convex optimization over (sparse) Radon measures [e.g. Bredies & al 2013] CL-OMPR: greedy and gridless [Keriven, Bourrier, G. & Perez 2016]

MP (Mallat & Zhang 93) > OMP (Pati & al 93) > OMPR (Jain 2011) > CL-OMPR

similar to Frank-Wolfe [Bredies & al 2013] CL-AMP: hybrid approximate message passing [Byrne, G. & Schniter 2017]

20

ˆ θ(z) = arg min

θj∈Rd min αj kz k

X

j=1

αjΦ(θj)k2

m ≈ O(k2d log(R/✏))

compare FoCM 2017 version:

m ≈ O(k2d2 log(R/✏))

slide-42
SLIDE 42
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Nicolas Keriven

CL-OMPR & the SketchMLbox

SketchMLbox (sketchml.gforge.inria.fr)

  • Mixture of Diracs (« K-means »)
  • GMMs with known covariance
  • GMMs with unknown diagonal covariance
  • Soon:
  • Mixtures of alpha-stable
  • Gaussian Locally Linear Mapping [Deleforge 2014]
  • Handles generic mixtures

with user-defined

26/28

Mpθ, rθMpθ

21

p =

k

X

j=1

αjpθj

slide-43
SLIDE 43
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Sketch size: theory vs experiments

In theory, sufficient to have Empirically ?

22

m & O(k2d)

slide-44
SLIDE 44
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Sketch size: theory vs experiments

K-means GMMs, diagonal cov. GMMs, known cov. In theory, sufficient to have Empirically ?

Relative loss Relative loglike Relative loglike

22

E`(X, ΘSketch) E`(X, ΘLloyd)

m & O(k2d)

slide-45
SLIDE 45

Learning from random moments: the concept Compressive Statistical Learning (guarantees) Recent developments & perspectives

  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

23

slide-46
SLIDE 46
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Private sketched learning ?

“Natural” privacy of an aggregated estimator:

role of sketch size

sufficiently large for “task-level” information-preservation sufficiently small for “sample-level” information loss?

Towards guaranteed differential privacy ?

randomized sketching function ?

noise on training samples noise on random features partial random features combinations of the above …

24

z = 1

n n

X

i=1

Φ(xi)

Ψ(xi) = Φ(xi + ξi)

Ψ(xi) = Φ(xi) + ξi Ψ(xi) = diag(di) · Φ(xi) ∈ Cm

kdik0 = αm, α < 1

slide-47
SLIDE 47
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Hot from the oven

Compressive k-means in Sketch size Private sketching

independent draws of for each training sample

Tradeoff privacy / size of training set / quality

25

Ψ(xi) = diag(di) · Φ(xi) ∈ Cm

Rd

m = 10kd

k = d = 10

di

kdik0

10kd

kd

d

d/10

d/100

slide-48
SLIDE 48
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Hot from the oven

Compressive k-means in Sketch size Private sketching

independent draws of for each training sample

Tradeoff privacy / size of training set / quality

25

Ψ(xi) = diag(di) · Φ(xi) ∈ Cm

Rd

m = 10kd

k = d = 10

di

E`(X, ΘSketch) E`(X, ΘLloyd)

kdik0

10kd

kd

d

d/10

d/100

slide-49
SLIDE 49
  • R. GRIBONVAL

Inverse Problems and Machine Learning, Caltech, February 2018

Summary

✓ Dimension reduction ✓ Empirical success ✓ Statistical guarantees ➡compressive PCA ➡compressive k-means ➡compressive GMM ✓ Sketching framework

z

see also [Bruna

✤Next challenges:

  • provably good recovery algorithms ?
  • sketches for other learning tasks ?
  • privacy guarantees ?

26

  • 1
  • 0.5
0.5 1
  • Dim. 5
  • 1
  • 0.5
0.5 1
  • Dim. 6
  • 1
  • 0.5
0.5 1
  • Dim. 5
  • 1
  • 0.5
0.5 1
  • Dim. 6
  • 1
  • 0.5
0.5 1
  • Dim. 5
  • 1
  • 0.5
0.5 1
  • Dim. 6
  • 1
  • 0.5
0.5 1
  • Dim. 5
  • 1
  • 0.5
0.5 1
  • Dim. 6
  • 1
  • 0.5
0.5 1
  • Dim. 5
  • 1
  • 0.5
0.5 1
  • Dim. 6
  • 1
  • 0.5
0.5 1
  • Dim. 5
  • 1
  • 0.5
0.5 1
  • Dim. 6
slide-50
SLIDE 50

TH###NKS #

27 toolbox sketchml.gforge.inria.fr preprint arxiv.org/abs/1706.07180