Organizing Deep Networks Edouard Oyallon advisor: Stphane Mallat - - PowerPoint PPT Presentation

organizing deep networks
SMART_READER_LITE
LIVE PREVIEW

Organizing Deep Networks Edouard Oyallon advisor: Stphane Mallat - - PowerPoint PPT Presentation

DATA 1 Organizing Deep Networks Edouard Oyallon advisor: Stphane Mallat following the works of Laurent Sifre, Joan Bruna, collaborators : Eugene Belilovsky, Sergey Zagoruyko, Bogdan Cirstea, Jrn Jacobsen, 2 DATA Classification


slide-1
SLIDE 1

DATA Edouard Oyallon

1

Organizing Deep Networks

advisor: Stéphane Mallat

following the works of Laurent Sifre, Joan Bruna, …

collaborators: Eugene Belilovsky, Sergey Zagoruyko, Bogdan Cirstea, Jörn Jacobsen, …

slide-2
SLIDE 2

DATA

Classification of signals

  • Let , random variables
  • Problem: Estimate such that
  • We are given a training set to build
  • Say one can write , Classifier being

built with

  • 3 ways to build :


Supervised Unsupervised Predefined

2

n > 0 (X, Y ) ∈ Rn × Y ˆ y (xi, yi) ∈ Rn × Y ˆ y ˆ y = Classifier(Φx) (Φxi, yi) (xi)i (xi, yi)i Geometric priors Φ Y = { } n = 2

Classifier

,

w

<latexit sha1_base64="eqfykHkxP29JLdLAGokHSA1pBNI=">AAAB53icbZDLSgMxFIbP1Futt6pLN8EiuCozCupKC25ctuDYQjuUTHqmjc1khiSjlNIncONCxa3P4hu4821MLwtt/SHw8f/nkHNOmAqujet+O7ml5ZXVtfx6YWNza3unuLt3p5NMMfRZIhLVCKlGwSX6hhuBjVQhjUOB9bB/Pc7rD6g0T+StGaQYxLQrecQZNdaqPbaLJbfsTkQWwZtB6erz9LQCANV28avVSVgWozRMUK2bnpuaYEiV4UzgqNDKNKaU9WkXmxYljVEHw8mgI3JknQ6JEmWfNGTi/u4Y0ljrQRzaypianp7PxuZ/WTMz0UUw5DLNDEo2/SjKBDEJGW9NOlwhM2JggTLF7ayE9aiizNjbFOwRvPmVF8E/KZ+V3ZpXqlzCVHk4gEM4Bg/OoQI3UAUfGCA8wQu8OvfOs/PmvE9Lc86sZx/+yPn4AYu+jmw=</latexit> <latexit sha1_base64="O0lQJ2xoQXj8ACW3DbgTbPBIgF0=">AAAB53icbZDLSgNBEEVr4ivGV9Slm8YguAozCupKA25cJuCYQDKEnk5N0qbnQXePEoZ8gRsXKm79Fv/AnX9jZ5KFJl5oONxbRVeVnwiutG1/W4Wl5ZXVteJ6aWNza3unvLt3p+JUMnRZLGLZ8qlCwSN0NdcCW4lEGvoCm/7wepI3H1AqHke3epSgF9J+xAPOqDZW47FbrthVOxdZBGcGlavP01z1bvmr04tZGmKkmaBKtR070V5GpeZM4LjUSRUmlA1pH9sGIxqi8rJ80DE5Mk6PBLE0L9Ikd393ZDRUahT6pjKkeqDms4n5X9ZOdXDhZTxKUo0Rm34UpILomEy2Jj0ukWkxMkCZ5GZWwgZUUqbNbUrmCM78yovgnlTPqnbDqdQuYaoiHMAhHIMD51CDG6iDCwwQnuAFXq1769l6s96npQVr1rMPf2R9/ABGL474</latexit>

w ˆ y = arg inf˜

y E(|˜

y(X) − Y |)

slide-3
SLIDE 3

DATA

  • Caltech 101, etc

3 3

Training set to
 predict labels

"Rhino" Not a "rhino"

"Rhinos"

High Dimensional classification

− → ˆ y(x)? Estimation problem (xi, yi) ∈ R2242 × {1, ..., 1000}, i < 106

slide-4
SLIDE 4

DATA

High-dimensional variabilities

  • Claim: In , the variance is huge.


Ex.:


  • Claim: Small deformations (not parametric) can have

huge effects:
 Ex.:


  • The variance is high, and the bias is difficult to
  • estimate. There are also few available samples…


How to handle that?

4

X ∼ N(0, In) 9C > 0, 8n, P(kXk t))  2e− t2

Cn

Lτx(u) = x(u − τ(u)) ⌧(u) = ✏, C ⇢ R2, k1C Lτ1Ck = 2 Rn, n 1

then

τ ∈ C∞ x ∈ L2(Rn),

define

x y kx yk2 = 2 E(X) = 0

slide-5
SLIDE 5

DATA

Image variabilities

5

Geometric variability Class variability

Groups acting on images: translation, rotation, scaling

Intraclass variability Extraclass variability

Other sources : luminosity, occlusion, small deformations

Not informative

I − τ High variance: how to reduce it? Lτx(u) = x(u − τ(u)), τ ∈ C∞

slide-6
SLIDE 6

DATA

Fighting the curse of dimensionality

  • Objective: building a representation of such that a

simple (say euclidean) classifier can estimate the label : 
 
 


  • Designing consist of building an approximation of a

low dimensional space which is regular with respect to the class:

  • Necessary dimensionality reduction


6

Φx x Φ

Φ w

<latexit sha1_base64="eqfykHkxP29JLdLAGokHSA1pBNI=">AAAB53icbZDLSgMxFIbP1Futt6pLN8EiuCozCupKC25ctuDYQjuUTHqmjc1khiSjlNIncONCxa3P4hu4821MLwtt/SHw8f/nkHNOmAqujet+O7ml5ZXVtfx6YWNza3unuLt3p5NMMfRZIhLVCKlGwSX6hhuBjVQhjUOB9bB/Pc7rD6g0T+StGaQYxLQrecQZNdaqPbaLJbfsTkQWwZtB6erz9LQCANV28avVSVgWozRMUK2bnpuaYEiV4UzgqNDKNKaU9WkXmxYljVEHw8mgI3JknQ6JEmWfNGTi/u4Y0ljrQRzaypianp7PxuZ/WTMz0UUw5DLNDEo2/SjKBDEJGW9NOlwhM2JggTLF7ayE9aiizNjbFOwRvPmVF8E/KZ+V3ZpXqlzCVHk4gEM4Bg/OoQI3UAUfGCA8wQu8OvfOs/PmvE9Lc86sZx/+yPn4AYu+jmw=</latexit> <latexit sha1_base64="O0lQJ2xoQXj8ACW3DbgTbPBIgF0=">AAAB53icbZDLSgNBEEVr4ivGV9Slm8YguAozCupKA25cJuCYQDKEnk5N0qbnQXePEoZ8gRsXKm79Fv/AnX9jZ5KFJl5oONxbRVeVnwiutG1/W4Wl5ZXVteJ6aWNza3unvLt3p+JUMnRZLGLZ8qlCwSN0NdcCW4lEGvoCm/7wepI3H1AqHke3epSgF9J+xAPOqDZW47FbrthVOxdZBGcGlavP01z1bvmr04tZGmKkmaBKtR070V5GpeZM4LjUSRUmlA1pH9sGIxqi8rJ80DE5Mk6PBLE0L9Ikd393ZDRUahT6pjKkeqDms4n5X9ZOdXDhZTxKUo0Rm34UpILomEy2Jj0ukWkxMkCZ5GZWwgZUUqbNbUrmCM78yovgnlTPqnbDqdQuYaoiHMAhHIMD51CDG6iDCwwQnuAFXq1769l6s96npQVr1rMPf2R9/ABGL474</latexit>

kΦx Φx0k n 1 ) ˆ y(x) = ˆ y(x0) RD Rd ˆ y y D d

slide-7
SLIDE 7

DATA Averaging makes euclidean distance meaningful in high dimension

7

x y kx yk2 = 2

Averaging is the key to get invariants

x y Translation Rotation

slide-8
SLIDE 8

DATA

An example: Invariance to translation

8

  • In many cases, one wish to be invariant globally to translation, a

simple way is to perform an averaging:
 


  • Even if it can be localized, the averaging keeps the low frequency

structures: the invariance brings a loss of information!

  • Bias issue! How do we recover the missing information?

Ax = Z Laxda = Z x(u)du A

It’s the 0 frequency!

ALa = A Lax(u) = x(u − a) Translation operator

slide-9
SLIDE 9

DATA

Necessary mechanism: Separation - Contraction

  • In high dimension, typical distances are huge, thus an

appropriate representation must contract the space:


  • While avoiding the different classes to collapse:

9

kΦx Φx0k  kx x0k 9✏ > 0, y(x) 6= y(x0) ) kΦx Φx0k ✏ ✏

Φ

slide-10
SLIDE 10

DATA

Deep learning: Technical breakthrough

10

  • Deep learning has permitted to solve a large number of

task that were considered as extremely challenging for a computer.

  • The technique that is used is generic and its success

implies that it reduces those sources of variability.

  • Previous properties hold for deep learning.
  • How, why?
slide-11
SLIDE 11

DATA

11

x0 x1 x2

Classifier

Ref.: ImageNet Classification with Deep Convolutional Neural Networks. A Krizhevsky et al.

DeepNetwork ρ0W0 ρ1W1 ρJ−1WJ−1 xJ = Φx xj+1 = ρjWjxj

linear operator non-linear operator

The kernel is learned xj+1(u, ) = ⇢( X

˜ λ

xj(., ˜ ) ? wj,λ,˜

λ(u))

slide-12
SLIDE 12

DATA

Why mathematics about deep learning are important

  • Pure black box. Few mathematical results are available.

Many rely on a "manifold hypothesis". Clearly wrong:


Ex: stability to diffeomorphisms


  • No stability results. It means that "small" variations of

the inputs might have a large impact on the system. And this happens.

  • No generalisation result. Rademacher complexity can

not explain the generalization properties.

  • Shall we learn each layer from scratch? (geometric

priors?) The deep cascade makes features are hard to interpret

12

Ref.: Intriguing properties of neural networks.

  • C. Szegedy et al.

Ref.: Understanding deep learning requires rethinking generalization

  • C. Zhang et al.

Ref.: Deep Roto-Translation Scattering for Object Classification. EO and S Mallat

slide-13
SLIDE 13

DATA

Organization is a key

13

Answers Questions Answers Questions Answers Questions Answers Questions

Organizing questions Organizing answers

Both

  • Consider a problem of questionnaires: people answer to

0 or 1 to some question. What does structuration means?

neighbours become meaningful: local metrics

In general, works tackle only

  • ne of the aspect

Ref.: Harmonic Analysis of Digital Data Bases

Coifman R. et al. structuration à changer

slide-14
SLIDE 14

DATA

Organization permits creation of invariance

  • As (all) the sources of regularities are obtained,

interpolating new points is possible (in statistical terms: generalisation property!)
 


  • In the previous case, one can build a discriminative and

invariant representation: Haar wavelets on graphs for example.

14

Answers Questions

regularity

+

  • +

Ref.: Harmonic Analysis of Digital Data Bases

Coifman R. et al.

slide-15
SLIDE 15

DATA

Organising the CNN representation: Local Support Vectors

  • Let’s consider a CNN of depth J.
  • Local Support Vectors of order k at depth j:

representations at depth j that are well classified by a k- NN but not by a l-NN for l<k


  • They give a measure of the separation-contraction via:

15

2-LSV 4-LSV 0-LSV k-LSV, k>6 x(l)

j : l-NN at depth j

Γk+1

j

=

  • xj 2 Γk

j |card{y(x(l) j ) 6= y(x(l) j ), l  k + 1} > k

2

Ref.: Building a Regular Decision Boundary with Deep Networks

EO

Local dimension is intractable!

slide-16
SLIDE 16

DATA

Complexity measure

16

# of k-local support vectors at different depth n

Slow decay to stationary regime indicates high complexity (separation) Small amount indicates contraction

slide-17
SLIDE 17

DATA

An organisation of the representation

  • There is a progressive localisation which explains why a

1-NN (or a Gaussian SVM) works better with depth:

  • How do the representation got localized? Necessary

variability reduction

17

linear metrics are more meaningful in low dimension

slide-18
SLIDE 18

DATA

Identifying the variabilities?

  • Several works showed a Deepnet exhibits some

covariance:

  • Manifold of faces at a certain depth:
  • Can we use these?

18

Ref.: Understanding deep features with computer-generated imagery, M Aubry, B Russel Ref.: Unsupervised Representation Learning with Deep Convolutional GAN, Radford, Metz & Chintalah

slide-19
SLIDE 19

DATA

Linearizing variabilities

  • Weak differentiability property:
  • A linear projection (to kill ) build an invariant

19

sup

L

kΦLx Φxk kLx xk < 1 ) 9 ”weak” ∂xΦ ) ΦLx ⇡ Φx + ∂xΦL + o(kLk)

Φ

A linear operator Displacement

+ projection

L L

example: Scattering Transform

slide-20
SLIDE 20

DATA

Symmetry group hypothesis

  • To each classification problem corresponds a canonic

and unique symmetry group : 


  • We hypothesise there exists Lie groups and CNNs such

that:

  • Examples are given by the euclidean group:

20

∀gj ∈ Gj, φj(gj.x) = φj(x) where xj = φj(x) ∀x, ∀g ∈ G, Φx = Φg.x G G0 ⊂ G1 ⊂ ... ⊂ GJ ⊂ G

High dimensional

G0 = R2, G1 = G0 n SL2(R)

Ref.: Understanding deep convolutional networks

S Mallat

slide-21
SLIDE 21

DATA

Structuring the input with the Scattering Transform

  • Scattering Transform is a local descriptor of

neighbourhood of amplitude .

  • It is a representation built via geometry with limited
  • learning. (~SIFT)
  • Successfully used in several applications:
  • Digits
  • Textures

21

Small deformations +Translation

Ref.: Invariant Convolutional Scattering Network, J. Bruna and S Mallat Ref.: Rotation, Scaling and Deformation Invariant Scattering for texture discrimination, Sifre L and Mallat S.

Rotation+Scale

All variabilities are known

2J SJ

slide-22
SLIDE 22

DATA

Wavelets

  • Wavelets help to describe signal structures. is a

wavelet iff

  • They are chosen localised in space and frequency.
  • Wavelets can be dilated in order to be a multi-scale

representation of signals, rotated to describe rotations.

  • Design wavelets selective to an informative

variability.

22

ψ

<latexit sha1_base64="vZM3p0oDK8G//ZjqteTWKP152iI=">AAAB6nicbZDLSgMxFIbPtF5qvY2XnZtgEVyVGRcqrgpuXFZwbKEdSibNtKFJZkgyQhkKPoEbFypufQ4fwp1vY6btQlt/CHz8/znknBOlnGnjed9OqbyyurZe2ahubm3v7Lp7+/c6yRShAUl4otoR1pQzSQPDDKftVFEsIk5b0ei6yFsPVGmWyDszTmko8ECymBFsCqubatZza17dmwotgz+HWuPws/wIAM2e+9XtJyQTVBrCsdYd30tNmGNlGOF0Uu1mmqaYjPCAdixKLKgO8+msE3RinT6KE2WfNGjq/u7IsdB6LCJbKbAZ6sWsMP/LOpmJL8OcyTQzVJLZR3HGkUlQsTjqM0WJ4WMLmChmZ0VkiBUmxp6nao/gL668DMFZ/bzu3fq1xhXMVIEjOIZT8OECGnADTQiAwBCe4AVeHeE8O2/O+6y05Mx7DuCPnI8frjmPpA==</latexit> <latexit sha1_base64="MKReSqAJM21mHrPcc9EWVyc1loM=">AAAB6nicbZC7SgNBFIbPJl5ivMVLZzMYBKuwa6FiFbCxjOAaIVnC7GQ2GTIzO8zMCiHkFWwsVGx9DmsbGzvfxtkkhSb+MPDx/+cw55xYcWas7397heLS8spqaa28vrG5tV3Z2b01aaYJDUnKU30XY0M5kzS0zHJ6pzTFIua0GQ8u87x5T7VhqbyxQ0UjgXuSJYxgm1ttZVinUvVr/kRoEYIZVOv770X1+ZE1OpWvdjclmaDSEo6NaQW+stEIa8sIp+NyOzNUYTLAPdpyKLGgJhpNZh2jI+d0UZJq96RFE/d3xwgLY4YidpUC276Zz3Lzv6yV2eQ8GjGpMkslmX6UZBzZFOWLoy7TlFg+dICJZm5WRPpYY2LdecruCMH8yosQntROa/51UK1fwFQlOIBDOIYAzqAOV9CAEAj04QGe4NkT3qP34r1OSwverGcP/sh7+wFQoJGe</latexit>

ψ

<latexit sha1_base64="mduu63BoHMqGld23Vba2fG/RufU=">AAAB6nicbZC9TsMwFIVvyl8pfwFGGCwqJKYqYQDGChbGViK0UhtVjuu0Vm0nsh2kKuorsDAAYuUheA42Nh4Fp+0ALUey9Omce+V7b5Rypo3nfTmlldW19Y3yZmVre2d3z90/uNdJpggNSMIT1Y6wppxJGhhmOG2nimIRcdqKRjdF3nqgSrNE3plxSkOBB5LFjGBTWN1Us55b9WreVGgZ/DlU68cfzW8AaPTcz24/IZmg0hCOte74XmrCHCvDCKeTSjfTNMVkhAe0Y1FiQXWYT2edoFPr9FGcKPukQVP3d0eOhdZjEdlKgc1QL2aF+V/WyUx8FeZMppmhksw+ijOOTIKKxVGfKUoMH1vARDE7KyJDrDAx9jwVewR/ceVlCM5rFzWv6Vfr1zBTGY7gBM7Ah0uowy00IAACQ3iEZ3hxhPPkvDpvs9KSM+85hD9y3n8AnNqQXA==</latexit> <latexit sha1_base64="KoiyL79C8JDW2tvOqjdwVaE3Hps=">AAAB6nicbZDNSgMxFIXv+FvrX9WlIsEiuCozLtRl0Y3LFhxbaIeSSTNtaJIZkoxQhi7dunGh4taH6HO48xl8CTNtF9p6IPBxzr3k3hsmnGnjul/O0vLK6tp6YaO4ubW9s1va27/XcaoI9UnMY9UMsaacSeobZjhtJopiEXLaCAc3ed54oEqzWN6ZYUIDgXuSRYxgk1vtRLNOqexW3InQIngzKFePxvXvx+NxrVP6bHdjkgoqDeFY65bnJibIsDKMcDoqtlNNE0wGuEdbFiUWVAfZZNYROrVOF0Wxsk8aNHF/d2RYaD0Uoa0U2PT1fJab/2Wt1ERXQcZkkhoqyfSjKOXIxChfHHWZosTwoQVMFLOzItLHChNjz1O0R/DmV14E/7xyUXHrXrl6DVMV4BBO4Aw8uIQq3EINfCDQhyd4gVdHOM/Om/M+LV1yZj0H8EfOxw95uZHC</latexit>

ψ ∈ L2(R2, C) and R

R2 ψ(u)du = 0

<latexit sha1_base64="mrzq5xaVNdN46w+JQS8GOBQT3ZU=">AAACQnicbVBBaxNBFH7baptGW9N69DI0CAmUsJtD7aUQmosHD1GaJpiNYXZ2kgydnV1m3paGZaE/zYt/wJt49+KhivTmwdkkQkx8MPB933vfzJsvSKQw6LpfnK3tR493dkt75SdP9w+eVQ6Prkycasa7LJax7gfUcCkU76JAyfuJ5jQKJO8F1+2i37vh2ohYXeIs4cOITpQYC0bRSqPKez8xgvhCET+iOGVUZm/yD83anAVB9s6Sk7+kndd95LeYEapCklsXjrLVyZwU19XSepieu6NK1W248yKbwFuCaqv+8PUOADqjymc/jFkacYVMUmMGnpvgMKMaBZM8L/up4Qll13TCBxYqGnEzzOYZ5OSlVUIyjrU9CslcXXVkNDJmFgV2sljYrPcK8X+9QYrjs2EmVJIiV2zx0DiVBGNSBEpCoTlDObOAMi3sroRNqaYMbexlG4K3/uVN0G02ThvuW6/auoBFleAFHEMNPHgFLXgNHegCg4/wDe7hh/PJ+e78dH4tRrecpec5/FPO7z/pLLQC</latexit> <latexit sha1_base64="hbb0Ca2ujpjCwTZDix9084r07Wg=">AAACQnicbVC7TsMwFHV4U14FRhaLCqmVUJUwAAtSBQsDQ0GUIppSOY4LVh0nsm8QVZSNP2JkYeEH2BA7CwMgxMaA04LE60qWzjn3Hvv6eJHgGmz7zhoYHBoeGR0bz01MTk3P5GfnDnQYK8pqNBShOvSIZoJLVgMOgh1GipHAE6zudbayfv2MKc1DuQ/diDUDciJ5m1MCRmrlj9xIc+xyid2AwCklItlJj1eKPeZ5yZ4hy19kKy25wM4hwUT6ODUuaCXfJ1OcXVeMS368YbfyBbts9wr/Bc4nKFRKb/dnl1cX1Vb+1vVDGgdMAhVE64ZjR9BMiAJOBUtzbqxZRGiHnLCGgZIETDeTXgYpXjKKj9uhMkcC7qnfHQkJtO4GnpnMFta/e5n4X68RQ3u9mXAZxcAk7T/UjgWGEGeBYp8rRkF0DSBUcbMrpqdEEQom9pwJwfn95b+gtlJeLdu7TqGyifo1hhbQIioiB62hCtpGVVRDFF2jB/SEnq0b69F6sV77owPWp2ce/Sjr/QNzfLXq</latexit>

ψj,θ = 1 22j ψ(rθ(u) 2j )

<latexit sha1_base64="4yWyX8g1ftSjdJSEzIePbvZr7Ik=">AAACJXicbVDLSsNAFJ34rPVVdekmWIQWpCRF1E2h6MZlBWsLTQ2T6aSddvJg5kYoIV/jxl9x46KK4MpfcdJkoa0HBg7nnMude5yQMwmG8aWtrK6tb2wWtorbO7t7+6WDwwcZRILQNgl4ILoOlpQzn7aBAafdUFDsOZx2nMlN6neeqJAs8O9hGtK+h4c+cxnBoCS71LBCyex4fGbBiAJOGpYrMInNJK4/xvVxkqR+JROFnYUqUTW1x0nVLpWNmjGHvkzMnJRRjpZdmlmDgEQe9YFwLGXPNELox1gAI5wmRSuSNMRkgoe0p6iPPSr78fzMRD9VykB3A6GeD/pc/T0RY0/KqeeopIdhJBe9VPzP60XgXvVj5ocRUJ9ki9yI6xDoaWf6gAlKgE8VwUQw9VedjLCqBFSzRVWCuXjyMmnXaxc14+683LzO2yigY3SCKshEl6iJblELtRFBz+gVzdC79qK9aR/aZxZd0fKZI/QH2vcP94GmQg==</latexit> <latexit sha1_base64="4yWyX8g1ftSjdJSEzIePbvZr7Ik=">AAACJXicbVDLSsNAFJ34rPVVdekmWIQWpCRF1E2h6MZlBWsLTQ2T6aSddvJg5kYoIV/jxl9x46KK4MpfcdJkoa0HBg7nnMude5yQMwmG8aWtrK6tb2wWtorbO7t7+6WDwwcZRILQNgl4ILoOlpQzn7aBAafdUFDsOZx2nMlN6neeqJAs8O9hGtK+h4c+cxnBoCS71LBCyex4fGbBiAJOGpYrMInNJK4/xvVxkqR+JROFnYUqUTW1x0nVLpWNmjGHvkzMnJRRjpZdmlmDgEQe9YFwLGXPNELox1gAI5wmRSuSNMRkgoe0p6iPPSr78fzMRD9VykB3A6GeD/pc/T0RY0/KqeeopIdhJBe9VPzP60XgXvVj5ocRUJ9ki9yI6xDoaWf6gAlKgE8VwUQw9VedjLCqBFSzRVWCuXjyMmnXaxc14+683LzO2yigY3SCKshEl6iJblELtRFBz+gVzdC79qK9aR/aZxZd0fKZI/QH2vcP94GmQg==</latexit>

ψj,θ

<latexit sha1_base64="h0os0qPCCwGOdy3Iax5X9vow0TE=">AAAB93icbVBNS8NAEN34WetHox69LBbBg5REinosePFYwdhCE8Jmu2nXbjZhdyLU0F/ixYOKV/+KN/+N2zYHbX0w8Hhvhpl5USa4Bsf5tlZW19Y3Nitb1e2d3b2avX9wr9NcUebRVKSqGxHNBJfMAw6CdTPFSBIJ1olG11O/88iU5qm8g3HGgoQMJI85JWCk0K75meZh8XDmw5ABmYR23Wk4M+Bl4pakjkq0Q/vL76c0T5gEKojWPdfJICiIAk4Fm1T9XLOM0BEZsJ6hkiRMB8Xs8Ak+MUofx6kyJQHP1N8TBUm0HieR6UwIDPWiNxX/83o5xFdBwWWWA5N0vijOBYYUT1PAfa4YBTE2hFDFza2YDokiFExWVROCu/jyMvHOGxcN57ZZbzXLNCroCB2jU+SiS9RCN6iNPERRjp7RK3qznqwX6936mLeuWOXMIfoD6/MHXVOTCg==</latexit> <latexit sha1_base64="h0os0qPCCwGOdy3Iax5X9vow0TE=">AAAB93icbVBNS8NAEN34WetHox69LBbBg5REinosePFYwdhCE8Jmu2nXbjZhdyLU0F/ixYOKV/+KN/+N2zYHbX0w8Hhvhpl5USa4Bsf5tlZW19Y3Nitb1e2d3b2avX9wr9NcUebRVKSqGxHNBJfMAw6CdTPFSBIJ1olG11O/88iU5qm8g3HGgoQMJI85JWCk0K75meZh8XDmw5ABmYR23Wk4M+Bl4pakjkq0Q/vL76c0T5gEKojWPdfJICiIAk4Fm1T9XLOM0BEZsJ6hkiRMB8Xs8Ak+MUofx6kyJQHP1N8TBUm0HieR6UwIDPWiNxX/83o5xFdBwWWWA5N0vijOBYYUT1PAfa4YBTE2hFDFza2YDokiFExWVROCu/jyMvHOGxcN57ZZbzXLNCroCB2jU+SiS9RCN6iNPERRjp7RK3qznqwX6936mLeuWOXMIfoD6/MHXVOTCg==</latexit>

Isotropic

| ˆ ψ|

Non-Isotropic

| ˆ ψ| VS

slide-23
SLIDE 23

DATA The Gabor wavelet

23

φ(u) = 1 2πσ e− kuk2

ψ(u) = 1 2πσ e− kuk2

2σ (eiξ.u − κ) (for sake of simplicity, formula are given in the isotropic case)

Heisenberg principle! Good localisation in space and Fourier

slide-24
SLIDE 24

DATA

Wavelet Transform

  • Wavelet transform :
  • Isometric and linear operator of , with

  • Covariant with translation :
  • Nearly commutes with diffeomorphisms
  • A good baseline to describe an image!

24

Wx = {x ? j,θ, x ? J}θ,j≤J kWxk2 = X

θ,j≤J

Z |x ? j,θ|2 + Z x ? 2

J

ω1 ω2

Ref.: Group Invariant Scattering, Mallat S

L2 WLa = LaW La k[W, Lτ]k  Ckrτk

slide-25
SLIDE 25

DATA

Filter bank implementation

  • f a Fast WT
  • Assume it is possible to find and such that
  • Set:
  • The WT is then given by
  • A WT can be interpreted as a deep cascade of

linear operator, which is approximatively verified for the Gabor Wavelets.

25

ˆ φ(ω) = 1 √ 2 ˆ h(ω 2 )ˆ φ(ω 2 ) ˆ ψθ(ω) = 1 √ 2 ˆ gθ(ω 2 )ˆ φ(ω 2 )

Ref.: Fast WT, Mallat S, 89

h g and xj(u, 0) = x ? j(u) = h ? (x ? j−1)(2u) xj(u, ✓) = x ? j,θ(u) = gθ ? (x ? j−1)(2u) and Wx = {xj(., θ), xJ(., 0)}j≤J,θ

slide-26
SLIDE 26

DATA Implementation of a WT

26

h gθ

gθ h

h

There is an oversampling

x0 x1 x2 x3

h ≥ 0

step by step of the construction(and add modulus also)

ˆ φj = 1 √ 2 ˆ h( . 2)ˆ φj−1 ˆ ψj,θ = 1 √ 2 ˆ gθ( . 2)ˆ φj−1

slide-27
SLIDE 27

DATA

Scattering Transform

27

  • Scattering transform at scale J is the cascading of

complex WT with modulus non-linearity, followed by a low pass-filtering:


  • Mathematically well defined for a large class of

wavelets.

with λi = {ji, θi}, ji ≤ J

<latexit sha1_base64="UmL5ObII+VZe+lGikFghFEB1KH4=">AAACDXicbVA9SwNBEJ3z2/gVtdNmMQoWEu4sVAQhYCNWEYwRcuHY20zM6t6Hu3NCOFJa2fhXbCxUbO3t/DduEgu/Hiy8eW+G2XlhqqQh1/1wRkbHxicmp6YLM7Nz8wvFxaUzk2RaYE0kKtHnITeoZIw1kqTwPNXIo1BhPbw67Pv1G9RGJvEpdVNsRvwilm0pOFkpKG74yja3eCAP/PwykFs+dZBs6fe2mK19hdfsOCiW3LI7APtLvC9SqqzcroNFNSi++61EZBHGJBQ3puG5KTVzrkkKhb2CnxlMubjiF9iwNOYRmmY+OKfHNqzSYu1E2xcTG6jfJ3IeGdONQtsZceqY315f/M9rZNTea+YyTjPCWAwXtTPFKGH9bFhLahSkupZwoaX9KxMdrrkgm2DBhuD9PvkvqW2Xd8ruiVeq7MMQU7AKa7AJHuxCBY6gCjUQcAcP8ATPzr3z6Lw4r8PWEedrZhl+wHn7BCTZnHU=</latexit> <latexit sha1_base64="XZvxeIR+93C6VoTUkYgk/eMNrAA=">AAACDXicbVC7SgNBFJ31bXxF7bQZXAQLCbsWKoIQsBFBiGCMkF2W2cmNGZ19OHNXCEuwsrLxV2wsVNLa2/kj1k4eha8DA+eecy937glTKTQ6zoc1Mjo2PjE5NV2YmZ2bXyguLp3pJFMcqjyRiToPmQYpYqiiQAnnqQIWhRJq4dVBz6/dgNIiiU+xnYIfsYtYNAVnaKSguO5J09xggdj38stAbHrYAjSl19mkpvYkXNOjoGg7JacP+pe4Q2KXV+7sz9vjbiUovnuNhGcRxMgl07ruOin6OVMouIROwcs0pIxfsQuoGxqzCLSf98/p0HWjNGgzUebFSPvq94mcRVq3o9B0Rgxb+rfXE//z6hk2d/1cxGmGEPPBomYmKSa0lw1tCAUcZdsQxpUwf6W8xRTjaBIsmBDc3yf/JdWt0nbJOXHt8h4ZYIqskjWyQVyyQ8rkkFRIlXByTx7JM3mxHqwn69XqDlpHrOHMMvkB6+0Ledue9A==</latexit> <latexit sha1_base64="XZvxeIR+93C6VoTUkYgk/eMNrAA=">AAACDXicbVC7SgNBFJ31bXxF7bQZXAQLCbsWKoIQsBFBiGCMkF2W2cmNGZ19OHNXCEuwsrLxV2wsVNLa2/kj1k4eha8DA+eecy937glTKTQ6zoc1Mjo2PjE5NV2YmZ2bXyguLp3pJFMcqjyRiToPmQYpYqiiQAnnqQIWhRJq4dVBz6/dgNIiiU+xnYIfsYtYNAVnaKSguO5J09xggdj38stAbHrYAjSl19mkpvYkXNOjoGg7JacP+pe4Q2KXV+7sz9vjbiUovnuNhGcRxMgl07ruOin6OVMouIROwcs0pIxfsQuoGxqzCLSf98/p0HWjNGgzUebFSPvq94mcRVq3o9B0Rgxb+rfXE//z6hk2d/1cxGmGEPPBomYmKSa0lw1tCAUcZdsQxpUwf6W8xRTjaBIsmBDc3yf/JdWt0nbJOXHt8h4ZYIqskjWyQVyyQ8rkkFRIlXByTx7JM3mxHqwn69XqDlpHrOHMMvkB6+0Ledue9A==</latexit> <latexit sha1_base64="KBet0gCnjjeHoNqWcqCFkvXlYrc=">AAAB43icbVBNS8NAEJ34WeNX9eplsQieSuJBPQpePFYwttCGstls2qWb3bA7EUroH/DgRfHqf/Lmv3H7AWrrg4HHezPMzEsKKSwGwZe3tr6xubVd2/F39/z9g8P60aPVpWE8Ylpq00mo5VIoHqFAyTuF4TRPJG8no9up337ixgqtHnBc8DinAyUywSg6qdWvN4JmMANZJeGCNGCBfv2zl2pW5lwhk9TabhgUGFfUoGCST/xeaXlB2YgOeNdRRXNu42p25oScOSUlmTauFJKZ+nuiorm14zxxnTnFoV32puJ/XrfE7DquhCpK5IrNF2WlJKjJ9GeSCsMZyrEjlBnhbiVsSA1l6JLxXQbh8serJLpoXjaD+58soAYncArnEMIV3MAdtCACBik8w6s39F68N+993rjmLSaO4Q+8j2+GborS</latexit> <latexit sha1_base64="gWxPYxqNM5VZOkDkIViilIi+kJI=">AAACDXicbVA9SwNBEJ2LXzF+RS1tFiVgEcLFQm0EwUasFIwKuXDs7c2ZNXsf7s4J4cgvsPGv2Fio2Nrb+W/cxICfDxbevDfD7LwgU9KQ6747pYnJqemZ8mxlbn5hcam6vHJm0lwLbIlUpfoi4AaVTLBFkhReZBp5HCg8D3oHQ//8BrWRaXJK/Qw7Mb9MZCQFJyv51ZqnbHPIfbnnFVe+rHvURbKlN6gzW3sKr9mRX91wG+4I7C9pjskGjHHsV9+8MBV5jAkJxY1pN92MOgXXJIXCQcXLDWZc9Pglti1NeIymU4zOGbCaVUIWpdq+hNhI/T5R8NiYfhzYzphT1/z2huJ/XjunaLdTyCTLCRPxuSjKFaOUDbNhodQoSPUt4UJL+1cmulxzQTbBig2h+fvkv6S11dhuuCdfYUAZ1mAdNqEJO7APh3AMLRBwC/fwCE/OnfPgPDsvn60lZzyzCj/gvH4Aofia9w==</latexit> <latexit sha1_base64="gWxPYxqNM5VZOkDkIViilIi+kJI=">AAACDXicbVA9SwNBEJ2LXzF+RS1tFiVgEcLFQm0EwUasFIwKuXDs7c2ZNXsf7s4J4cgvsPGv2Fio2Nrb+W/cxICfDxbevDfD7LwgU9KQ6747pYnJqemZ8mxlbn5hcam6vHJm0lwLbIlUpfoi4AaVTLBFkhReZBp5HCg8D3oHQ//8BrWRaXJK/Qw7Mb9MZCQFJyv51ZqnbHPIfbnnFVe+rHvURbKlN6gzW3sKr9mRX91wG+4I7C9pjskGjHHsV9+8MBV5jAkJxY1pN92MOgXXJIXCQcXLDWZc9Pglti1NeIymU4zOGbCaVUIWpdq+hNhI/T5R8NiYfhzYzphT1/z2huJ/XjunaLdTyCTLCRPxuSjKFaOUDbNhodQoSPUt4UJL+1cmulxzQTbBig2h+fvkv6S11dhuuCdfYUAZ1mAdNqEJO7APh3AMLRBwC/fwCE/OnfPgPDsvn60lZzyzCj/gvH4Aofia9w==</latexit> <latexit sha1_base64="jGrKMMN/j+MNpd5niMDb/Px4PNI=">AAACDXicbVC7SgNBFJ31GeMramkzGAIWIWwsVAQhYCNWEVwTyC7L7OQmGTP7cOauEJZ8gY2/YmOhYmtv5984eRSaeGDg3HPu5c49QSKFRtv+thYWl5ZXVnNr+fWNza3tws7urY5TxcHhsYxVM2AapIjAQYESmokCFgYSGkH/YuQ3HkBpEUc3OEjAC1k3Eh3BGRrJL5RcaZrbzBfnbnbni7KLPUBTusMyNbUr4Z5e+YWiXbHHoPOkOiVFMkXdL3y57ZinIUTIJdO6VbUT9DKmUHAJw7ybakgY77MutAyNWAjay8bnDGnJKG3aiZV5EdKx+nsiY6HWgzAwnSHDnp71RuJ/XivFzqmXiShJESI+WdRJJcWYjrKhbaGAoxwYwrgS5q+U95hiHE2CeRNCdfbkeeIcVY4r9rVdrJ1N08iRfXJADkmVnJAauSR14hBOHskzeSVv1pP1Yr1bH5PWBWs6s0f+wPr8Acbqm3E=</latexit> <latexit sha1_base64="BSW+/Z7XFttMFpRW/beCc1nU/m8=">AAACDXicbVC7SgNBFJ31GeNr1dJmMAQsQti1UBGEgI1YRTAmkA3L7OQmGTP7cOauEJZ8gY2/YmOhYmtv5984eRSaeGDg3HPu5c49QSKFRsf5thYWl5ZXVnNr+fWNza1te2f3Vsep4lDjsYxVI2AapIighgIlNBIFLAwk1IP+xcivP4DSIo5ucJBAK2TdSHQEZ2gk3y560jS3mS/OvezOFyUPe4Cm9IYlampPwj298u2CU3bGoPPEnZICmaLq219eO+ZpCBFyybRuuk6CrYwpFFzCMO+lGhLG+6wLTUMjFoJuZeNzhrRolDbtxMq8COlY/T2RsVDrQRiYzpBhT896I/E/r5li57SViShJESI+WdRJJcWYjrKhbaGAoxwYwrgS5q+U95hiHE2CeROCO3vyPKkdlY/LzrVbqJxN08iRfXJADolLTkiFXJIqqRFOHskzeSVv1pP1Yr1bH5PWBWs6s0f+wPr8Acc6m3I=</latexit> <latexit sha1_base64="XZvxeIR+93C6VoTUkYgk/eMNrAA=">AAACDXicbVC7SgNBFJ31bXxF7bQZXAQLCbsWKoIQsBFBiGCMkF2W2cmNGZ19OHNXCEuwsrLxV2wsVNLa2/kj1k4eha8DA+eecy937glTKTQ6zoc1Mjo2PjE5NV2YmZ2bXyguLp3pJFMcqjyRiToPmQYpYqiiQAnnqQIWhRJq4dVBz6/dgNIiiU+xnYIfsYtYNAVnaKSguO5J09xggdj38stAbHrYAjSl19mkpvYkXNOjoGg7JacP+pe4Q2KXV+7sz9vjbiUovnuNhGcRxMgl07ruOin6OVMouIROwcs0pIxfsQuoGxqzCLSf98/p0HWjNGgzUebFSPvq94mcRVq3o9B0Rgxb+rfXE//z6hk2d/1cxGmGEPPBomYmKSa0lw1tCAUcZdsQxpUwf6W8xRTjaBIsmBDc3yf/JdWt0nbJOXHt8h4ZYIqskjWyQVyyQ8rkkFRIlXByTx7JM3mxHqwn69XqDlpHrOHMMvkB6+0Ledue9A==</latexit> <latexit sha1_base64="XZvxeIR+93C6VoTUkYgk/eMNrAA=">AAACDXicbVC7SgNBFJ31bXxF7bQZXAQLCbsWKoIQsBFBiGCMkF2W2cmNGZ19OHNXCEuwsrLxV2wsVNLa2/kj1k4eha8DA+eecy937glTKTQ6zoc1Mjo2PjE5NV2YmZ2bXyguLp3pJFMcqjyRiToPmQYpYqiiQAnnqQIWhRJq4dVBz6/dgNIiiU+xnYIfsYtYNAVnaKSguO5J09xggdj38stAbHrYAjSl19mkpvYkXNOjoGg7JacP+pe4Q2KXV+7sz9vjbiUovnuNhGcRxMgl07ruOin6OVMouIROwcs0pIxfsQuoGxqzCLSf98/p0HWjNGgzUebFSPvq94mcRVq3o9B0Rgxb+rfXE//z6hk2d/1cxGmGEPPBomYmKSa0lw1tCAUcZdsQxpUwf6W8xRTjaBIsmBDc3yf/JdWt0nbJOXHt8h4ZYIqskjWyQVyyQ8rkkFRIlXByTx7JM3mxHqwn69XqDlpHrOHMMvkB6+0Ledue9A==</latexit> <latexit sha1_base64="XZvxeIR+93C6VoTUkYgk/eMNrAA=">AAACDXicbVC7SgNBFJ31bXxF7bQZXAQLCbsWKoIQsBFBiGCMkF2W2cmNGZ19OHNXCEuwsrLxV2wsVNLa2/kj1k4eha8DA+eecy937glTKTQ6zoc1Mjo2PjE5NV2YmZ2bXyguLp3pJFMcqjyRiToPmQYpYqiiQAnnqQIWhRJq4dVBz6/dgNIiiU+xnYIfsYtYNAVnaKSguO5J09xggdj38stAbHrYAjSl19mkpvYkXNOjoGg7JacP+pe4Q2KXV+7sz9vjbiUovnuNhGcRxMgl07ruOin6OVMouIROwcs0pIxfsQuoGxqzCLSf98/p0HWjNGgzUebFSPvq94mcRVq3o9B0Rgxb+rfXE//z6hk2d/1cxGmGEPPBomYmKSa0lw1tCAUcZdsQxpUwf6W8xRTjaBIsmBDc3yf/JdWt0nbJOXHt8h4ZYIqskjWyQVyyQ8rkkFRIlXByTx7JM3mxHqwn69XqDlpHrOHMMvkB6+0Ledue9A==</latexit>

Ref.: Group Invariant Scattering, Mallat S

retirer l’ordre 2

SJx = {x ? J, |x ? λ1| ? J, ||x ? λ1| ? λ2| ? J} x

φJ ψλ |.| φJ ψλ |.| φJ

Depth

  • rder 0
  • rder 1
  • rder 2
slide-28
SLIDE 28

DATA Scattering as a CNN

28

h

gθ h

h gθ

J = 3, θ ∈ {0, π 4 , π 2 , 3π 4 } x0 x1 x2 x3

Order 2 Order 1 Order 0 Modulus

h

h

h ≥ 0

Scattering coefficients are only at the output

Ref.: Deep Roto-Translation Scattering for Object Classification. EO and S Mallat

ˆ φ(ω) = 1 √ 2 ˆ h(ω 2 )ˆ φ(ω 2 ) ˆ ψθ(ω) = 1 √ 2 ˆ gθ(ω 2 )ˆ φ(ω 2 )

slide-29
SLIDE 29

DATA

Analytic wavelets and modulus?

  • For any translations :
  • A modulus removes the phase!

29

ω ω0 | ˆ ψ(ω)| ω0

Rectificatio Modulu

Real and imaginary Phase artifact

Ref.: Group Invariant Scattering, Mallat S

Non-linear projection

the infinitesimal generator

  • f translations is the derivative…

\ Lax ? (!) = eiωT aˆ x(!) ˆ (!) = X

n

(i!T a)n n! ˆ x(!) ˆ (!) ≈ X

n

(i!T

0 a)n

n! ˆ x(!) ˆ (!) = eiωT

0 a [

x ? (!)

ωT a ˆ ψ(ω) ≈ ωT

0 a ˆ

ψ(ω)

slide-30
SLIDE 30

DATA

Information loss Reconstruction

30

Ref.: Mallat S, Bruna J

arg inf

y kS3x S3yk

x ˜ y invariance up to pixels 23

slide-31
SLIDE 31

DATA

Wavelets on Lie group

31

  • Discovering more complex groups is necessary to build more

complex invariants:


  • A wavelet is defined by and can be

dilated via

  • Theorem: Let G be a compact Lie group, for appropriate

mother wavelet and then
 
 
 is an isometry and covariant with the action of G

  • Proposition: almost commutes with deformations but is

not invariant to translation…

k[W, Lτ]k  Ckτk ψλ = Lλψ Λ Wx = { Z

G

x, x ?G λ}λ∈Λ ψ

Ref.: Group Invariant Scattering, Mallat S Ref.: Stein, E. M. Topics in harmonic analysis related to the Littlewood-Paley theory. 


W ψ ∈ L2(G), ˆ ψ(e) = 0 … vs

Translation Scattering does not see the difference

R2 , → SO2(R) o R2 , → ...

Ref.: Deep Roto-Translation Scattering for Object Classification. EO and S Mallat

slide-32
SLIDE 32

DATA

An ideal input for a modern CNN

  • Scattering is stable:
  • Linearize small deformations:
  • Invariant by local translation:
  • For , , has a topology that is structured by


, and this structures the first layer also:

32

CNN x kSJx SJyk  kx yk Deformations SJ SJx(u, λ) λ u SO2(R) SJ(g.x)(u, λ) = SJx(g−1u, g−1λ) , g.SJx(u, λ) Lτx(u) = x(u − τ(u)) kSJLτx SJxk  Ckrτkkxk |a| ⌧ 2J ) SJLax ⇡ SJ ∀u∀g ∈ SO2(R), g.x(u) , x(g−1u) if then,

Ref.: Scaling the Scattering Transform: Deep Hybrid Networks EO, E Belilovsky, S Zagoruyko

slide-33
SLIDE 33

DATA

How much learning is really required?

33

Dataset Type Paper Accuracy Caltech101 Scattering 79.9 Unsupervised Ask the locals 77.3 Supervised DeepNet 91.4 CIFAR100 Scattering 56.8 Unsupervised RFL 54.2 Supervised DeepNet 65.4

Identical Representation

104

101

256 × 256

images classes color images

32 × 32 100 5.104

color images images classes CALTECH CIFAR Group representations are competitive with representations learned from data without labels

Ref.: Deep Roto-Translation Scattering for Object Classification. EO and S Mallat

slide-34
SLIDE 34

DATA

Benchmarking ImageNet

  • Cascading a modern CNN leads to almost state-of-the-

art result on Imagenet2012:

  • Demonstrates no loss of information + Less layers

34

ResNet x SJ

Ref.: Scaling the Scattering Transform: Deep Hybrid Networks EO, E Belilovsky, S Zagoruyko

slide-35
SLIDE 35

DATA

Shared Local Encoder

  • It is equivalent to encode the non-overlapping

scattering patches: the output of the 1x1 is a local descriptor of an image that leads to AlexNet performances.

35

convolution

24 24 24

1 × 1

W1 W1 W1 W2 W2 W2 W3 W3 W3 W4 W5 W6 S4 S4 S4

Good generalization

  • n Caltech101

Extremely constrained

Ref.: Scaling the Scattering Transform: Deep Hybrid Networks EO, E Belilovsky, S Zagoruyko

slide-36
SLIDE 36

DATA

Benchmarking Small data

  • Adding geometric prior regularises the CNN input, in

the particular case of limited samples situations, without reducing the number of parameters.

  • State-of-the-art results on STL10 and CIFAR10:

36

STL10: 5k training, 8k testing, 10 classes +100k unlabeled(not used!!) Cifar10, 10 classes keeping 100, 500 and 1000 samples and testing on 10k

Ref.: Scaling the Scattering Transform: Deep Hybrid Networks EO, E Belilovsky, S Zagoruyko

slide-37
SLIDE 37

DATA

Invariance to rotation

  • We evaluate the angular energy propagated for given

frequencies:

  • They are all localised in the low-frequency domain:

invariance to rotation is learned. (supports symmetry group hypothesis)

37

Ω(ωθ1, ωθ2) = X |W1(., ωθ1, ωθ2)|2

Ref.: Scaling the Scattering Transform: Deep Hybrid Networks EO, E Belilovsky, S Zagoruyko

slide-38
SLIDE 38

DATA

Multiscale Hiearchical CNN

  • Can we structure the next layers?
  • Introduce a CNN that is convolutional along each

direction, finally averaged:


  • For , we refer to the variable as an attribute that

discriminates previously obtained tensor.

  • performs an averaging along .

38

xj+1(v1, ..., vj, vj+1) = ⇢j(xj ?v1,...,vj vj+1)(v1, ..., vj) xj+1 = ρjWjxj vj xj Wj vj−2

Ref.: Multiscale Hierarchical Convolutional Networks J Jacobsen, EO, S Mallat, Smeulders AWM

xJ = X

vj,j≤J−2

xJ−1(v1, ..., vJ−1)

slide-39
SLIDE 39

DATA

Flattening the variability

  • An explicit invariant of any translations along

is built.

  • Completely structures the axis of the "channels" via

convolutions.

  • It aims at mapping the symmetries of into the

translations along .

39

(v1, ..., vj) Φx = xJ Gj = Rj, j ≤ J

Φ

Organizing the channels indexes

Ref.: Multiscale Hierarchical Convolutional Networks J Jacobsen, EO, S Mallat, Smeulders AWM

slide-40
SLIDE 40

DATA

Reducing the number of parameters

40

CIFAR10 CIFAR100

This implies an effective structuration

Ref.: Multiscale Hierarchical Convolutional Networks J Jacobsen, EO, S Mallat, Smeulders AWM

slide-41
SLIDE 41

DATA

Organization of the representation?

  • We observe that representations at several layers are

translated:

41

x x Φx Φx

Ref.: Multiscale Hierarchical Convolutional Networks J Jacobsen, EO, S Mallat, Smeulders AWM

slide-42
SLIDE 42

DATA

Conclusion

42

  • Structuration should be the topic of future research to

improve Deep neural networks

  • Check my webpage for softwares and papers: http://

www.di.ens.fr/~oyallon/

Thank you!