On Learning Sets of Symmetric Elements [1] [2] [1,3] [3] Haggai - - PowerPoint PPT Presentation

on learning sets of symmetric elements
SMART_READER_LITE
LIVE PREVIEW

On Learning Sets of Symmetric Elements [1] [2] [1,3] [3] Haggai - - PowerPoint PPT Presentation

ICML 2020 On Learning Sets of Symmetric Elements [1] [2] [1,3] [3] Haggai Maron, Or Litany, Gal Chechik, Ethan Fetaya [1] Nvidia Research [2] Stanford University [3] Bar-Ilan University Motivation and Overview Set Symmetry Previous


slide-1
SLIDE 1

ICML 2020

On Learning Sets of Symmetric Elements

Haggai Maron, Or Litany, Gal Chechik, Ethan Fetaya

[1] Nvidia Research [2] Stanford University [3] Bar-Ilan University

[1] [2] [1,3] [3]

slide-2
SLIDE 2

Motivation and Overview

slide-3
SLIDE 3

Set Symmetry

{ }

Input set Previous work (DeepSets, PointNet) targeted training a deep network over sets

Deep Net

x1 x2 ⋮ xm , x1 x2 ⋮ xm , x1 x2 ⋮ xm , …

slide-4
SLIDE 4

Set+Elements symmetry

{ }

Input set

Main challenge: What architecture is optimal when elements of the set have their own symmetries?

Deep Net

, , …

Both the set and its elements have symmetries.

slide-5
SLIDE 5

Deep Symmetric sets

{ }

Input image set Output

slide-6
SLIDE 6

Set symmetry: Order invariance/equivariance

{ }

=

slide-7
SLIDE 7

Set symmetry: Order invariance/equivariance

{ }

=

{ }

slide-8
SLIDE 8

Set symmetry: Order invariance/equivariance

{ }

=

{ }

slide-9
SLIDE 9

Element symmetry: Translation invariance/equivariance

{ }

slide-10
SLIDE 10

{ } { }

Element symmetry: Translation invariance/equivariance

slide-11
SLIDE 11

{ } { }

Element symmetry: Translation invariance/equivariance

slide-12
SLIDE 12

Applications

Modalities 1D signals 2D images 3D pointclouds Graph

slide-13
SLIDE 13

This paper

A principled approach for learning sets of complex elements (graphs, point clouds, images) Characterize maximally expressive linear layers that respect the symmetries (DSS layers) Prove universality results Experimentally demonstrate that DSS networks outperform baselines

slide-14
SLIDE 14

Previous work

slide-15
SLIDE 15

Deep sets [Zaheer et al. 2017]

slide-16
SLIDE 16

CNN CNN CNN

Siamese

Deep sets [Zaheer et al. 2017]

slide-17
SLIDE 17

CNN CNN CNN

Features Siamese

Deep sets [Zaheer et al. 2017]

slide-18
SLIDE 18

CNN CNN CNN Deep Sets

Features Siamese Deeps sets block

Deep sets [Zaheer et al.]

slide-19
SLIDE 19

Previous work: information sharing

Aittala and Durand, ECCV 2018 Sridhar et al., NeuriPS 2019 Liu et al., ICCV 2019

CNN CNN CNN CNN CNN CNN

Information sharing layer

slide-20
SLIDE 20

Our approach

slide-21
SLIDE 21

Invariance

Many Learning tasks are invariant to natural transformations (symmetries) More formally. Let be a subgroup: is invariant if , for all e.g. image classification

H ≤ Sn f : ℝn → ℝ f(τ ⋅ x) = f(x) τ ∈ H

τ f f

“Cat”

slide-22
SLIDE 22

Equivariance

Let be a subgroup: Equivariant if , e.g. edge detection

H ≤ Sn f(τ ⋅ x) = τ ⋅ f(x)

τ τ f f

slide-23
SLIDE 23

Invariant neural networks

Equivariant FC Invariant

  • Invariant by construction
slide-24
SLIDE 24

Deep Symmetric Sets

with symmetry group Want to be invariant/equivariant to both and the ordering Formally the symmetry group is

x1, …, xn ∈ ℝd G ≤ Sd G H = Sn × G ≤ Snd

G

slide-25
SLIDE 25

Main challenges

  • What is the space of linear equivariant layers for specific

?

H = SN × G

slide-26
SLIDE 26
  • What is the space of linear equivariant layers for a given

?

  • Can we compute these operators efficiently?

H = SN × G

Main challenges

slide-27
SLIDE 27
  • What is the space of linear equivariant layers for a given

?

  • Can we compute these operators efficiently?
  • Do we lose expressive power?

H = SN × G

  • invariant networks

H

  • invariant continuous functions

H

Continuous functions

Main challenges

slide-28
SLIDE 28
  • What is the space of linear equivariant layers for a given

?

  • Can we compute these operators efficiently?
  • Do we lose expressive power?

H = SN × G

  • invariant networks

H

  • invariant continuous functions

H

Continuous functions Gap?

Main challenges

slide-29
SLIDE 29

Theorem: Any linear −equivariant layer is of the form where are linear

  • equivariant functions

We call these layers Deep Sets for Symmetric elements layers (DSS)

SN × G L : ℝn×d → ℝn×d L(X)i = LG

1 (xi) + ∑ j≠i

LG

2 (xj)

LG

1 , LG 2

G

  • equivariant layers

H

slide-30
SLIDE 30

DSS for images

are images is the group of circular translations

  • equivariant layers are convolutions

x1, …, xn G 2D G

Single DSS layer

slide-31
SLIDE 31

DSS for images

are images is the group of circular translations

  • equivariant layers are convolutions

x1, …, xn G 2D G

CONV CONV CONV

Single DSS layer

slide-32
SLIDE 32

DSS for images

are images is the group of circular translations

  • equivariant layers are convolutions

x1, …, xn G 2D G

CONV CONV CONV

+

Single DSS layer

slide-33
SLIDE 33

DSS for images

are images is the group of circular translations

  • equivariant layers are convolutions

x1, …, xn G 2D G

CONV CONV CONV CONV

+

Single DSS layer

slide-34
SLIDE 34

DSS for images

Siamese part Information sharing part

CONV CONV CONV CONV

+ +

Single DSS layer

slide-35
SLIDE 35

Expressive power

Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS networks for

  • equivariant functions.

SN × G

slide-36
SLIDE 36

Expressive power

Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS networks for

  • equivariant functions.
  • Main tool:
  • Noether’s Theorem (Invariant theory)
  • For any finite group

, the ring of invariant polynomials is finitely generated.

  • Generators can be used to create continuous unique encodings for elements in

SN × G H ℝ[x1, . . . , xn]H ℝn×d/H

slide-37
SLIDE 37

Results

slide-38
SLIDE 38

Signal classification

slide-39
SLIDE 39

Image selection

slide-40
SLIDE 40

Shape selection

slide-41
SLIDE 41

Conclusions

A general framework for learning sets of complex elements Generalizes many previous works Expressivity results Works well in many tasks and data types