Unrolling Inference: The Recurrent Inference Machine Max Welling - - PowerPoint PPT Presentation

unrolling inference
SMART_READER_LITE
LIVE PREVIEW

Unrolling Inference: The Recurrent Inference Machine Max Welling - - PowerPoint PPT Presentation

Unrolling Inference: The Recurrent Inference Machine Max Welling University of Amsterdam / Qualcomm Canadian Institute for Advanced Research ML @ UvA (2 fte) (12fte) Machine Learning in Amsterdam (12fte) (3fte) (10fte) (4fte) Overview


slide-1
SLIDE 1

Unrolling Inference:

The Recurrent Inference Machine

Max Welling

University of Amsterdam / Qualcomm Canadian Institute for Advanced Research

slide-2
SLIDE 2

ML @ UvA

(12fte) (12fte) (3fte) (10fte) Machine Learning in Amsterdam (4fte) (2 fte)

slide-3
SLIDE 3

Overview

  • Meta learning
  • Recurrent Inference Machine
  • Application to MRI
  • Application radio astronomy
  • Conclusions
slide-4
SLIDE 4
  • Train an optimizer to choose the best parameter updates by solving many optimization

problems and learn the patterns.

  • Unroll gradient optimizer, then abstract into a parameterized computation graph, e.g. RNN

2016

slide-5
SLIDE 5

2017

  • Learning a planning algorithm to execute the best actions by

solve many different RL.

slide-6
SLIDE 6
  • One shot learning: meta-learn a learning

algorithm to classify from very few examples 2017

slide-7
SLIDE 7

2017

  • Learning a NN architecture using active learning / reinforcement learning

Bayesian optimization

slide-8
SLIDE 8
slide-9
SLIDE 9

The Recipe

  • Study the classical iterative optimization algorithm
  • Unroll the computation tree and cut it off at T steps (layers)
  • Generalize / parameterize the individual steps
  • Create targets at the last layer
  • Backpropagate through the ”deep network” to fit the parameters
  • Execute the network to make predictions
slide-10
SLIDE 10

Learning to Infer

  • Unroll a known iterative inference scheme (e.g. mean field, belief propagation)
  • Abstract into parameterized computation graph for fixed nr. iterations, e.g. RNN
  • Learn parameters of RNN using meta-learning (e.g. solving many inference problems)
slide-11
SLIDE 11

Graph Convolutions

Thomas Kipf

slide-12
SLIDE 12

Convolutions vs Graph Convolutions

vs.

slide-13
SLIDE 13

vs.

Convolutions vs Graph Convolutions

slide-14
SLIDE 14

Graph Convolutions

slide-15
SLIDE 15

Graph Convolutional Networks

Kipf & Welling ICLR (2017)

slide-16
SLIDE 16

Application to Airway Segmentation

(work in progress, with Rajhav Selvan & Thomas Kipf)

slide-17
SLIDE 17

Inverse Model

Inverse Model

Inverse Problems

Forward Model

Measurement

Forward Model

Quantity of interest

w/ Patrick Putzky

slide-18
SLIDE 18

The Usual Approach

generative model (known) prior (learn)

  • bservations

advantage: model P(X) and optimization are separated. disadvantage: accuracy suffers because model and optimization interact…

slide-19
SLIDE 19

Learning Inference: Recurrent Inference Machine

  • Abstract and parameterize computation graph into RNN
  • Integrate prior P(X) in RNN
  • Add memory state s
  • Meta learn the parameters of the RNN
slide-20
SLIDE 20

Recurrent Inference Machine (RIM)

Learn to optimize using a RNN. external information memory state CNN/RNN

slide-21
SLIDE 21

Recurrent Inference Machine

+

Embedding 5x5, 64 GRU 3x3, 64, atrous 2 Time 3x3 conv

slide-22
SLIDE 22

Recurrent Inference Machines in Time

Objective

slide-23
SLIDE 23

Simple Super-Resolution

Time

slide-24
SLIDE 24

Reconstruction from Random Projections

32 x 32 pixel image patches Fast Convergence on all tasks

slide-25
SLIDE 25

Image Denoising

Denoising trained on small image patches, generalises to full-sized images

slide-26
SLIDE 26

Super-resolution

LR HR Bicubic Interpolation RIM

slide-27
SLIDE 27

Super-resolution

slide-28
SLIDE 28

Square Kilometer Array

Up to 14.4 Gigapixels With thousands of Channels

Jorn Peters

slide-29
SLIDE 29
slide-30
SLIDE 30
slide-31
SLIDE 31
slide-32
SLIDE 32

Deep Learning for Inverse Problems

E.g. MRI Image Reconstruction

w/ Kai Lonning & Matthan Caan

slide-33
SLIDE 33

Example of training data point, 30x30 image patch Testing done on full images, sub- sampling masks shown for 6x, 4x and 2x acceleration

http://sbt.science.uva.nl/mri/about/

(Slides and website made by Kai Lonning)

slide-34
SLIDE 34
slide-35
SLIDE 35

A full brain RIM reconstruction, starting from the 4 times sub-sampled corruption

  • n the left, attempting to recover the target
  • n the right.
slide-36
SLIDE 36

Each time-step in the Recurrent Inference Machine produces a new estimate, here shown to the left, from the 3x accelerated corruption until the 10th and final

  • reconstruction. Target is in the middle,

while the error (not to scale) is shown to the right.

slide-37
SLIDE 37
slide-38
SLIDE 38
slide-39
SLIDE 39
slide-40
SLIDE 40
slide-41
SLIDE 41

Conclusions

  • Meta learning is interesting new paradigm that can improve classical
  • ptimization and inference algorithms by exploiting patterns in classes of problems.
  • RIM is a method that unrolls

inference and learns to solve inverse problems.

  • Great potential to improve

& speed up radio-astronomy and MRI image reconstruction.

  • Application to MRI-linac?