Regularization of Inverse Problems Yoeri Boink *, Srirang Manohar , - - PowerPoint PPT Presentation

regularization of inverse problems
SMART_READER_LITE
LIVE PREVIEW

Regularization of Inverse Problems Yoeri Boink *, Srirang Manohar , - - PowerPoint PPT Presentation

Deep Inversion, Autoencoders for Learned Regularization of Inverse Problems Yoeri Boink *, Srirang Manohar , Leonie Zeune *, Leon Terstappen , Stephan van Gils*, Christoph Brune* Biomedical Photonic Imaging Medical Cell Biophysics


slide-1
SLIDE 1

/ 43

Deep Inversion, Autoencoders for Learned Regularization of Inverse Problems

Yoeri Boink*†, Srirang Manohar†, Leonie Zeune*‡, Leon Terstappen‡, Stephan van Gils*, Christoph Brune* Biomedical Photonic Imaging† Medical Cell Biophysics‡ Applied Mathematics*

08-06-2018 Regularisation methods for PAT 1

slide-2
SLIDE 2

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

OUTLINE

2

  • 1. Robustness of learned primal-dual (L-PD) reconstruction in photoacoustic tomography (PAT)
  • 2. Functional learning with an unrolled gradient descent (GD) scheme

guaranteed convergence and stability!

  • 3. Learn latent representation of data space and image space via variational autoencoder (VAE)
slide-3
SLIDE 3

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

Classical Inverse Problems Parameter Estimation

VARIATIONAL METHODS AND DEEP LEARNING

3

Partial Differential Equations Deep Variational Networks Measurements Hidden Parameters Complex Data Science

Generalization? Regularization Theory?

slide-4
SLIDE 4

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

VARIATIONAL METHODS AND DEEP LEARNING

4

Chen, Pock - Trainable Nonlinear Reaction Diffusion, 2016

Norms nonconvex Differential

  • perators

Scale-space, Harmonic analysis Regularization theory Inverse problems Vector fields, multimodality Time dependent modeling Activation functions ReLU, sigmoid Convolutions functions per layer Scattering networks Generalization properties GANs VAEs? Multiple populations? Residual? Skip connections?

Deep networks Variational methods Deep Residual Neural Networks are connected to Partial Differential Equations

Chen et al - Neural Ordinary Differential Equations, 2018 Ciccone et al - Stable Deep Networks from Non-Autonomous DEs, 2018 Mallat - Understanding deep convolutional networks, 2016 Haber, Ruthotto - Stable architectures for deep neural networks, 2018

slide-5
SLIDE 5

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

bi-level, non-convex large-scale, real-time training data, generalization accurate, stable, reproducible uncertainty, parameters

Optimization Regularization Architecture

MATHEMATICS OF DEEP LEARNING

5

DL

Vidal et al. Mathematics of Deep Learning, 2018

slide-6
SLIDE 6

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

DEEPER IN INSIGHTS IN INTO DEEP IN INVERSIO ION

6

slide-7
SLIDE 7

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

4TU PROPOSAL PRECISION MEDICINE 2018

Challenges: C1: sparse/big data; C2: multi-dimensionality, heterogeneity C3: non-linearity; C4: super-resolution Deep learning for model-driven imaging

“classical” physics-based reconstruction enriched by deep learning → latent operator parameters, robustness “black-box” deep learning enriched by physical constraints → more natural, latent data structures, robustness

natural robust performance simple

7

slide-8
SLIDE 8

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

FIN INDING NEEDLES IN IN A HAYSTACK

8

slide-9
SLIDE 9

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

FIN INDING NEEDLES IN IN A HAYSTACK

9

slide-10
SLIDE 10

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

Denoising and Scale Analysis by Local Diffusion

  • given noisy input
  • denoise image by minimizing the Total Variation (TV) flow

→Time-discrete case: solving in every step the ROF [Rudin,92] problem:

10

slide-11
SLIDE 11

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

Denoising and Scale Analysis by Local Diffusion

Idea: Solution of nonlinear eigenvalue problem transformed to a peak in the spectral domain → time-point t where the eigenfunctions are completely removed depends on size and height of the disc, thus scale indicator

[Gilboa 2013,2014],[Horesh, Gilboa 2015],[Burger et al., 2015,2016] [Aujol, Gilboa, Papadakis, 2015, 2017]

11

slide-12
SLIDE 12

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

Contrast Invariance of L1-TV Denoising

12

slide-13
SLIDE 13

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

Zeune et al - Multiscale Segmentation via Bregman Distances and Nonlinear Spectral Analysis, 2017

13

slide-14
SLIDE 14

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

CNN CLASSIFICATION FOR CANCER-ID ID

14

Convolution network with max-pooling layers Fully- connected layers Input: image with three fluorescent channels

32 64 128 CTC probability No CTC probability

0.9722 0.0278 0.010 0.990

slide-15
SLIDE 15

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion
  • assume we have a trained AE with tied weights, i.e.
  • x and f(x) live in the same vector space, i.e.

is a vector field pointing from x to reconstructed f(x)

  • under some (fulfilled) assumptions, G(x) is the gradient field of an energy E(x)

with

AUTOENCODERS AND GRADIENT FLOWS

15

slide-16
SLIDE 16

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

Convolutional AE

Trained on no noise data, 6000 training sets, 1000 test sets 3 Convolution (32 filters in total) + Pooling blocks for Encoder + Decoder → 4963 parameters GT Std.dev 0.5 Std.dev 0.2 Std.dev 0.1 Std.dev 0.05 No noise

16

slide-17
SLIDE 17

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

Convolutional AE

Trained on noisy data (std dev 0.2), 6000 training sets, 1000 test sets 3 Convolution (32 filters in total) + Pooling blocks for Encoder + Decoder → 4963 parameters GT Std.dev 0.5 Std.dev 0.2 Std.dev 0.1 Std.dev 0.05 No noise

17

slide-18
SLIDE 18

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

Convolutional AE

Trained on no noise data, 6000 training sets, 1000 test sets 3 Convolution (32 filters in total) + Pooling blocks for Encoder + Decoder → 4963 parameters GT No noise Different Radius Different Radius, Trained on correct

18

slide-19
SLIDE 19

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

DEEP LEARNING OF CIR IRCULATING TUMOR CELLS

Leonie Zeune

19

Zeune et al - Deep learning for tumor cell classification (2019)

slide-20
SLIDE 20

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

TUMOUR BREAST ULTRASOUND DETECTORS

PHOTOACOUSTICBREAST IM IMAGING

(ILLUSTRATIONS MADE BY SJOUKJE SCHOUSTRA –BMPI GROUP , UNITWENTE)

20

LASER

Folkman, 1996

angiogenesis

slide-21
SLIDE 21

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

TUMOUR BREAST PULSE

  • LIGHT ABSORPTION
  • TEMPERATURE RISE
  • EXPANSION
  • PRESSURE RISE
  • ULTRASOUND WAVE
  • SIGNALS
  • RECONSTRUCTION

ULTRASOUND DETECTORS PHOTOACOUSTIC EFFECT

21

LASER

Folkman, 1996

angiogenesis

PHOTOACOUSTICBREAST IM IMAGING

(ILLUSTRATIONS MADE BY SJOUKJE SCHOUSTRA –BMPI GROUP , UNITWENTE)

slide-22
SLIDE 22

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

We make use of a projection model with calibration2:

2 Wang, Xing, Zeng, Chen -Photoacoustic imaging with deconvolution (2004) 1 Van Es, Vlieg, Biswas, Hondebrink, Van Hespen, Moens, Steenbergen,

Manohar - Coregistered photoacoustic and ultrasound tomography of healthy and inflamed human interphalangeal joints (2015)

PHOTOACOUSTIC TOMOGRAPHY

22

Figure: 2D slice-based imaging with rotating fibres and sensor array.1

PAT-operator

Folkman, 1996

angiogenesis

slide-23
SLIDE 23

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

INVERSE RECONSTRUCTION METHODS

23

FBP

  • perator

data reconstruction

▪ Direct reconstruction: filtered backprojection (FBP) ▪ Iterative reconstruction: total variation (TV)

(solved with PDHG)

▪ Learned post-processing: U-Net 3

PDHG

  • perator

data reconstruction reconstruction

FBP

  • perator

data

U-Net

3 Jin, McCann, Froustey, Unser - Deep Convolutional Neural

Network for Inverse Problems in Imaging (2017)

slide-24
SLIDE 24

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

MODEL-BASED REGULARISED RECONSTRUCTION -CHOOSING FUNCTION SPACES

Kingsbury - The dual-tree complex wavelet transform a new efficient tool for image restauration and enhancement (1998) Rudin, Osher, Fatemi - Nonlinear total variation based noise removal algorithms (1992) Bredies, Kunisch, Pock - Total Generalised Variation (2010) Boink, Lagerwerf, Steenbergen, van Gils, Manohar, Brune - A framework for directional and higher-order reconstruction in photoacoustic tomography (2018)

?

24

slide-25
SLIDE 25

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

FROM MODEL-DRIVEN TO DATA-DRIVEN

Hauptmann, Lucka, Betcke, Huynh, Cox, Beard, Ourselin, Arridge - Model based learning for accelerated, limited-view 3D photoacoustic tomography (2017)

▪ No regularisation parameter; ▪ Better robustness to noise; ▪ Faster reconstruction. However, ▪ No proven stability or convergence.

Meinhardt, Möller, Hazirbas, Cremers - Learning Proximal Operators: Using Denoising Networks for Regularizing Inverse Imaging Problems (2017) Adler, Öktem – Learned Primal-Dual Reconstruction (2017)

25

slide-26
SLIDE 26

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

= = +

CNN

+ +

26

slide-27
SLIDE 27

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

MODEL CHOICES PAT SIMULATION AND RECONSTRUCTION NETWORK

Training Resolution 1.5625 mm Number of pixels 192x192 Number of sensors 32

  • 768 training images
  • 192 test images
  • scaled between 0 and 1

DRIVE dataset: Staal, Abramoff, Niemeijer, Viergever, Ginneken - Ridge based vessel segmentation in color images of the retina (2004)

‘large’ network small network # primal-dual iterations (N) 10 5 # primal/dual channels (k) 5 2 # hidden layers 2 2 # channels in hidden layers 32 32 activation functions ReLU ReLU filter size convolutions 3x3 3x3

27

slide-28
SLIDE 28

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

ROBUSTNESS TO IMAGE CHANGES

28

slide-29
SLIDE 29

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

ROBUSTNESS TO IMAGE CHANGES

▪ Strong noise removal and background identification; ▪ L-PD is robust against many changes in image; ▪ Training with more variety in data has positive effect on quality.

29

slide-30
SLIDE 30

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

Initial pressure Segmentation

INCLUDING HIGHER-LEVEL TASK: SEGMENTATION

30

slide-31
SLIDE 31

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

INCLUDING HIGHER-LEVEL TASK: SEGMENTATION

reconstruction segmentation ▪ Learned post-processing: U-Net ▪ joint reconstruction and segmentation with L-PD

segmentation reconstruction

FBP

  • perator

data

U-Net

reconstruction

L-PD

  • perator

data segmentation

31

slide-32
SLIDE 32

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

INCLUDING HIGHER-LEVEL TASK: SEGMENTATION

Figure: Reconstructions and segmentations using a 32 detector setting.

32

slide-33
SLIDE 33

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

EXPERIMENTAL RESULTS: ONE-SIDED SAMPLING

33

slide-34
SLIDE 34

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

34

EXPERIMENTAL RESULTS: UNIFORM SAMPLING

slide-35
SLIDE 35

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

ROBUSTNESS TO SYSTEM CHANGES

Figure: Reconstruction from data of 64 detectors.

How to achieve generalisability towards imaging operator uncertainty? Latent structures?!

35

slide-36
SLIDE 36

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

CONVERGENCE OF PARTIALLY LEARNED METHODS

5 Banert, Ringh, Adler, Karlsson, Öktem – Data-driven

nonsmooth optimization (2018)

Several recent papers give convergence proofs for: ▪ methods that use explicit (Tikhonov-like) regularisation in the form of a neural network.4 ▪ methods with a proximal structure.5 ▪ methods where the learned part only has influence on the null-space.6

6 Schwab, Antholzer, Haltmeier – Deep Null Space Learning for

Inverse Problems: Convergence Analysis and Rates (2018)

4 Li, Schwab, Antholzer, Haltmeier - NETT Solving Inverse

Problems with Deep Neural Networks (2018) Lunz, Öktem, Schönlieb – Adversarial Regularizers in Inverse Problems (2018) 36

slide-37
SLIDE 37

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

LEARNED (UNROLLED) GRADIENT DESCENT

Goal: learn a nonlinear function (functional) such that its minimiser is our desired reconstruction where , and Hi represents a convolutional layer.

37

slide-38
SLIDE 38

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

▪ ▪ ▪

▪ Train network with , f or .

LEARNED (UNROLLED) GRADIENT DESCENT

Learning of well-known gradient flows possible? Diffusion, convection-diffusion, thin-film, etc.

38

slide-39
SLIDE 39

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

LEARNED (UNROLLED) GRADIENT DESCENT

NN

NN NN NN

39

slide-40
SLIDE 40

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

TRAINING DETAILS: COIL DATASET

Columbia Object Image Library

Nene, Nayar, Murase – Columbia Object Image Library (COIL-100) (1996)

Ground truth Noisy input

40

slide-41
SLIDE 41

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

▪ #GD-steps = N = 20 ▪ #layers = 3 ▪ #channels = 8 ▪ batch size = 9 ▪ #epochs = 50 ▪ Adam optimiser rate 2∙10-2 →10-3 ▪ kernel size = 3x3

TRAINING PARAMETERS AND RESULT

NN NN NN

Ground truth Noisy input L-GD output

41

slide-42
SLIDE 42

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

LEARNED (UNROLLED) GRADIENT DESCENT

iterations 1-5 iterations 6-10 iterations 11-15 iterations 16-20

42

slide-43
SLIDE 43

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

LEARNED (UNROLLED) GRADIENT DESCENT: Q2

iterations 1-5 iterations 6-10 iterations 11-15 iterations 16-20

LEARNED “DATA FIDELITY” OPERATOR

43

slide-44
SLIDE 44

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

LEARNED (UNROLLED) GRADIENT DESCENT: Q1

iterations 1-5 iterations 6-10 iterations 11-15 iterations 16-20

LEARNED “REGULARIZATION” OPERATOR

44

slide-45
SLIDE 45

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

LEARNED (UNROLLED) GRADIENT FLOW

Ground truth Noisy input L-GD output (20 iterations) L-GD output (40 iterations)

Learning Loss Functional

45

slide-46
SLIDE 46

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

LEARNING GRADIENT FLOWS, VARIATIONAL NETWORKS (PHYSICS-INFORMED)

46 Chen, Pock – Trainable Nonlinear Reaction Diffusion (2015) Kobler et al – Variational Networks: Connecting Variational Methods and Deep Learning (2017) Lusch et al - Deep learning for universal linear embeddings of nonlinear dynamics (2017) Yang et al - Physics-informed deep generative models (2018)

slide-47
SLIDE 47

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

DEEP DISCRIMINATIVE VS GENERATIVE MODELS

Genevay, Peyre, Cuturi – GAN and VAE from an Optimal Transport Point of View (2017) Mescheder, Nowozin, Geiger – Adversarial Variational Bayes: Unifying VAEs and GANs (2017) 47

slide-48
SLIDE 48

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

(BLIND) DECONVOLUTION PROBLEM IN PHOTOACOUSTICS

LEARNING UNCERT AINTY -CALIBRA TION IN FORWARD MODEL

48

slide-49
SLIDE 49

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

LEARNEDDATA-SPACE WITH VARIATIONAL AUTOENCODER

Dai, Wipf - Diagnosing and Enhancing VAE Models (2019)

49

Kingma, Welling – Auto-Encoding Variational Bayes (2013)

slide-50
SLIDE 50

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

LEARNEDDATA-SPACE WITH VARIATIONAL AUTOENCODER

r

50

slide-51
SLIDE 51

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

LEARNED DATA-SPACE WITH VARIATIONAL AUTOENCODER

ground truth reconstruction

51

slide-52
SLIDE 52

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

LIMITATION: GEOMETRIC VAE LATENT SPACE REGULARIZATION

Regularization of latent variables z

52

slide-53
SLIDE 53

Imaging and ML, IHP Paris, Apr 3, 2019

  • C. Brune - Deep Inversion

Thanks for your attention

CONCLUSIONS

  • Nonlinear eigenvalue problems offer spectral decompositions similar to Autoencoders
  • Robustness. Deep learning improves PAT reconstruction quality particularly in uncertain cases.
  • Functional learning. Mathematical theory on stability and convergence not available. Construct

learned methods that learn the functional to be minimised via a gradient flow.

  • Generative networks. Generative networks can be successfully used for learning latent

variables in inverse problems. VAEs show geometric limitations for latent space interpolation.

Boink, van Gils, Manohar, Brune - Sensitivity of a partially learned model‐based reconstruction algorithm (2018) Boink, Manohar, Brune - A partially learned algorithm for joint photoacoustic reconstruction and segmentation (2019) 53

Zeune et al - Multiscale Segmentation via Bregman Distances and Nonlinear Spectral Analysis (2017) Zeune et al - Deep learning for tumor cell classification (2019)