Deep Learning Theory with Application to Cancer Research
Leonie Zeune, Stephan van Gils, Guus van Dalum, Leon Terstappen, Christoph Brune
Inverse Problems and Machine Learning, Pasadena, USA, Feb 9-11, 2018
Deep Learning Theory with Application to Cancer Research Leonie - - PowerPoint PPT Presentation
Deep Learning Theory with Application to Cancer Research Leonie Zeune, Stephan van Gils, Guus van Dalum, Leon Terstappen, Christoph Brune Inverse Problems and Machine Learning, Pasadena, USA, Feb 9-11, 2018 Deep Learning as a Black Box
Leonie Zeune, Stephan van Gils, Guus van Dalum, Leon Terstappen, Christoph Brune
Inverse Problems and Machine Learning, Pasadena, USA, Feb 9-11, 2018
April 2017: ”No one really knows how the most advanced algorithms do what they do. That could be a problem.” June 2017: ”Artificial intelligence is a black box that thinks in ways we don’t understand. That’s thrilling and scary.” Breakthrough Technologies 2013 Deep Learning 2017 Reinforcement Learning Google Trend Benchmark 2010-2017 2010 2012 2014 2016
Calculus of Variations
Partial Differential Equations Inverse Problems Optimization Uncertainty Quantification Regularization MODEL-BASED
Theory
DATA-BASED
Application
Graphs / Networks Deep Learning Big Data Biomedical Imaging Segmentation Clustering Classification Life Sciences
◮ Learning variational networks ◮ Learning unrolled, proximal schemes (LISTA, learned PD)
θ
θ
θ (f)) − f] ◮ related:
◮ Network architecture: Which activation functions (nonlinearities, norms)
◮ Network as a generalized ODE: How can we add robustness to the learning of
◮ Nonconvex network learning: What is the optimal selection and amount of
◮ cells dissociate from primary tumor
and invade blood circulation
◮ rare cell events, challenging to detect ◮ circulating tumor cell (CTC) count has
prognostic value for survival outcome
◮ no overall CTC definition exists yet
7
8
9
10
11
1 2||u − un||2 2 + αTV(u) → minu
12
[Gilboa, 2013] [Gilboa, 2014] [Horesh, Gilboa 15] [Burger et al., 2015],[Gilboa et al., 2015],[Burger et al., 2016]
Example taken from [Burger et al., 2015].
13
JCV(c1, c2, C) =
(f(x) − c1)2dx +
(f(x) − c2)2dx + α · Length(C) → min
C,c1,c2
14 [Osher, Sethian, 88], [Mumford, Shah 89], [Chan, Vese 01], [Ambrosio, Tortorelli 90] → related to Level-set method (Hamilton-Jacobi)
ϕ∈C∞ (Ω;R2) ||ϕ||∞<1
JCV2(c1, c2, u) =
u
dx + α TV(u) → min
u∈BV(Ω),c1,c2 u(x)∈{0,1}
For fixed c1, c2 corresponds to ROF with binary constraint ([Burger et al. 12]): min u∈BV(Ω)
u(x)∈{0,1} 1 2 ||u(x) − r(x)||2 2 + αTV(u) with r(x) = (f(x) − c2)2−(f(x) − c1)2− 1 2.
15
1 2||u − f||2 2 + αTV(u)
◮ Inverse scale space for nonlocal filtering through Bregman iterations
u∈BV(Ω)
2 + α(TV(u)− < u, pk >)
◮ How can we construct an ”inverse scale space” for segmentation?
16 [Osher et al. 05]
S(t) =
H
17
18
18
18
18
18
18
18
18
18
18
18
18
18
18
18
18
18
18
18
18
18
18
18
18
18
18
18
18
18
18
18
19
19
19
19
19
19
σ = 0.25, c1 = 1.01 σ = 0.5, c1 = 1.09 σ = 0.75, c1 = 1.25 σ = 1, c1 = 1.44
20
ϕ∈C1
C(Ω;RD) ϕ(x)·n<γ(n)∀n∈Rd −
x∈Rd
21
22
◮ successive peaks (usually) correspond to comparable scales. ◮ reconstruction of different segmentation scales based on S ◮ non-eigenfunctions reshaped over time
[Florack, Kuijper, Topological structure of scale-space images, 2000] 23
activation function (differentiable) bias convolution kernel input “image” pixel pos.
N 2
i=1
2
Trained on noise-free data, 6000 training sets, 1000 test sets 3 Convolution (32 filters in total) + Pooling blocks for Encoder + Decoder à 4963 parameters GT Std.dev 0.5 Std.dev 0.2 Std.dev 0.1 Std.dev 0.05 No noise
Trained on noisy data (std dev 0.2), 6000 training sets, 1000 test sets 3 Convolution (32 filters in total) + Pooling blocks for Encoder + Decoder à 4963 parameters GT Std.dev 0.5 Std.dev 0.2 Std.dev 0.1 Std.dev 0.05 No noise
Trained on noise-free data, 6000 training sets, 1000 test sets 3 Convolution (32 filters in total) + Pooling blocks for Encoder + Decoder à 4963 parameters GT No noise Different radius Different radius, trained on correct
◮ Spectral decomposition for nonlinear denoising & segmentation ◮ Scale/Stability: Multiscale properties of CNNs ◮ Eigenfunctions/Invariants: Inverse NF equations, kernel properties of CNN ◮ ACCEPT open source software tool for CANCER research ◮ Deep Learning for Cancer ID is possible, however multimodal and multiclass
◮ Outlook: Combination of CNN autoencoder, classifier and GAN
◮ Zeune, et al., Multiscale Segmentation via Bregman Distances and Nonlinear
◮ Zeune, et al., Combining Contrast Invariant L1 Data Fidelities with Nonlinear
◮ Zeune, et al., Quantifying HER-2 expression on Circulating Tumor Cells by
◮ Eissa, et al., Cross-scale effects of neural interactions during human neocortical
◮ Boink, et al., A framework for directional and higher-order reconstruction in
37
[Nikolova 02, Chan 05, Darbon 05, Duval 09, Zeune 17] 38