Splines and imaging: From compressed sensing to deep neural nets - - PDF document

splines and imaging from compressed sensing to deep
SMART_READER_LITE
LIVE PREVIEW

Splines and imaging: From compressed sensing to deep neural nets - - PDF document

Splines and imaging: From compressed sensing to deep neural nets Michael Unser Biomedical Imaging Group EPFL, Lausanne, Switzerland Plenary talk: Int. Conf. Signal Processing and Communications (SPCOM20), IISc Bangalore, July 20-23, 2020


slide-1
SLIDE 1

Splines and imaging: From compressed sensing to deep neural nets

Michael Unser Biomedical Imaging Group EPFL, Lausanne, Switzerland

Plenary talk: Int. Conf. Signal Processing and Communications (SPCOM’20), IISc Bangalore, July 20-23, 2020

Variational formulation of inverse problems

2

noise

n

Linear forward model

s

Integral operator

H y = Hs + n

Problem: recover s from noisy measurements y

srec = arg min

s∈RN ky Hsk2 2

| {z }

data consistency

+ λkLskp

p

| {z }

regularization

, p = 1, 2

Reconstruction as an optimization problem

slide-2
SLIDE 2

Formal linear solution:

s = (HT H + λLT L)−1HT y = Rλ · y

Linear inverse problems (20th century theory)

3

Equivalent variational problem

s? = arg min ky Hsk2

2

| {z }

data consistency

+ λkLsk2

2

| {z }

regularization

Interpretation: “filtered” backprojection

R(s) = kLsk2

2: regularization (or smoothness) functional

L: regularization operator (i.e., Gradient)

Formal linear solution:

s = (HT H + λLT L)−1HT y = Rλ · y

Andrey N. Tikhonov (1906-1993)

min

s

R(s)

subject to

ky Hsk2

2  σ2

Dealing with ill-posed problems: Tikhonov regularization

Learning as a (linear) inverse problem

4

but an infinite-dimensional one …

minf∈H R(f)

subject to

M

X

m=1

|ym − f(xm)|2 ≤ σ2

Introduce smoothness or regularization constraint

R(f) = kfk2

H = kLfk2 L2 =

Z

RN |Lf(x)|2dx: regularization functional

(Poggio-Girosi 1990) (Wahba 1990; Schölkopf 2001)

kernel estimator Regularized least-squares fit (theory of RKHS)

fRKHS = arg min

f∈H

M X

m=1

|ym f(xm)|2 + λkfk2

H

!

<latexit sha1_base64="T+C9yVxj8piky8k/0T8uhwPugc=">AL2XiclVZNb9tGEKXSr0RuG6e9FOhlUFuAk8aCqKBNWtRACeBjSapaydOYK+jLsmltPHyI7tLW8qah96KXvuv+g/6S3rtLCnJOUDgEJxMzb94Yzu7PjpYIr3ev907rywYcfzJ1WvtpU8/+/z68o0v9lWSZ89xORyJceVUzwmD3XAv2MpWMRp5gL7zjTet/cKk4kn8TE9SdhTRYcxD7lONpsHy30RlnmK+2WXDTFDJ37IABKNKr6s3GZVMQcg1EFjTI5bICSQh7P68tXczbxNPVgNB4bIqLDlGyTgKhV0ovREMCBUDoFEPB6YEAiPgeylPmzlRLBQrwEqRwMTbj5qydQ2M5gMohgHcI1sg8wHkQ3ieTDkT571YdvEYKfFVAgZyE5Q9WS7FW/xNxcHSyv9Lq94oHF3f6snL/K6d4dgY3rv1LgsTPIhZrX1ClDt1eqo8MlZr7guEXZoql1D+mQ2aKVOfQVMAYSLxF2sorDUcjZSaRB4iI6pHqumzxgt91iJVqOqnhfl7arlMNPhvSPD4zTLPbLgMJMgE7A1hcCLpmvxQSo7+N3ZVRjvDWGZ+6RscHb5TWH8qlgwUb3h/6RGTFxwjT6JYvZqZ9EY0DQ0IacTEJWEgzoXNDVDh7b3faHSj2mwKEgpcJwbT6Ee0EMbg9i1yZKBsmIsiNHq5cW/3une/u93LFzCSTaYXhcB5Q9h1WA8kbFSBiOpxymoh0dCs4ijqyS16ByIzAQzVpWN8/bsJW8wW/nLM0/DvSw3fv57cdt0XZLbE1jK90kJwv+Pu86AHaBsFrZ15AYMGSncO+yWe91v8cl9UzjXhcCVtlVGU3WT1f0rca8C6Rd+rcPdeZ513zePJuoTszoapMKZAf4lkg52K3zEW7xUWC4m/KaVNUZAjsAYAV92L6hsJlRaoqRSHmIs1i86Etp1UAY+srsWFcInybxWn40C54oU6c4jVQIb5ErBiqLScSFskoesFeynxOBdhumIiyEVRVeLybm+JWsK3P8xu8wN5fFBHDQRm7g7Zs7NpnO34lzgPqg4F2ifVpxPm84tLjwm9QyCPdNsNTHsDbYzWziNG1IyYcgJXiqS03iI94rZyG1Try54nKS5KS7Sx02uB3PXg6Zre+7abroCHupL1hQe4TA2ey+gmPFnaB9aux9IFZQXhLfJ06pBxAzJe2KNhFxAnDYQJInYkC7E7B03cMeLza+BEAskY8yJbizDCGUzrwHenFxP8jK3QXk9kf0VtyZgysMzezpwgc78SjeNJgZ9leAruQHFipqAnJWV47o1+wn0yLr2/pOguCGRUIEozY3MjK8pmja9fz42v58btOC+mW7lyLhzx45MvNKufGl2ZsyPElxOTnARlSbAjVf8nCcliuSdMZoHmIHiBPBI65xcGl31osH9pjGljwE7OWQ4mQBCsdKYNrv4tMmiuGYGg/1CLs9G2uUw5nIuK6Ps07VaftQbEuBQ0Dhq98Us7FXaYnUlqHbxQSJ9S8cn3YhopEOeCo6DYHw8WX/X7XvdPt/9pfuX+vnBmdq87XzjfOmuM6d537zpaz4zx3/Favtd8atH5bOlz6femPpT9L6JXWdM2XTu1Z+us/+a1U5w=</latexit>

Given the data points (xm, ym) ∈ RN+1, find f : RN → R s.t.

f(xm) ≈ ym for m = 1, . . . , M

slide-3
SLIDE 3

OUTLINE

■ Introduction ✔

■ Image reconstruction as an inverse problem ■ Learning as an inverse problem

■ Continuous-domain theory of sparsity

■ Splines and operators ■ gTV regularization: representer theorem for CS

■ From compressed sensing to deep neural networks

■ Unrolling forward/backward iterations: FBPConv

■ Deep neural networks vs. deep splines

■ Continuous piecewise linear (CPWL) functions / splines ■ New representer theorem for deep neural networks

5

2

Part I: Continuous-domain theory of sparsity

6

(Fisher-Jerome 1975) L1 splines (U.-Fageot-Ward, SIAM Review 2017) gTV optimality of splines for inverse problems

slide-4
SLIDE 4

Splines are analog, but intrinsically sparse

7

Spline theory: (Schultz-Varga, 1967)

: spline’s innovation L = d dx ak xk xk+1

Definition The function s : Rd → R (possibly of slow growth) is a nonuniform L-spline with knots {xk}k∈S

⇔ Ls = X

k∈S

akδ(· − xk) = w L{·}: differential operator (translation-invariant) δ: Dirac distribution

Spline synthesis: example

8

L = D = d dx

x x1

wδ(x) = X

k

akδ(x − xk)

a1 x

s(x) = b1p1(x) + X

k

ak

+(x − xk)

b1

Null space:

ND = span{p1}, p1(x) = 1 ρD(x) = D−1{δ}(x) =

+(x): Heaviside function

slide-5
SLIDE 5

Spline synthesis: generalization

9

Requires specification of boundary conditions

Finite-dimensional null space: NL = span{pn}N0

n=1 xk

Spline’s innovation:

wδ(x) = X

k

akδ(x − xk)

<latexit sha1_base64="FLgF8iONCx2uG1p2Ptp/DtdeHMQ=">ALgHiclVZtT9xGEDa0zctd25D2Y7+sCqeSKFzPRC0tKhISCQIpTSkJAKj09qeu1tu/cLuGu5Y+XdW/SX9VKmztu/wC1TE0p2seZ59ZjwzO7tuzJlUvd7fC4uf7Fg4ePHrfaX3719ZOlp98cygRHrz3Ih6Jjy6VwFkI7xVTHD7GAmjgcvjgjncM/uEShGR+E5NYzgL6DBkA+ZRhab+0sVRbJb+IAkLw+gys246Fwn1yYrjMxlzOpVqyoFc9R0fuKrzjGZPNtyZBL0x0Q7WQxagJ/S/jglJdJaGTQGxJ+t9JeWe91e9pDmi128LFvFc9B/+vgfx4+8JIBQeZxKeWr3YnWmqVDM45C2nERCTL0xHYLOHKakgyafDCKBv1CRzFrh0UDKaeAiM6BqJOuYMd6KGYuQA1n16rpB2ipbThM1+OVMszBOFIReHtAg4URFxJSB+EyAp/iUM/D70qowngrCu/sM2CN8srgPQoB3+r+v6mR4BvwSFuIAQrwoCGjoa2dA8anPgxowlWqHTmYvbc6rQ7J2kISpBI34RyU3EQ7Vm6ArZCXLEiGEfdTLYZuqu0Xve7GTy96aYMjYFpwel0k5D+klYNxeQK5G4ykGienLnaugoCls14xbOwikXDQxitM0tbsJa0pG/f3Vy7Cva82fv4naZt03VPb5VjKT0kJ0v9Pu6qgRiATV2J74V5ONW7RkcTeged6rdf9GZdUM429zjlZaeVRSdz6AlZulqwbH+QuJ3f62bjxM8+7YuH0bkcvZ47KbnIH6SnuBefG2XN9W7fYKJD9FZomRVmGiNkAZNm+Xb7m4b5Oyl6yQsyd1IvNhqacxgPRpr4CB8Y9wjdZLMInrUyXVIVjKsrC94gVQzXlRMEsGdksOIrBY5QTMw0jng+CshcWHqbmnzhm9LmuPqx/IAtPqoyTOmMHu2MG7tTBwxLY0D4pgQ3ZtyXwbR3cY9wFoWYUnJl6r86BCxnpnAKG1IA184lHiqC0XCI54reSs1QLy94E8UploE5E1d69UcelWH9ufQfh3yWagKzG9428VDnoEof8Juowvca7MeS+fnFSTXqNOpUiY1yqTRo3WGaDCuagwnCmBIGzG74xpv3Bx+NQZviEwJ6q2DCMU9bz6eHIyNU3z3Pr58eQcL9sVBzrfPLOnQy5xg+P5l7OPYi/jzJckDsk2zEF6fdcnCDBr9hn0xy9I8Y4UxABkjNwPMjZCVzRjPz+fG87lxP/SZp4tWDrQ9Bw5E5OZ26Ql9MFPejXC5c8V8GFGlM9Z8yetJnK+I4pmifo0TIw4C5jCi0urs5Y95AgUjuQhwVlOYrxZEMmugYDyuvi0HAl4mwyHaoTHiYK3eGdSNu2h3edMmjmUGhKgZeADKueFLPbqVQCpY1Cd71WSEXdtHR8mIGKQinBq6Jdvxg2X47Xu/bLbu/P9eXt7eLS+Mj6zvreWrVsa8PatvasA+u95Vl/Wf8uPFh42F5sr7Z/bNs5dXGhWPOtVXnam/8BFGE2Ig=</latexit>

s(x) = X

k

akρL(x − xk) +

N0

X

n=1

bnpn(x)

<latexit sha1_base64="Lhlkq8HrqncLMb/40xnAnOuT6Ag=">ALe3iclVZbT9xGFDb0liy9kPaxL6NuVkorGzSlN6QkEgQSCmlEgE3q7G9uzusONLZsawy2j+ZN/6S9qXSj1jexdfoNoYsbLO9813js85c2a8hFEhbfuvpeUPvzo408ePGytfPrZ51+sPvryTMQp98mpH7OYv/WwIxG5FRSycjbhBMceoy8ca7Bn9zRbigcfRaThPSC/EwogPqYwm/ioT9wzNHm6jVyRhn01ki5maziJNC4P9YuH8V91WcZMyNMmwMwHiKvkXuz9lfJhJtO/oPdi3NfL6EUr6Ue6kv9q2u3b2oOaLU7y0reI56j96+LcbxH4akj6DAtx4diJ7CnMJfUZ0S03FSTB/hgPicrC0qgDpgANYg7/kUSZtcLDoRDT0ANmiOVI1DFjvBMzFi4GourV80LdKlsuUjn4oadolKSRH4e0CBlSMbIVAFlBNfsinCvg/flWIJ8VYUXjs9ZYI3yuA8DEjwXb3x82eGhF2RSTgnETk2o/DEeBcgc4pGwakAFOmdTKFYPZe6vT6qCsIwQCKvJSxogUP4HdBQ40UF7YMB3GLNCKDz2tnHW7u/V83dYNDifTgmN3gZD/A60cjMdSkruBSKpxMuxB0oSUj3rKMPWyOUpI8p4JRPdmr3omrJxv7hyEe6i2vD576Vt0rWgtseglO+TEqD/n3ZVQY6ISD0B7QWbWyvYyiMBvUPW1Ibd/R6WVDMNvc4YetzKoxLvUszJ49slm8YHus/JvX62bv3M8y5pNL3f0bOZo7Kb3IG+gL3g3jpbU3d1iwMC2U+haVKUZQiZDYDazt3yNQ+LOil7yQoxd1IvNh2achoPSJn6chgYC4RvsliEj1qZLqoKJ5iXhReIFUI15QTBLBnZLDhJiE8xQ2YaxiwfBGUvNDrW5he5ZvR5njqufyCNzquM8zpjF7pjBu7WweMS2NA+L4EN2cMSeFgH9ynzCJczCsxMtV/nkHcwzkzhJDQkJ0y5V3CocIqjIZwraluboV5eAOefhlTzEL2qa72YQy/q0MEcOqhDAY1kgQUNb3twvlPCy5+w1+gC78ash9IFeQXRDeh0qpRJjTJp9GidwRuM6xrDjUMyxI2YvXGN24OvxqDNUQmkBNZWwYR8npeAzg5qZzqPLdBfjy5Z2n4kDlm2f2dNAVbHA4/3L2SeJn/FmGC3IHZTumIP2aSw5u0fAX6JNJjv6WAJwJ8DBj5GZC5kaSlc0YLy/nxsu58SAKqK+KVg6VMweOeOzlduFzdTRT3othuXtNAzLCUmWs+ZKXkyRfESczRfUSJkAUMxpSCReXVmcje9AJkTCShwhmOUrgZoEvSGISL8LT8sVBC6S0VCOYNqTiQR3cCdSjuPDXacMmjkUmVLAJSDqifF7GIqJAdpo9DdrBVSYk+Xjg8zUEFI7gqOvWLYfPlbLPrPOvav3/X3tkpLo0PrK+tb6wnlmNtWTvWvnVknVq+9af1z9LS0nLr35X2ytrKek5dXirWfGVnpXn/wFo2zMd</latexit>

Green’s function of L: ρL(x) = L−1{δ}(x)

L: spline-admissible operator (LSI)

Proper continuous counterpart of

10

`1(Zd)

Space of real-valued bounded Radon measures on Rd

M(Rd) =

  • C0(Rd)

0 =

  • w 2 S0(Rd) : kwkM =

sup

ϕ2S(Rd):kϕk∞=1

hw, ϕi < 1 ,

where w : ϕ 7! hw, ϕi

M

= R

Rd ϕ(r)w(r)dr

S(Rd): Schwartz’s space of smooth and rapidly decaying test functions on Rd S0(Rd): Schwartz’s space of tempered distributions δ(· x0) 2 M(Rd) with kδ(· x0)kM = 1 for any x0 2 Rd

Basic inclusions

8f 2 L1(Rd) : kfkM = kfkL1(Rd) ) L1(Rd) ✓ M(Rd)

slide-6
SLIDE 6

Representer theorem for gTV regularization

11

Convex loss function: F : RM × RM → R (P1)

arg min

f∈ML(Rd)

M X

m=1

|ym hhm, fi|2 + λkLfkM ! ’

Representer theorem for gTV-regularization The extreme points of (P1 ) are non-uniform L-spline of the form

fspline(x) =

Kknots

X

k=1

akρL(x xk) +

N0

X

n=1

bnpn(x)

with ρL such that L{ρL} = δ, Kknots  M N0, and kLfsplinekM = kak`1.

L: spline-admissible operator with null space NL = span{pn}N0

n=1

gTV semi-norm: kL{s}kM = supkϕk∞1hL{s}, ϕi Measurement functionals hm : ML(Rd) ! R (weak⇤-continuous)

with ν(f) =

  • hh1, fi, . . . , hhM, fi
  • (P1’)

arg min

f∈ML(Rd)

  • F
  • y, ν(f)
  • + λkLfkM
  • V

ν : ML → RM (U.-Fageot-Ward, SIAM Review 2017) ML(Rd) =

  • f 2 S0(Rd) : kLfkM < 1

Example: 1D inverse problem with TV(2) regularization

12

x

no penalty

Generic form of the solution

L = D2 = d2 dx2 ρD2(x) = (x)+: ReLU ND2 = span{1, x}

sspline(x) = b1 + b2x +

K

X

k=1

ak(x − τk)+

with K < M and free parameters b1, b2 and (ak, τk)K

k=1 τk

Total 2nd-variation: TV(2)(s) = supkϕk∞1hD2s, ϕi = kD2skM

sspline = arg min

s∈MD2(R)

M X

m=1

|ym hhm, si|2 + λTV(2)(s) !

slide-7
SLIDE 7

Other spline-admissible operators

13

L = Dn

(pure derivatives)

polynomial splines of degree (n − 1)

L = Dn + an−1Dn−1 + · · · + a0I

(ordinary differential operator)

exponential splines Fractional Laplacian:

(∆)

γ 2

F

! kωkγ )

polyharmonic splines Fractional derivatives:

L = Dγ

F

← → (jω)γ ⇒

fractional splines

(Dahmen-Micchelli 1987) (Schoenberg 1946) (U.-Blu 2000) (Duchon 1977) (Ward-U. 2014)

Elliptical differential operators; e.g,

L = (−∆ + αI)γ ⇒

Sobolev splines

Recovery with sparsity constraints: discretization

14

Auxiliary innovation variable: u = Ls

Constrained optimization formulation

(Ramani-Fessler, IEEE TMI 2011)

LA(s, u, α) = 1 2 ky Hsk2

2 + λ

X

n

|[u]n| + αT (Ls u) + µ 2 kLs uk2

2

ssparse = arg min

s∈RN

✓1 2ky Hsk2

2 + λkuk1

subject to u = Ls

<latexit sha1_base64="6+x230BfxcZesH0JRr3xpFp+hw=">AL23iclVbNbhs3EJbSv0RqG6c9kLUFpCmsaBV2roNaiCAE8MGHNe1YzuwV1G5u5REm/sTkmtLZnjqrei1T9Un6Jv02CF3Je+u7EIhIJuY+eab2ZnhkF7CqJCdzj/1Ox98+NHn9y912h+tn95cefHEk4pT75NCPWcxfe1gQRiNyKlk5HXCQ49Ro698w2jP74gXNA4eiUnCemFeBjRAfWxBF/6e8VN6AiYXgi5IQRpFxvgITuK5eHSCSYC6LROnIxHyI3pFf5QiXRvtvdjVyGRnIhwq5A4595WjV1SB8Z1ETvWr/bxn8u373TRd9CwYQXYAtxsavOAm0+xKlBuMgl9PhSH7jhl48Vkik3hnxJZIx0qhoYIlTrdftZgc8rPSXljvtjl1ofuPkm+Vavb6D+796waxn4Ykj7DQpw6nUT2FOaS+ozohpsKkmD/HA+Jsq41aoEoQIOYwy+SyEpLOBwKMQk9QIZYjkRVZ4Q36oyEi4Eoe/W8UDeKktNUDn7sKRolqSRnwU0SJnJkCkwCiHhLEJwr4P35ViCfGWGF45PWCN+YlhfAxI8F6+6duT40IuyAS9JxE5NKPwxBHgXIHOKRsEpABTpmEIojBdN9oNVrINpxAEVeyhiR4inIXcBAf2bFC9NhzKB+fOhp5TzutNe+f9zRcxhOJjm0wZA9gNYMRiPpSRzA5GU42TYgzMhSUj1tGsMGlqTp4wo45WMdWO60RVm435x5jzcRbnh89+L26RrQW6PQSnfJyUA/z/uMoMcETiPAtoLZodWcB5HMCJ8kitdto/gEk509DrjKGVRhaVeJtiTlauTbrGB7rNya1+1q79zPIuaTS53dGTqaOim8yBPoWz4F47e6Ru6hYHCOyfnNOkyGYImQOAlp2b6SseFnVS9GILMXNSLTYdmnIaD0iZ+nIYGAuEb7KYh48alheViWHsF4kXiBVCNeUEQpsMOwsOEuJTzJCZhjHLBkHRC1wg2vyFewVGn+ep/eoH0uikjDipIjagO6bKjapyv6Cc4z4pKOdodwvK3apyizKPcDmFwMxUW1UMeQvjzBROQkNywpR7AZcKpzgawr2i1rUZ6kWDnTjRyl6O1Wu5zPV86pqe6barqoCGslcF8x524TnAyW8+Ambc13gXRl7KF2QVRBdAU+rDBlXIO5Hq0i+BzisoJw45AM8VzM3nkFdz4/CoINkcyhpzIihlEyKt5DeDmpHKis9wG2fXkHi07JQcqOzT1UIXcMDh/svQB4lv8dM5+AWsicmB73MKAfX2vBn6JNxpv0lAbUl4KFZGJCZkJiy2aEZ2cz4dlMuB0F1Fd5K4fKmSn2eOxlcuFztTdl3ozB3L2kARlhqSxqZvJinGQWcTJlVC9gAkQxoyGV8HBptFbtQgdEwkgeIpjlKIGXBRL0iAi/TashisIvFOjoRzBtCdjCe7gTaSc73x46xSVZg5FphTwCLC68k0xfcKyYHaMLS7lUJK7OnC9WEGKhDBF8Fb0am+DOc3R9286Td+bW7/Oxp/mq8W/uq9nXtYc2prdWe1bZqe7XDml936sf13+q42Wv+3vyj+WcGvVPb6slVbzr/8AhixYMg=</latexit>
slide-8
SLIDE 8

Linear step Proximal step

ADMM algorithm

sk+1 =

  • HT H + µLT L

−1 z0 + zk+1

with

zk+1 = LT µuk − αk

αk+1 = αk + µ

  • Lsk+1 − uk

= pointwise non-linearity

For k = 0, . . . , K

LA(s, u, α) = 1 2 ky Hsk2

2 + λ

X

n

|[u]n| + αT (Ls u) + µ 2 kLs uk2

2

Discretization: compatible with CS paradigm

15

ssparse = arg min

s∈RK

✓1 2ky Hsk2

2 + λkuk1

subject to u = Ls uk+1 = prox|·|

  • Lsk+1 + 1

µαk+1; λ µ

  • <latexit sha1_base64="t+H4B0KHM64O7imGe2vW8MC3Lw=">AJyHiclVZb9s2Fa6W+zd0u2lwF64xgG6XgzJ3ZuQ4AKYp0yEOXNm2RyAso6VgmTF1KUokdli973h/c39gv2KFkJ5aUDK4AWwTPx+87OufwkEHOmVSu+8/arY8+/uTz9Y73c+/+PKrzduf/NaZoUI4SjMeCbeBlQCZykcKaY4vM0F0CTg8CaY7Fn7mzMQkmXpKzXLYZjQOGUjFlKFU6cbf/e0H4xIYf7Ukwe2dF+yakDXoDRvkhILrKpOdXv/TDK1Htj/IDF98pFB6Z8yfnaB/5I0FB7OJsUCMt4JGcJvohPeT6mFeo3UsF8jk5GdAFm8Q89crqx6fbd8iHtgTcfbO7ecrnxent9Z4fZWGRQKpCTqU8dxcDTUVioUcTNcvJOQ0nNAYdPlhmzhVERGmcBfqkg5W8PRFq/EZlQNZNm528znZSqNGToWZpXihIw0poVHCiMmIjTyImIFR8RmgYor8FVehHjeGVN9TWKbu8ZpAh5RDt9H8ZDPUY+BkotAtI4TzMkoSmkfZHNGF8FsGIFlxhVOVoMe5udbdIWQmSIJQEBeg5K847yMGC6dKeVLEmDKjRwY7T10+9s/PXRNCyNgNse4fQRUP4QtO2Orp5KxNVTzk9MAi1VBwkyt1ogvCg7aqsLUdBcD02C28qsz91dlRs/4O4bhW5A4pvJDQoLw/+OuM6gxyCKQWF64qY0m2h9LrB24rx+5/Z9xST3SWMOck1638kq+K6iA3tWSgdUgN4ncqLN9pXMZd8XS2c1CjxdCyzKVgDnBveBfid3X1WLhwTl35zThqiMELEbgGx619M3FYVWVYpE3Ep0kw2i206rQLRNr8CW8cK7tsozt0n3ZKX1IlzKpaJV/AVXbXpRMIyGUveJlDyCgnVXeuGsGyCksPjf0nvm1yQaAPmx/I0uM64riJ2MPqWBj3msbDJWOL+3jJ2KLdZzwAoRYQbIt6v4mBd9ixbG4U1pwArv0zPA8Eo2mMR4LeQXx9wUGWG12edQdNrqeXpqdN0/NL0/OmKWKpmtuiltozPLoZiOVPeNZKdHBh1y8doRetGps2ENMW4ryB8LMEYtpyKJg0cJN282oguCXZelQ+5CUo3OYxwf5AcjytiGQXQECFfXy6vgS8lKSxGmMHgak6ZxGen9r7MUxMzWhrO43ARs4tbfXus7jkSCWQ2jL0Bw03FQ3MUkuymxSJMLR4r/Cat4j24PWg7z3uD/4YbO4+qS4YzrznXPXued4zraz6+w7L5wjJ3T+Xbuz9v3a3c7vnbxz3plV0Ftr8zXfOrWn89d/EsSR6A=</latexit>

Example: ISMRM reconstruction challenge

16

L2 regularization (Laplacian)

  • M. Guerquin-Kern, M. Häberlin, K.P. Pruessmann, M. Unser, IEEE Trans. Medical Imaging, 2011.

TV regularization

slide-9
SLIDE 9

OUTLINE

■ Introduction ✔ ■ Continuous-domain theory of sparsity ✔ ■ From compressed sensing to deep neural networks

■ Unrolling forward/backward iterations: FBPConv


■ Deep neural networks vs. deep splines

■ Continuous piecewise linear (CPWL) functions / splines ■ New representer theorem for deep neural networks

17

Linear step Proximal step

ADMM algorithm

sk+1 =

  • HT H + µLT L

−1 z0 + zk+1

with

zk+1 = LT µuk − αk

αk+1 = αk + µ

  • Lsk+1 − uk

= pointwise non-linearity

For k = 0, . . . , K

LA(s, u, α) = 1 2 ky Hsk2

2 + λ

X

n

|[u]n| + αT (Ls u) + µ 2 kLs uk2

2

Discretization: compatible with CS paradigm

18

ssparse = arg min

s∈RK

✓1 2ky Hsk2

2 + λkuk1

subject to u = Ls uk+1 = prox|·|

  • Lsk+1 + 1

µαk+1; λ µ

  • <latexit sha1_base64="t+H4B0KHM64O7imGe2vW8MC3Lw=">AJyHiclVZb9s2Fa6W+zd0u2lwF64xgG6XgzJ3ZuQ4AKYp0yEOXNm2RyAso6VgmTF1KUokdli973h/c39gv2KFkJ5aUDK4AWwTPx+87OufwkEHOmVSu+8/arY8+/uTz9Y73c+/+PKrzduf/NaZoUI4SjMeCbeBlQCZykcKaY4vM0F0CTg8CaY7Fn7mzMQkmXpKzXLYZjQOGUjFlKFU6cbf/e0H4xIYf7Ukwe2dF+yakDXoDRvkhILrKpOdXv/TDK1Htj/IDF98pFB6Z8yfnaB/5I0FB7OJsUCMt4JGcJvohPeT6mFeo3UsF8jk5GdAFm8Q89crqx6fbd8iHtgTcfbO7ecrnxent9Z4fZWGRQKpCTqU8dxcDTUVioUcTNcvJOQ0nNAYdPlhmzhVERGmcBfqkg5W8PRFq/EZlQNZNm528znZSqNGToWZpXihIw0poVHCiMmIjTyImIFR8RmgYor8FVehHjeGVN9TWKbu8ZpAh5RDt9H8ZDPUY+BkotAtI4TzMkoSmkfZHNGF8FsGIFlxhVOVoMe5udbdIWQmSIJQEBeg5K847yMGC6dKeVLEmDKjRwY7T10+9s/PXRNCyNgNse4fQRUP4QtO2Orp5KxNVTzk9MAi1VBwkyt1ogvCg7aqsLUdBcD02C28qsz91dlRs/4O4bhW5A4pvJDQoLw/+OuM6gxyCKQWF64qY0m2h9LrB24rx+5/Z9xST3SWMOck1638kq+K6iA3tWSgdUgN4ncqLN9pXMZd8XS2c1CjxdCyzKVgDnBveBfid3X1WLhwTl35zThqiMELEbgGx619M3FYVWVYpE3Ep0kw2i206rQLRNr8CW8cK7tsozt0n3ZKX1IlzKpaJV/AVXbXpRMIyGUveJlDyCgnVXeuGsGyCksPjf0nvm1yQaAPmx/I0uM64riJ2MPqWBj3msbDJWOL+3jJ2KLdZzwAoRYQbIt6v4mBd9ixbG4U1pwArv0zPA8Eo2mMR4LeQXx9wUGWG12edQdNrqeXpqdN0/NL0/OmKWKpmtuiltozPLoZiOVPeNZKdHBh1y8doRetGps2ENMW4ryB8LMEYtpyKJg0cJN282oguCXZelQ+5CUo3OYxwf5AcjytiGQXQECFfXy6vgS8lKSxGmMHgak6ZxGen9r7MUxMzWhrO43ARs4tbfXus7jkSCWQ2jL0Bw03FQ3MUkuymxSJMLR4r/Cat4j24PWg7z3uD/4YbO4+qS4YzrznXPXued4zraz6+w7L5wjJ3T+Xbuz9v3a3c7vnbxz3plV0Ftr8zXfOrWn89d/EsSR6A=</latexit>
slide-10
SLIDE 10

Identification of convolution operators

19

Normal matrix: A = HT H (symmetric)

  • deconvolution microscopy (Wiener filter)
  • parallel rays computer tomography (FBP)
  • MRI, including non-uniform sampling of k-space

Generic linear solver:

s = (A + λLT L)−1HT y = Rλ · y

<latexit sha1_base64="+pqfqhf0LVMNd1ASFUzd5fPpbPA=">AJ+3iclVZdb9s2FJXbfcTe1qbd416IJQG6LjFkN1u2bhk6pOhSoA9d2rRFIjegpGuZM/VRkqrjEfw1exmGve7H7N/sUpJTSU4GV4AlmvfwnEvey0v6GWdSue6/nWvXP/jwo4/Xur1Pv3sxs31W7dfyDQXARwHKU/FK59K4CyBY8Uh1eZABr7HF760wNrf/kWhGRp8lzNMxjFNErYmAVUYdfZ+p+/QAKCBcQSUEFkyhF+3uT05BseiGTGadzqeYctOePiT72itkT0Xkj7Tb/3Z3m7h9t3iZOwXoZ/O1x9GHkJLi/xPz+nV+Oq13hkYU/w7XHTPLSl+j8yZrgYaLwhTRSrzJiFn6xsoUDxkuTGoGhtO9Tw9u7X2gxemQR5DogJOpTwduJkaSoUCziYnpdLyGgwpRHoYkaGbGFXSMapwF+iSNHbwNFYynsIzKmaiLbNt5me0V+PvRpolWa4gCUqhc6JSomNCgmZgEDxOaFBgP7mVKEfDQYZUA7hfv/74UhPAIOkcAYCgzcL0jimSai9MY0Zn4cwpjlXuMRyvGj3tnpbpEgGSRBK/JxzUPI+9nuIwdAXU9VxHqU8NBpDa/Rg2+3vfbPtmiWMgHmFcfsIKH8Iqzvj8xKGfSk6SenPuargpiZKpe0RviRzTzKrCuVl8TYvYq9OXHm7IjVO/r2o7WKtRu1zjOP7rAfC/4e6SaAmIHNfYgbhnjYad81EYt7AXb2D+xOHNFcZ05Rz3NyFiMSdLmDz3YihlSBXaVwps/dO5mLNFUvmV+rcW+jUVUp+czoYLdpW6+LE8Gdk3sq+K0yVhw2swnG4PL2VsCq2rURMpgLDTacWaRjaQVINqGVmBlWMX5nQvS1rS5M2oqPOu5qkNZcVHihLwLIOAU5sDUt5uf/rKiw5MvZNPFvCfF8ftefHkpMm4qSNOMDMWBgP2sajmnGJ+6RmXKI9ZNwHoRYQrIb6sI2BN1iobGgUJpwAr23WO0Fo0mEBV/vI7454EmaGVxNEePJ1OJ6eGF62DY9vjA9bptClqjKFi6pPcJDm4GoT+HRUqD9czseoxOWQSLnSyk2bSGmS4hZC+GlMUTUOrS1UzkGSjcnRHBbU0yPF+IZL8DARX08el5EvAmkURqUmZeEoKdlxvES3VhcfmQSuDoCeZgf9hyR1Hf1IoFJjry4JlUV1FwrmYsVJZgF83dLcLw5E9zkgAehXhO0vC3XKrC4QJI0jE2kjCdNZnQXzlmS2XwbRjOKRjkQYIpGeYBzRofsIvigZgAJLoSgkaAZHt543Ri0LxfLjRfD/uBe3/1uPHgx+riseZ84Xzp3HEGzp7zwDl0njrHTtC50dnt7Hd+6pruH92/un+X0GudasznTuPp/vMf17Wf+w=</latexit>

Justification for use of convolution neural nets (CNN)

<latexit sha1_base64="ViBa+27vzi8AM+b0KOueVjoCxPg=">ALU3iclVbtw2EFXctE3kpnXSx74QdRZIg2Sxctq4LRIgBPDLlLXtWMnsGUYlDTapU1dQlL2rgl9UL+mTwWaPynQlw4l7VoXu9gQvghzDs8MZ8ghvZQzqQaDzcWPrn56Wef37ptL35x58uvlu7e25dJnzY8xOeiHcelcBZDHuKQ7vUgE08ji89U7XDP72DIRkSfxGTVI4iugwZiHzqULT8dKaKzP/ICvf8mkmkEkTATJAkJH4SnyU8K8wxZIJy/KckebC2tfVdbh8vLQ/6g2KQ7odTfSxb1dg+vnv7HzdI/CyCWPmcSnoDFJ1pKlQzOeQ2y46Tql/SoegizXmpIemoAgqTGJFCmuDRyMpJ5GHzIiqkWxjxnglZixChrLp1fOi3K5bDjMV/nikWZxmCmK/DCjMOFEJMYklARPgKz4h1PdxXRlVG9D4Y1zpE3wZnoDkD7lEDzv/7RypEfAz0AhLiCGcz+JIhoH2g1pxPgkgJBmXOXaleH02+7ZPVIUWhKkEi/jHIvzM9pd5OC+KHKlo2yY8CDXYujl2nk06K/+8GiQdzgCJhVn0EdC+Yu0ejAez6B0g5E04+TUw72oIGIlaKGnRNXZBy08Qrj3J5+5C1l435+5SrcebVx+R+lbdI1p7bHsZQfkxKk/592U0GNoDyi5gTmh3JHvwEP9eNB/ilOamca9zjm5b5dRyfcZFXD/csqK8UGuc3Ktn9VLP7O8KxZPrnf0ZOqo7qZ0kB/iWXAvnT3UV+0WBwWKP5WmSVGRIWIOAFl2rpZveZjXSd1LUYiZk3ax2dCU03g2tRXYMOYI3yTxSp8Yhe6pCmcUlEXniNWDNWUEwWLZBS9YDcFn2GLNt0w4WUjqHth8U5u/hLXtD7P0zvtBbL4oMk4aDPWcHdMwbU2uFMDO9oHNbAju1UDt9rgBuMeCDWlYM/UG20OvMd2ZgqncEMK4No9w0tFMBoP8V7Rz3PT1OsTXidpjqkWEXnd1no5g162oc0ZtNmGAharCgs63tbx2mYg6ktY7+wC78LMx9IFZQXJBer0mpRxizLu7NE2Q3QY5y2Gm0QwpJ2YvdMW7Tb/FoM3hEZY05UaxpGKNp5DfDmZGqSl7kNyuvJ3V92Gg50eXimo0fO8IDj/Veyd1O/4E8zXJF7pDgxFenXUjK8RKNnuE/GJfpbinAhIKCUZoBZkYoymaMJycz48nMuBkHzNfVo60MwO2ReKVdukLvT1VXk9wunvOAhRpQvWbMqrcVrOSNKpon6FHSBOIuYwoeL3XtcDLILClvykGAvJym+LIhkF0BA+X0ctisB34fxUI2w28NYoTt8E2nex/fOnXQ9KHYlAIfAQXWvCm702pBEobhf5Kq5CKent+jANFYVwRfhWdNovw+7H/krfedIf/L6y/OJZ9Wq8ZX1jfWs9sBxr1XphbVjb1p7lW39Yf1p/Wx/sv+x/FxcWb5bUhRvVnK+txli8x/AFycm</latexit>

(see Theorem 1, Jin et al., IEEE TIP 2017)

Recognizing structured matrices

L: convolution matrix ⇒ LT L: symmetric convolution matrix L, A: convolution matrices ⇒ (A + λLT L) : symmetric convolution matrix

<latexit sha1_base64="2K/zYgzd9XbfE3rfemGABsqH6M=">AMt3iclVZtT9xGEDbXt+R6aUL7sV9WhZPSlJx8R1uaCKREJAgklLeEoEpWq/n7hbWL6zXcMfKf6f/qf+kHztrnw/bBy2x7sXaeWae2Zmd2XEjwWNl23/PNT7/Isv3rwsPl169E3j5/Mf3sYh4lkcMBCEcqPLo1B8AOFcCPkYSqO8K+OCerxv5h0uQMQ+DfTWO4MSng4D3OaMKl07n5/5y4sQ1H2B6F1iI0mseDEisZMJUIsEjPlWSM4jTpuPCgAfa+JVqJ2PXfjKQME4JcWQiQNudlV9glBb/aZMQ7agwiFa68GIOAL6yqcS7az1OsvgE4cr8I3YNmLJB8NCbjMfOY2YLDrvyPbiS8LC4DIUiXE+92uEouEesTZNZpUyvCKTFZQ5c9MtGMx74PZiO32KiQLGX/r28jwyDcTfUaJGfcIcYfo+W2H9cJP/D384deB1hzhnF5BEVNh0IvDzWpHn6ZMHu2NlDZl+6k5cFa/LsnM4/MfxQpb4ECgmaBwfd+1InWgqFWcCMK4JxpyczoAnWUyJW1c8kg/lPgNFMlWKzjqx7gNF5Ho9zCuy8zirTKzIuN+XGV1XUxveU4Uf3fTjQPokRBwHKH+onAYBzeInHJTAlxoQyhvtKqEJ/Kxb2uyfaOG/UK4KYUQHeWudF70QPQVyCQrmEAK5Y6PsUI+30qc/F2IM+TYQ53G/eG+2m2SFVNMErcRAhQ8UtcdxCDtTethVB4qZYDN9XdJVMDS3Y6gzH1kmPsDgLyL8LKzrgigZwGPan6KaiL9Y7nhU+L0KDTagU2pyVYs2zo72+5KO972sbtf5JtE6572nYFpvJTQoLw/7JdtaCGkHdBU5apxqY1jPHswDP93O78irVSONZF4IsNnOvYmwAEhZvVHqGg9xFcifPyg3PNO6KB+O7iZYLojJNTpAeYy04N2TP9G2npWuatfmZ2DQhyiJETAGQhe7t5msM9yUps+T3REFSTzYfmHQaBqJNfiU2jHu4b6I4cZ80M7ukajismz4Hr6iqyadaDALRtYL9iJgnArT1N1Q5I2gzMKD3dT8Ese0PtfVu/UN8uCoijiqI9bxdBTC9bpwtyScsX1UEs6YfV8Svq8LN7lwQaoCgj1Tb9YxcIHtzCRO4YGUILRziZeK5DQY4L2i1LT1MsK2GUYqilT7brt5MRW/qoq2paKsu8nigJjJvhm0DRyMOsryFjZlT4F4bfUydl2eQXKOdhUyqkFGM2e0jpAziKsawgl9GNAZn93zGu58tvnVEGLGyAhjompq6KGsx9XDm5OrcZrH1suvJ+dwoVsh0HnxFE+bXGKB4/2Xo/ciluGLCE/AbZJVzAT0LjfZv5H6q/m0ZaS/RyjODEg/Q+TLANFyNJmFs/Opotn08WtwONMT46yr7tTwY4M3Xw9ZlLvFJY3QlR3rgHQ6p0hpqvB1FuUYFRb1W+wAQSi4zxUOLs328+whe6CUGZCxl5MIJwsS82sgoFgHn6YTA87gwUANsdvDSCEdzkS6+3M2ypaEpg8FJhU4BGSy6k1RzPQ4hqNpY6HTqyVSUTctXR+moaKhfFbs1ifD2ZfDXqe73LH/6C28Wp1MjQ+s760frKdW1qxXlmb1o51YLHGo8ZyY7Wx1nrROm31W8Mc2pib6HxnVZ7Wxb9iXaIO</latexit>

Applicable to

<latexit sha1_base64="A9LUv0MPS0qEpvFm8ec/eSoPFj4=">AMuXiclVZbU9tGFDb0lqhuE9rHvuwUPJOmic2TWnT0CZDwsAMSm3ZECUWa2O7YXVJbsrsNnR/+lf6j/pY89KlpFkaInGF81+37nsOXvOHi8WXOlO5+5+Y8+/uTz+7cdT5vfvHlvfsLXx2oKJEM9lkIvnOowoED2Ffcy3gXSyBp6At97ZmsXfnoNUPAr39DiG4AOQt7njGpcOlmY+6vlqsSzH2BmB1iE8CUPB0RpmTCdSPBJQLXkDFTquB4MeGisY6lxM/MmSAYSxikhrkwEmE575QmM0uI/dQgxro5iBfFqD0bEFdDXAZWoZ7XoaAuFxDYOGOhSUfDAu8w4LUaWU4WXJfk62lp4RF4XkEut+7tgIofcJ9Ym7Y0WplNEFmaygyJ97ZCKpxkEAdifX6KhaeZT9v7jOGobhZnsPrBT5HveIGfBpyfx3S+R/HMjtv4gx64xi+oiOHBdCPw82cU7uL3banewhsy/dyctiY/Jsnyzc/cf1I5YEGomqFJH3U6sjw2VmjMBmMwEg07ZGR2AyVKZkhYu+aQfSfyGmSrFR4NFO7CQya6PVR1zC5ei9kVqfqatXzML/laNE9386NjyMEw0hyx3qJwKDQezxJT6XwLQYE8oY7iuhGv2taNjrHhvrvBWvAIpRAf5q+fesRmCOAeNuIQLlgUBQj7fZpwMXYhz5NhD3gql+8Oy2nRbJyUgSpxEuEAK2e4rqLHKy+aTFEwk+NHip6T6yRfCok85wbMHknE4bCfkXaWVnPJFAbgY9qfopqIcVj+eFT6vQstNqCTrTGqxptuZvr7mo71vqxu1/kG4brlvq9gSm8kNCgvT/0l3VoIeQt0FblanBrjVUeHbgoXncaf+ItVI41kXgiw5uVcK61/C0pVIz9ogNxm50c7KlZ1p3DUPxzcbWi4Mlc3kBtIjrAX3ythDc91p6dpubX8mOm2IsgRWwBksXu9+pqF2xopW8kvisJIPdl8YNpLRBj8yuxYdzCfRvFifvEyfSquKYyrLiW/iKrtp0osIsGFkv2I2BcSpsT/cikTeCshUe7qT2l7i29Xme2alvkIeHVcZhnbGp6MA1+rgTgmc0X1YAmfUvimBb+rgBhceSF1QsGeajToH3mM7s4nTeCAlCOe46UiOQ0HeK+Y1dQ29bLAVhSnGoZkK26rpdT6GUd2pxCm3XI56GeYP6MtXUcjI8hbWZ06Bd2nlMXV+nkFymdpRo0IZ1SijmTNaZ8gZxkWN4UYBDOiMz95ZjXc2/xqDGjZIQx0TUx9FDW4+rjzcn1OM1j6+fXk3uw2K0YMHnxFE+LnGOB4/2Xs3djlvGLCE/ILZJVzIT0OlfZv0KDZ/mwY9HfY4QzBTLIGPkywHQRsrTZxdPT6eLpdHEz9Dkzk6McmO4U2JaRl68rJs12oXk9QnH3gvswpNpkrKnIq1GcS0RxodG8wg4QRoIHXOPg4rQeZw/ZBa3thIy9nMQ4WRDFL4GAZm18HFcBTuHhQA+x28NIozmciUz3BzvLlkHbh0KbChwCMqx6UxRTPc7hqNpqaPdqidTUS0vXh2oqCifFbv1yXD25aDX7i63O3/0Fp8/m0yNdxrfNL5tPGh0GyuN542NxnZjv8Hm780/mf91/rfmL03aHDZPc+r83ETm60blap/AWAaomw=</latexit>

Connection with deep neural networks

20

LISTA : learning-based ISTA FBPConvNet structures ISTA with sparsifying transformation

X

Unrolled Iterative Shrinkage Thresholding Algorithm (ISTA) (Gregor-LeCun 2010)

slide-11
SLIDE 11

Recent advent of Deep ConvNets

21

CT reconstruction based on Deep ConvNets

Input: Sparse view FBP reconstruction Training: Set of 500 high-quality full-view CT reconstructions Architecture: U-Net with skip connection (Jin et al., IEEE TIP 2017) (Jin et al. 2016; Adler-Öktem 2017; Chen et al. 2017; ... )

Dose reduction by 7: 143 views

Reconstructed from

from 1000 views

X-ray computer tomography data 


slide-12
SLIDE 12

Dose reduction by 7: 143 views

(Jin et al, IEEE Trans. Im Proc., 2017) Reconstructed from

from 1000 views

2019 Best Paper Award X-ray computer tomography data 


Dose reduction by 20: 50 views

Reconstructed from

from 1000 views

(Jin-McCann-Froustey-Unser, IEEE Trans. Im Proc., 2017) X-ray computer tomography data 


slide-13
SLIDE 13

OUTLINE

■ Introduction ✔ ■ Continuous-domain theory of sparsity ✔ ■ From compressed sensing to deep neural networks ✔ ■ Deep neural networks vs. deep splines

■ Background ■ Continuous piecewise linear (CPWL) functions / splines ■ New representer theorem for deep neural networks

25

Deep neural networks and splines

26

(Glorot ICAIS 2011) (Poggio-Rosasco 2015) (LeCun-Bengio-Hinton Nature 2015) (Montufar NIPS 2014)

Deep ReLU nets = hierarchical splines

ReLU is a piecewise-linear spline

(Strang SIAM News 2018)

Preferred choice of activation function: ReLU

<latexit sha1_base64="i1WjvXK+ND2ZjaRy5L3GdcKVaYU=">AJiXiclVbdbts2Fa7v9jdT9reFNgNsdTAUDSG5GJNF7RAgBRFCuQiS5e2SGQEFHUkc6Z+SlJxXEKPsqfZ7fYAe5sdSnZqScngEpBN8Hz8vkOew0MGueBKu+6/t25/8eVX3+z0evf+fa73/YvHvrcoKyeCEZSKT7wOqQPAUTjTXAt7nEmgSCHgXTPet/d0FSMWz9Hc9z2Gc0DjlEWdU49D5o6vikABM0cSIpASQsImGWdAsohQpvlFBSRkTLb2SXHcHhS9s83t9yhWzXS7XiLztbeA6dqR+d3N3b9MGNFAqlmgip15rm5HhsqNWcCyr5fKMgpm9IYTLWukgxwKCRJvFLNalGziaKDVPAkQmVE9U2YHr7OdFTp6NjY8zQsNKauFokIQnRG7STkEpgWc9wBhv4WVKMfDQbFqIDwxfDX0dhMQFyAxhVISGHGsiShaWj8iCZczEOIaCF0aXwVLfv9QX9AqtgoglASFEKAVrs47iMGQ1kt1SRFnImwNDIOSuM9doc7vzx2yw5GwnyBcYcIqD+ErToTiAJqGfSk6aegAaPhoSjqSa16JL4shBgrCpclsv/skVs1dcnXni7JjUu/rOo7WatRx0IjOPn7AfC/4e6SaAnUB8qe2BKQ4w/UZg38Mhsu8OnOKW5y5imQpCHtU/qQ0ElPw0Y2QlyE0aN8rsfJK52nPN0/mNOk+WOqsqNX95o2Xfav1yFyXJ57dE/uz4LTJWHazCdb3vXsLYF1NVZE6mAsNdpx5rGNpBUgxoZWYmVYx/ntK+9rWtLkzalc5V3PUxvKBR+pSsCbHBingtgalon6/K+q8PS4tL/EtyUsCMxe308PW0iTtuIfcyMpXG/bTxeMXa4T1eMHdoDLgKQegnBamgO2hj4gIXKhkZjwkQxr/Aai85TWMs+OYF4psTDrO8xN2UCTlsc728Mr1sm15fmV63TSFP9cIWdtRe4R3KQa4u4VUn0MGlnY/RCesgkctOik1biGkHMWsh/CyBmFqHBtVI29A4+mMCR5rkuP9QhT/CAQ0G2Lr+wrwYk9jPakzLw3BrstlSacuLN8CSkucPcEcHI5a7mgalCvFAhMdecqGiIZLPeOhtvM9tPYGhEdknhUkBbwJ8Zqk4R+F0pW/FdA+F2boWDbrD1ruqinPbUcNYhmFG90JMIySws8MFh/bF7EICeAaS4D5LGkuZ4d+Nrw2u/Lbqdt6Oh92Q4+m20tfe8fnY4G86Pzk/Oz47n7Dh7zoFz5Jw4zPnT+cv52/mnd6fn9Z71dmvo7VuLOfedRuvt/weGzXkP</latexit>

ReLU works nicely with dropout / `1-regularization Networks with hidden ReLU are easier to train State-of-the-art performance

<latexit sha1_base64="TUO79kdUG89HwGXwCRHxGijJfmo=">AKY3iclVZtb9s2EJa7t0Td1qTbhwHDgMSA0NRe7KDLVuxAFSFCkQDF2ytEUiI6Cks8yZeilJxXEF/ZR93X7TfsD+x46S7VpyMrgEZBO8u+c53h2P9FLBlXacf1r3Pvjwo48/2di073/62ecPtrYfvlRJn089xORyNceUyh4jOea4GvU4ks8gS+8sZHRv7qGqXiSfy7nqY4iFgY8yH3malq+3WluthyOPcBW5WyLmURZKnBYArswE5j28KcqfwgbIXZ2kCtODPt6AK3CoIyYJ4aDf3cMIXK4xMmLHiCUPR3O540eFXYrhFE/OYZLIsYKY+yimMOF6BIFM0iT8D3suijEVW+3IzHMBJP8benvzPxX1JVxaTXiQYBxhckAjLFUYJOQEvG5zZnmnsJMOHmGHSQ0pymEiIxb7aLsYB1UEwL7a2nG6TjlgdKbTXYOv7LK8eJqe+OJGyR+FmGsfcGUuw5qR7kRMJ9gbTnjOLB/DELMS/jW0CblgIgfvpiDeVqTY9FSk0jzQjpkeqKTOLt8kuMz38aZDzOM0xn5FNMxEGQ1KPwRcoq8p4Mz3yd+MghJADUH5TGBw0P25P8hHK5R0w4kxjxk4jCFeTukEVcTAMcskyYolHD+dxu20oq04BqYKXCYFaPaF1l3SoSBcFloigyGXoUWU9dr7Pzx2ihUdU4SVjtMlheojtWVnPJFhRUOe1P0UzKODQdni8o2sWsrA2rqezZf9EANuzrA8+PzHrQtPn3gjbBWg/aE5TH94kHqf8PdB2Azo7KPEUVRIexyKkXjBTVDT7KO073RzKpR5nKVAg6yWJepPR6dx9Z9E3FHAXx50+9oFjHXPJ7eybM351lmqfCLy95gPjdcj/Lb6mTR/maYphLTFP5sNO7Hb1BsC7HEkmVjDlHM8NJk0BJCb1ErqDOs431l4X8FCHTelZruEu56nJpUzPChbwFmKPmcCTA9LRHX+l1l4fFqYX3BNC/O8/LS5Px5f1DUumhpHVBlz4VFTeLokXMG+WBKuwB5z4aHUcxXqhvlxUwfUKMyqdFUcBJF7l5Tt5ecxSE1/PyA9OsGJ0laUDRlBCdNrKcL0dOm6PlC9LwpCnisZ7Jghe0ZvQ7oClzewrOVRHs3xp6yE1RJgpuVEhs3NMYrGpOGhptEGDLjULtTDjhDTaczBDrWkNL9Aoq/pTta+10atquQnixqEdV5cV0k9O+qdCvS/MXzlKS7IeUQ12+w13NPOKpWZBhU4RY1E42e8EAb+x5JN9vAhzBNMoiRbkK6JlnwR6Z06W+pCMmQJnGQTOx2w1015qnplv0a0ITRjU5AlCGZBJlPWyZ/TAw8ercgPVXIlIWSpXR302uj13xbrE5e9ru9vW7/t/7O4S/Vs8PasL62vrW+s3rWvnVoHVsvrHPLb01af7b+av29+a935of1mp3mvNbL6wasP+5j9qusf</latexit>

Deep nets as Continuous PieceWise-Linear maps

ReLU ⇒ CPWL CPWL ⇒ Deep ReLU network

<latexit sha1_base64="dizmXLExvdH+WGD4SBANcx9kuzw=">AKwniclVZtb9s2EFbStUu0t6b7MmBfiCUGhqIxbAdbtmABgiUoUiDsnRpikRBQEknmTVFKiQV29X0Nwfsv+zDjpLtWnIyuATiELrj8xzvHh7p5xp0+n8s7L6JPHTz5dW3c/+/yL796uvHsjZaZCuA8kFyqtz7VwJmAc8Mh7epApr4HC78waG1X9yB0kyKP804heuExoJFLKAGP91srEhPZ76GID8CSIkAownV5FAKw0QmM01OGQRwTRsnyAHVShqS5cz4eYidyGWOReGUieZLGCcUGIpzIOeRdGRflTuOu5Z2SqId3vwYh4HCKTUIUA+732DiTEYwYSa+5Ys2Jxf2rvBEnhtko7+RXGUoRkywPOb7pbeyRVTAQs5RASmqZK0qBPIqmITiFg0ZiJmCiIM04Ve19umESZCOyE8inob3T0e2YQ9MyUqXkcIscnl6cuJX9DE7OH7bacNaJrJchdkcSjVwPRBhlSni3jzd7LQ75SCLk+5ksnwjVO05uNtT0vlEGWgDABp1pfdTupuc6pMizgJXIMHE0GNAY8rIOBWnhp7DMRISFJOXmh9NtB4nPnom1PR102Y/3me7ykz03XORJoZEFGWcGEmsukjIFASGjwkNAow3owbjqCHogGK59ts/967zPvA7MLgDBQKGgUwSionyIpowPg4hohm34tLRdO623BYpRY0qRSX4Geo2D387qEP6nMmRMnDIlexjwp80Wnv/vCiUyz4WLFWPp02OlR/6DYfjM8zqGgwknqcnPp47lAFbHYCrHcxkb9ltSdg8r9oAFv25YGnR2s5aNz8R0HbZC0H7XOs48fkA93/B7oOYPpQNSN7Qouc5F5fo27geb7daf+IS+pZRplyjoevJNG3GVWw9WFz1KQhzgepNn9QDPLObC8YM8O1OeZYKv7jqXk/nlut5fp9OZm1ygmnFWGJa5ZPN7v3oDYJlOeZIqmJMOZp1ZrGtpCUguS2tws6wTPDbs+grWFLHTbEVz+EuF6kt5QSPlC3gte3vlBPbwySvzv8CxNnhf0lnm1hvp+fNfHxGXd47LpcYjKmBoPm8azOeMC9uWcQH2mHEflJm6YDfMj5s+cIuNypbGoOAU8Ny7w26vGBUxNvx8H/3rC05kWmA2VUJOmlhHM9NR0/RqZnrVNIVMmIktXGB7iY8PBmp+Cy8XCu2P7HqsTlgViYwWJDZoeAwWPIYND08mEFMbUGu7HOQ1GOveTzWJMX7hWj2HgiYoI3D9Tgi0jEpl8pT4Rg91W+KRp9YfqI0kbh6j5qsN1rhGOoX8w1CxQ64hQ1EgMjM2Shseu7aF1vERaRsczwLYA3IV6TNHyXaVPGWzoSGeFEhHKID5J6uHrAUtstezWgIcUbHYGwQkqGWYBbxnhsDnx8bgAIzIOisaIp3t342ug23xaLkze9dnen3fujt3nwS/XscNacb53vnO+drPrHDjHzqlz7gQrf6/8u/p49Yl75L5zb1dua6uTNZ87dSG+9d/ZobpnQ=</latexit>

ReLU(x; b) = (x − b)+

slide-14
SLIDE 14

27 layers nodes (n, `) …. ….

neuron

(n − 1, `)

Linear step: RN`−1 ! RN`

f ` : x 7! f `(x) = W`x + b`

Nonlinear step: RN` ! RN`

σ` : x 7! σ`(x) =

  • σ(x1), . . . , σ(xN`)
  • Learned

zn,` = σ

  • wT

n,`z`−1 + bn,`

  • Layers: ` = 1, . . . , L

Deep structure descriptor: (N0, N1, · · · , NL) Neuron or node index: (n, `), n = 1, · · · , N` Activation function: : R → R (ReLU)

fdeep(x) = (σL f L σL−1 · · · σ2 f 2 σ1 f 1) (x)

Feedforward deep neural network Continuous-PieceWise Linear (CPWL) functions

28

1D: Non-uniform spline de degree 1

<latexit sha1_base64="urJVSURcV8NBeTnP6fI4R8FVsfc=">AKtniclVbtw2EFXSW6LenPaxL0TtBYogXkgbtE4NBAhgJ7CBtHDt2AlsOQYlzWpU5eQlL1rQn/Rr+lr+xP9mw4l7VoXu9gQuwIxc3hmODMc0s84k8px/r13/5NP/v8iwcP7S+/+vqb1cefXck01wEcBikPBXvfCqBswQOFVMc3mUCaOxzeOtfbBn920sQkqXJGzXL4DSmUcLGLKAKRWcrQ0/mvoRAu9ub5Pc0Wc9Rm4qYyMxQktD8IgFA1ty1wj5bWXWGTjlIf+LWk1WrHntnjx5mXpgGeQyJCjiV8sR1MnWqVAs4FDYXi4ho8EFjUCX2ynIAEUhQS/wnyhSls4Gks5i31ExlRNZFdnhLfqjETIsWxb9f24sJuSk1yNn51qlmS5giSoHBrnKiUmBiSkAkIFJ8RGgS4r5wq9LfF8MY91cZ5s7ylkAHlED4f/jo61RPgl6BQLyCBqyCNY5qE2hvTmPFZCGOac1VoT47nc3tgD0iZU0kQSvyc1ByE+UeYjBfZax0nEcpDwstIr/Q7hNnuPHzE6foYQTMaowzRED1R1jTGZ/nUJlBT9p+cupj2SmIGaoqUoMuiCdyDtpYhWlhzydFh9mYX565dndZbtz+R3GbcC3J7XNM5ceEBOH/x91mUBOoDqQ5noUm2ptIrB14rNed4S+4pB1prHXOyZpdeSU/5FTA2s2SkbFB7jJyp52NGzuLuCuWzO429HRuqGmMlCc4Fnwbow91rdVi4sE5afmNCEqI0TMASCr7u30HQvLGmlaKROxMNJNotMOo0Fok1+BTaMJdw3UazdJ3bJS9rEGRVN4iV8RVdNOpGwDEbZCw4yCBjlxHTDlFeNoGmFJfuF+RLPtD7f1/vdDbLkuI047iK2sDrmyq2ucr+h7HEfN5Q92h3GfRBqDsG2qHe6GPiAHcvkRmHNCeDau8R7QzCaRHh16OeF6dvNBa/TrMBo4vX1usu1vVBtd1W7C9VuVxWyRNW6sGftFV7CDERzC696ifavzXrMTlgliVz3amzaQUx7iKsOwktjiGjPIf+ig7voN68OgvdIprhZYQYe+9o1W3hdVXL8zEgl3je8Dq0AdZUOLn0ajBA1IWcA36raIc364FqJaLWEMZSyM8P18IzxfC3SRkga7rK9buQrEnUr+Sy0DovaK6+DBbuNy7YiFMqNIlarHk5TSrVqTZnFG/xJOXpJzFTOGDwR6sl4McgMJWGBHsoSTDG51Idg0EVDEYXsS8AmWRGqCXRamCs3hW0S7owDfGE2lOf9JCKa6nFLX7tDzJ51UAqkNw3DUSaWiftFo26aRIRHuCN9obvdF1p8cjYbu06Hzx2j1xWb9Wntg/WD9aP1kudaG9cLasfasQyuw/rT+sv62/rGf2e9tsKMKev9eveZ7qzXs7D8tyud4</latexit>

τk τk+1

<latexit sha1_base64="YA93wTA38YDHkPJWQBgK3L9kl4I=">ALGniclVZLb9w2EFbSV6I+krTHXog6CxRpslg5bdwWDRDAiWEDaerasRPYWhiUNLtLm3qEpOxdE/oXPb9Mb0VvfbSv9JTh5J2LVF2sSGwC2K+j9+MZsghg4wzqQaDf65df+fd97/4MZN98OPv7k1u07n+7LNBch7IUpT8XrgErgLIE9xRSH15kAGgcXgUn6wZ/dQpCsjR5qWYZDGM6TtiIhVSh6cBXND/SJ195xdHtlUF/UA7SnXj1ZMWpx/bRnZv/+lEa5jEkKuRUykNvkKmhpkKxkEPh+rmEjIYndAy6DLQgPTRFZJQK/CWKlNYWj8ZSzuIAmTFVE2ljxngpZixCjmTbaxDEhdu0HOZq9O1QsyTLFSRhFdAo50SlxGSHRExAqPiM0DE78qpwnhbCi+9oTbBm+UtQIaUQ/S4/93qUE+An4JCXEACZ2EaxzSJtD+iMeOzCEY056rQvhzN527P7ZGyWpIglQ56Dk92j3kYPFLXOl43yc8qjQYhwU2rs/6K9c39QdDgCZjVn0EdC9UNaM5iA51C5wUjacXIa4IZSEDOEKlHDLogvcg7aeIVp4c4nhaVs3C+vXIe7rDZ+/ltpm3QtqR1wLOXbpATp/6fdVlATkHkgcXvhwSs0f5E4t6Be/rBoP8Il7QzjXudc3LXraKSb3Iq4O7FklXjg1zl5Eo/axd+FnlXLJld7ejh3FHTeWgOMSz4F84u6cv2y0eCpR/taZJUZkhYg4AWfEul7c8LOuk6aUsxMKJXWw2NuU0Hog29RXYMJYI32SxDp+4pS5pC2dUNIWXiBVDNeVEwTIZS/YzSBklBPTDVNeNYKmF5bsFOaf+Kb1BYHesT+QJQdtxoHNWMfdMQfXbXCnAXa0DxpgR/ZFA3xhg5uMByDUnI9U2/aHiD7cwUTuGFMC1f4qXimA0GeO9oh8Xpqk3FzxPswJTLWLy3NZ6uoCe2tDWAtqyoYglqsaijrcNvHsZiOYnbHR2QXBu1mPpoqC5Bx1em3K1KJMO3vUZogO48xi+GkMY9qJOTixeCfd5mcxeEdkijlR1jKMUNh5jfDmZGpWVLmNquvJ31/xWg50dXjmo0dO8YDj/Vexd7Ow5M8zXJN7pDwxNenHSnJ0gcY/4D6ZVuhPGcKlgIhLRmUGWBihLJsxHh8vjMcL41YSsVDXWznW3gLYFmlQ2WUo9PZceSPF5f4Zi2BClS5ZiyXPplm1Is3mivoZdoAk5SxmCh8ubu9BOcguKGzJY4K9nGT4siCSnQMBFfZxuL4EfOQlYzXBbg9The7wTaS9h/21EF87Tdh0osQUA58BJda+K+bPRqkEihuN/qpVSkWDonGBmJaKQvhN+Fr07Ldhd7K/2se4Bj9/vfLkUf1uvOF87nzhfOl4zprzxNl0tp09J3QS5xfnN+d391f3D/dP96+Kev1aveYzpzXcv/8DLTUS5Q=</latexit>

Partition: R = SK

k=0 Pk with Pk = [τk, τk+1), τ0 = −∞ < τ1 < · · · < τK < τK+1 = +∞.

<latexit sha1_base64="0MVRDrbEWDyFqlqOEDPeDF7T08Q=">ALkHiclVZtb9s2EFa6t9Zet7QDtg/7Qiw2sLWJISXYkhUx1iFtkCxd5yZNWiTKDEqibcbUS0kqsUPow37mfsF+wr7uKMmOXtLBFeCEunvudPd8UgnYlRI0/x76c5H3/y6Wd37zWan9/4svlBw9PRBhzlxy7IQv5WwcLwmhAjiWVjLyNOMG+w8gbZ7yj9W8uCRc0DF7LaUTOfTwM6IC6WIKov/xXD3Mwg/UT1LIPu7ZDh24c9dW4ayZ/HqBef9xCV1SOUAuW3TNb4rg/Xk3/qfFjK/mhtQqG+tXsrtk0GMjpdvpqbSPbC6XI3g7gLbU5AJvu4wzYadl2o7+8YnbM9EH1hZUvVoz86fUf3PsXiN3YJ4F0GRbizDIjea70h7iMJA07FiTC7hgPiUozlKA2iDw0CDn8AolSaQmHfSGmvgNIH8uRqOq08FadlnAxEGWvjuMnjaLkLJaDrXNFgyiWJHCzgAYxQzJEuizIo5y4k0Rdl34rhLiLfE8No6Vzp4bV5SCBcz4nU7P6+fqxFhl0SCnpOAXLmh7+PAU/YA+5RNPTLAMZOJsVgtm60G2UtolAEVOzBiR4gnIbcBAV6W5Un48DJmXKD50EmWtmp3NH1fNpIbhZJpjzA4Ash/AisE4LCaZG4ikHCfDnSyJD4FVUaq0QmyecyI0l7JGnMFkmFWbtfnDkPd1Fu+PwP4tbpWpDbYVDKD0kJwP+Pu8wgR0TEjoD2gl2eKTskYDeIY/Umtn5CUzKmYZeZwy1GlU4l2MOWndmKxrH+h9Tt7rZ/PGzvkgbT9zvamDkquskcJGewF+wbZ4/Ubd1iAUH6J+fUKUozhPQGQCvW7fQVD4s6KXpJCzF3Ui02Hepyag9I6fpyGBgLhK+zmIePGikvKhNHmBeJF4gVQtXlBMI0GeksOIqISzFDehqGLBsERS80OEz0X2Tr0ec46rD6gTQ4LSNOq4gd6I6ZcqeqPCwoa9ynBWN9mVB+bKq3KPMIVzOIDAz1V4VQ97BONOFk9CQnDBlX8KhwikOhnCuqG6ih3rR4EUYJZBq7qMXVa5nc9Wzqmp/rtqvqjwayFzn1bztwqFPCS9+wm6tC5xrbQ+l87IKomvgaZchkwpkUuvRKoLXEFcVhB36ZIhrMTvjCm5cH34VBKuRTCAnsmIGEfJqXj04OamcJluvex4sk9WrJIDlW2e2dNGl7DB4fzL0EeRm+JnGc7BbZTumBz0e0Y5uNH629Ank0z7RwTqlID7KSITEzIXkrRsWnhxMRdezIX7gUdlbeyr6y5osdDJ5MLl6vejHk3BHP7inpkhKVKUXOT5MoswijGaN6DhMgCBn1qYSLS6O9lj7oiEgYyUMEsxFcLNAgl4TRKTbgadhCwK3y2AoRzDtyUSCO7gTKWvDhbtOUanUKBLAZeAVFc+KWa3VSE5UGuGznqlkBI7SeH40AMViBIEV0WrejGsL07WO9ZGx3y1vJ0K7803jW+Nb4zvjcsY9N4auwZPePYcI1/lu4vfb30TfNhc6v5S/PXDHpnKbf5yig9zd/+AzviNzQ=</latexit>

The function fspline : R → R is a piecewise-linear spline with knots τ1, . . . , τK if

(i) : fspline is continuous R → R (ii) : for x ∈ Pk : fspline(x) = fk(x)

M

= akx + bk with (ak, bk) ∈ R2, k = 0, . . . , K

<latexit sha1_base64="WcCz5u0SXIVwjT2e5prbGafhU=">ANr3iclVdtb9s2EHabdWu9ZWu3LwP25bDYQNI6huW2a1csQIG+oEW6LkvTF7RKPEqibNbUS0kqsUvor+x/7Z/s46U5UhyOqQCElt3z91zvDseaS/lTKrB4J8LF9e+uPTlV5evtL/+Zv3b765e+/6VTDLh05d+whPxiOSchbTl4opTt+kgpLI4/S1N31g9K+PqZAsiQ/UPKWHERnHLGQ+USgaXbv498GEQpjFvnmHTjSrohApsZhfg/cfXBVgh8dYBIpIz69IRJum0ARCyQcMLUBKZxoiR0XEWykdNzA3zr2ZdtA7brkfHLNZmVbl2bew6ysaCznMAV2Sc6kH/zm06y8vPvN0Fi5P3wOPEn/ZA0KAHaEPjHuBaKxID/w5iZfSOeU8OUFtNk54YD+RAj0Z4u2UCBJRhSmBT/swzMSpYlQ+KnoDNI+OM7NrbGRaeSpjtDFLqchioiAoPH9whcpmhklIP+baMWbDwp9QM/ytsWAO3OJtvCDALUk2oz6SexYnGWZCZfyxwvLNHQWHYgTAR0Zi6LocwXLj/fG02xMnWnm7OtnSomRAzKXPohoCGQ0RmcAO80bRTlKqzibIevm8Z7+7+0bDTg850Z7Ao26n3V0E4wYMSchcqjmnZ9B6o4F17VgKV2bRSE93nPxo1/JuzrZtD0y3RjdKcjRBbgepcdG9Ipgt+3a02+m3waVxUPTJ6OrGoD+wD6x+cRZfNu7/2LP3ujalX9xBX4WYWf4nEj5zhmk6lAToZjPKRYnw9JhK2HvaJuvHLoCmyQyxL0XE1HImknEceIiOiJrKpM8IzdUYiZCjrJ6HPVKVvMtUePdQszjNFI39IqAw4BdYTYtBExQX/E5EN/HdWXYqwHUPBw4h9oEb8xrCukToOd/q/DQz2h/Jgq1Asa0xM/iSKCeXZDEjE+xz4hGTc7U4bl93YXt40dIrj34wC8jHOq5D2Uu4jBDljuYtxquRZjL9dOz+ze3iBfwZidXmBw5/QWfwirBuPxjBY0GEk9Tk48nHPYk2w5Pgw6r8+O9nJ4NDwb+vN7LgfTOX3j8j/Lt0nXOX3bwfc5KUH4/mue1ATKjNPUjv/c42DZiKxd+h1vT3o/4Im9Uxjr3MOKgsjfyQEUE7pyZDwGfIvkz51TnmXecTzOP010sySq0hQE+TvcC+4p2XV9Vrc45pgx/xY+TYpshsBsANhwznbfYDgvSZWlOFKkmax2diU0zCANvUVODOEb7J4iJ8aFu/UHeM517V8TlixVBNOdGhTYadBS9S6jPCwUzDhBeDoMrC4v1c2xPFjD7P0/vNBbL4bR3xtol4gN1RKh80lfsV5YrvtxXlitvnFeXzpvIJ4x4VqoTgzNRPmh7lJrCKWxIQbl2j/FQEYzEYzxX9E5uhnrV4FmS5tqels+avh4uVQ+bqdL1dOmKmCxWuiCFbHeCVkVFSX8HilC7yPxh5LFxQVhI/mjlWHzBqQ2UqPNhFiBXHSQLhJRMdkJWZv2sBNV4dfA8FXnMwJ6phGKZl4DPDmZmudFboPieHJfbTg1Al1snvLpwjFucDz/CvSL1Lf4MsMLcBfsjlmAfi9chqfa6Dfsk1mh/SNFtXUgIosoxJQuhdSWzQjfv18K3y+FT+OA+XrRypF2lo9kXiFXPpC75WeHydo7p6wgE6I0ha1NHk0SwuLJC096kc4AeKEs4gpvLi0u9v2gRdU4UgeA85ySPFmAZJ9pECV38en7UqKvz3isZrgtKczhXR4J9LOLXsfrijNHIpNKfASYHX1k6L8LSOVQNfGQ3/YKQiXl45PsxARUe4IrwrOs2b4eqXV8O+c7M/HO4cf9ucWlsXW791Pq5tdlyWnda91tPWnutly1/7dLajbVba7fXnfX60frfxXQixcWNj+0as86+w+6nu/p</latexit>

fspline(x) = ˜ b0 + ˜ b1x +

K

X

k=1

˜ ak(x − τk)+

with ˜

b0,˜ b1 ∈ R, (˜ ak) ∈ RK.

<latexit sha1_base64="Ilfefr2XTOkQ4J6sUC52y41BGYU=">ANy3iclVdtb9s2EHabvbTesrXblwH7cltkIGkdw3LXtQsWoEBf1iJtl6WvaJQalETZrKmXklRih9PH/an9k/2TfdyR8pvkdEgJLuHj53vDseaT/jTKpu958LF9c+fSzy9dbn7x5fpX1+5+s1LmeYioC+ClKfitU8k5SyhLxRTnL7OBCWxz+krf3TX6F8dUyFZmjxXk4wexWSQsIgFRKGof/Xi363nQwpRngRGAE7U156IQWaGsdgB7wA8leLDASaBQMZoQE+YpNsGQMQUCSdMDWGUpEqC4ymS92F+JX237s4eyo6fl0wBJtlVozqv43wg6KQA8ETOqe52bt2k42L2LJotsDi5Az4nwagNgoZtwDk0aQMuhiaKtCGYkGQunVDO0xPU5oOUh/aJpDJGN7OiCAxVRgT2AyiDjwmcZYKhU9Fx5B1wHVvbDU1LjqTNvtodDjNFIxEeg8fsfgMUVjo+x2bhq1YIPhTN8NYnTaIqDZcjbZFsYQoBpWG8sgTRL8jQ3EZtH2fMOXTo+slPNXAeiVIAz9lgC+/3RTo1pc7y1G/VH+PDo+5BGQPojGMN18Psjp0yKs4myNn5vGQ7v4G3PaYMz2u1OE7TnNEt3HS9kSEsmUk04PcOQpxgPKVJ30cD8w7X2PJnHfT3adYu3ezOd8WVzvG0rYLTVv+6A9z4noXULY+Is+NoLNnQSw4Aebi5Ytqzw7Z7TaYJHk7Asof6VjW6nawesvrjTl4073zXs2O9fvfwvLjnIYyagBMpD91upo40EYoFnBZNL8esYpVhWlbdgW0UBTaHESYr7IYKzgSzmJfUTGRA1lXWeEZ+qMRMhIVq36PpbPsuQwV9HtI82SLFc0CUqHopwDlovZ0BAyQPFJ0CANeVYxmHUGF47h5p47yZXlHIgHAa7nZ+6R3pIeXHVKFe0ISeBGkcE4yzF5GY8QkWFsm52bQymr03W7ijbIPBtpCE4OecUyV3UO4hBmtmvsFxFxZaDPxCu2zsdvdYgVjmkCJwU3Vnv4hbNkZn+e0NIOeVP3kxMceiFXM5p3FoItqW2nO+0qN2Zg/P/OsZ52TG5f/UdwmXOfktj3xY0KC8P/jrjKoIZW5L6k9GgoN2htKrB16TW93Oz/jlGqksdY5B+wm1ozErS6os5jSMzbgQ0Y+aOfWws487tg3Jx82dGNmaNlMaA4xL3gLYxd02dVi2tOIPNvymlCZCMEZgPAhns2fc3CeY0sWykPv5mRerLZwKTWABt8iuwYZzDfRPFqfvQtLxQJcYjcZn4HL6iqyadSGiDYXvBs4wGjHAw3TDlZSNYtsKSg0LbI8i0Pt/XB/UFsuRNFfGmjriL1TFT3q0rD5aUK9xvlpQrtE+XlE/ryoeM+1SoGQR7pn5Yx9iz1yROYUEKyrV3jIeKYCQZ4LmidwvT1JcnPE6zQtvz9XGd695cda+uejRXPaqrQpaoqS5csfYAr4uMiuUlPFipAv/UzMfUhWUG4dRcv6qQcQ0yXqnROkKsIE5qC+N6YCs+OyParjRavOrIfgKyRhjomrT0ENRj2uIJydTk6KMbVgeT97LDbdiQJebZzZacIwbHM+/Ev0sCyx+FuEpuAV2x0xBT0rKaKGNf8U6GZfa3zNUWwIRW0QpnQupDZtRvju3Vz4bi58lIQs0NSjrU7V+yL1C/lMhB6f8b8IMXp3gkL6ZAobVHzKfHWTkjzWaM+j52gCTlLGYKLy7N1rYd8IwqbMkDwF4OGd4sQLJTClQFHRxNT1L8XZIM1BC7PR0rNId3Iu3+ZK7Ky0rThxKTCrwEWF31pJj9zpFKILVh6PRqiVTEL5aOD9NQkQhXhHdFt34zXH152eu4Nzq9P3obd26Xl8bGpcb3jR8bmw23catxp/Gwsd940QjWflj7bW1/7Y/1J+ty/XT9zxJ68cJ0zreNylj/6z+A4vhu</latexit>
slide-15
SLIDE 15

CPWL functions in high dimensions

29

Partition of domain into a finite number of non-overlapping convex polytopes; i.e.,

RN = SK

k=1 Pk with µ(Pk1 \ Pk2) = 0 for all k1 6= k2

<latexit sha1_base64="PkX6ZMk/JBfwP85mvzOF51iXIU=">ALw3iclVZtT9xGED7St+RoG9J+a7+syp2URuF0BrWkL0hIBARtSq8QSAQmaG2P75Zbr53dNdyx8r/qn+kv6dfO2neH7YOKWOJk5n2mfHM7Ox6CWdKd7v/LDz46ONPv3s4aPm4udfPl46clXxypOpQ9Hfsxj+dajCjgTcKSZ5vA2kUAj8Mb7hl8TeXIBWLxWs9TuAson3BQuZTjabzpb97VOIyfCdxSI4okwQJnRMKAmZYBqISCMPpIVFLFZiVOM0SZjoE+N6IfFjcQkjksR8rOMEVPYLYR3oPHfdUwdGZ82We/Buf8P1WN9Pk3Mz3HCyd7+T3vmwRa6YHpCWG6VPewicOxlxfZqQ/J/V7PuNbouEsSUc9JC2BVAEGg1z5eWu51u/pD5F2fystyYPL3zJ4/+dYPYTyMQ2udUqVOnm+gzY7/d5A13VRBQv0h7YPJk5qRNpqC3H8YC01ya4VHI6XGkYfMiOqBqmPWeCtmLVKFqurV86KsWbacpjp8cWaYSFINwi8ClNOsDa2kiRgEnzNx4T6Pn5XSjXGW1F47ZwZG7xdXgGUTzkEG52fVs/MAPglaMQlCLjy4yiIjBuSCPGxwGENOU6M64Kp+/NdrN8s5SBKnESzkHrX5Gu4scbMQ8VyZK+zEPMiP7Xmac593O+g/Pu9kcR8J4wul2kFD8Ia0cjMdTKNxgJNU4OfWw+TVEDKFC1LKxlWTKwVivMqa05espmzd3195Eu59tfHzP0jbpue2h7HUn5ISpD+f9pVBT0AlXoK2wsHQ2Zwow8U9g48Myvdzo+4pJp7HW7RZtFVOp9SiW0bpasWh/kLid3+lm/8TPLu2ZifLejtamjspvCQXaKe8G9cfbM3NYtOLCy/GeiaVOUZ4jYDUCWndvlax7u6TsJS/EzEm92Kxvy2k9EGPrK3Fg3CN8m8VJ+KSZ65KqcEJlWfgesWKotpwomCcjnwWHCfiMcmKnYcyLQVD2wsRBZn+Ja0ef5mD+gcycVJlnNQZW9gdU3CrDh6UwDntkxI4J7tfAvfr4C7jePDpKQVnptmtc+A9jNbOI0NKYEb9xIPFcmo6O5YjYyO9TLC17FSYaplhF5Vd6OYNe1qG9GbRXhwI8rSdYMOdtB+8JDGT5E3bmusC7tuxdEFRQXKNOu0qZVSjOZ6tM6Qc4yrGsONI+jTuZi9Y03nB9+NQafExlhTnRtGUYo63kN8ORkepwVuQ2K48k9XnYqDkyxeaZPm1ziBsfzr2AfJn7On2Z4Qm6TfMdMSH8UkuENGv2KfTIq0D8ThHMBGeWMwgwM0JeNmu8uJgZL2bGPREw30xaOTLODOjJ2CvsypemN1XeiXG5e8UCGFBtctZsyfYoKVbEyVTRbOMEDFnEdN4cWm2V/KHILW9iaIs5wkeLMgil0DAe138Gm6CvBCKvp6gNMeRhrd4Z3IOGs+3nXKoJ1DwpYCLwE5Vj0phdcpSVKW4XOaq2QmnpZ6fiwAxWFMoJXRad+MZx/OV7tOGud7l+ry5svJpfGh41vG981njacxnpjs7Hb6DWOGv7CNwubC3sLvy1uLw4X5aIuqA8WJmu+blSexew/RKhKaA=</latexit>

The vector-valued function f CPWL = (f1, . . . , fM) : RN → RM is a CPWL if each component function fm : RN → R is CPWL.

<latexit sha1_base64="lrEqoBDRz57dUVgYykrYWn1kE=">ALoHiclVZtb9s2EFa6t1be1nT7uH0gmnjoisSw0m3pigUokDZIgCRN81okygxKomwm1EtJKrFD6D/s7+2PDPu4IyU7kpwMLgEbxN1zx3vTkd6KaNCdrt/z347PMvzq4SO79fU3z6ef/LdsUgy7pMjP2EJ/+BhQRiNyZGkpEPKSc48hg58S7Xtf7kinBk/hQjlJyHuF+TEPqYwmi3vxfhwOCrogvE758hVlGAhRmsa+VaNHdQWFPuTxC63sn2/nas7DnLlBIgVaCns7P79C7v6fu8iVid7sLCIqEDZY5Lo2DRHB/gD5SZQmMYlhTnsRYVxYWstV3H7s0vdDtds9D0xik3C1a59npPHv0LIflZB58hoU4c7qpPFeYS+ozktuJkiK/UvcJ8okLEdtEMFBEw4/iMtIazgcCTGKPEBGWA5EU6eFd+q0hItQ1L16XpTbVclZJsOX54rGaSZJ7BcBhRlDkA1dJRQDjVhI4R9H86VYQnx1hgOnXOlg9fmNYXwMSPBWuf3lXM1IOyKSNBzEpNrKESE40C5IY4oGwUkxBmTuXJFON7buNTNdAJeMAeRljRIpXIHcBA01mcqWirJ+wIFe87+XKWep2Vn9d6uZTGE5GJabAUDxA1g1GA+arnADkdTjZNiDxpYkoqAqSDU6Ry7PGFHaKxnm9niTN5i1+9mZy3Bn5YbjfxK3TteM3B6DUn5KSgD+f9x1BjkgIvMEMV9irpByBwJ6hzxXy93Ob2BSzT0OmNo0S6iEh8zMnircmK9oHuc3Kvn9VbP5O8SxqP7nf0Yuyo6qZwkJ/Bt+DeOnu7uoWBwjMX8mpU2QyhPQHgBacu+kbHmZ1UvViCjFx0iw27etyag9I6fpyGBgzhK+zWIaPbMOL6sQp5lXiGWKFUHU5gdAkw8yCg5T4FDOkp2HCikFQ9ULj/Vz/I1ePs9T+80D0vi0jhtItahO8bK9aZyv6Kc4j6tKdodyvK3aZykzKPcDmGwMxUm0M+QjTBdOQkNywpR7BZcKpzjuw72i1nI91KsG20maK3Nfbje53kxUb5qrYlq6kKaCxLXTDlbQPeAJTw6hE2prAu9H2ULqgqC6AZ52HTJsQIZTPdpE8CnEdQPhJhHp46mYvcsG7nJ6+DUQbIpkCDmRDTOIkDfzGsDNSeUoL3IbFNeTe7zg1Byo4uMZr3b5GhIF+iD1DX6c4RLcRuaLKUE7BWV4q43+gD4ZFtp3KagNAY8MohATMhESUzYtvLiYC8mwq04oL4qWzlSzkSxOvkAufq70x80YC5u41DcgAS2VQE5O3w7SwSNIxo3oLEyBOGI2ohIeL3V42Cx0QCSO5j2CWoxReFkjQG4KI9DuwbFcQeGzGfTmAaU+GEtzBm0g5v/jw1qkq9RyKdSngEWB09Zti/HgVkgO1ZuisNAopsZdXrg89UIEITgRvRaf5MpzeHK90nBed7vuVhdcvy1fjQ+sH6n1zHKsVeu1tWntWUeWb/0z9+Nce+6n1tPWZutd630BfTBX2nxv1Vbr9D+CBzrw</latexit>

The function fCPWL : RN → R is continuous piecewise-linear with partition P1, . . . , PK

(i) : fCPWL is continuous RN → R (ii) : for x ∈ Pk : fCPWL(x) = fk(x)

M

= aT

k x + bk with ak ∈ RN, bk ∈ R, k = 1, . . . , K

<latexit sha1_base64="+mpicMHbXSvxbJYFD/VzXZ9P+98=">ANsHiclVdtb9s2EHabrWu9ZWu3j/tyWBwgaR3DSteXBQtQIG3RomnpWlatEoMSqJsxtRLSqxS+i37Hftn+zjpTlSHI7ZAISy3fP3XO8Ox5pL+VMqn7/7ytXV76+to312+0v/1u9fsfbt768UgmfDpGz/hiXjnEUk5i+kbxRSn71JBSeRx+tab7Bn92zMqJEviQzVL6XFERjELmU8Uioa3rv51OKYQZrFvkMnHGpXRLA3eLuf74B7cPIKXJXgSweYBO16IfhJrFicJZmElFGfnjNJtw/ETmcMzWGlAgMxfobDJ2uGyRKdgfDF5269ERi7VZWK5dG76OspGgsxzAFRmnut97cI9O8/Izb6+Dxckd8DjxJ10QNOgC2tC4C7gcGivSBX9G4oV0RjlPzlGbjRIe2E+kQE+GeAvDIxFVmBXY8Me7JMoTYTCT0WnkPbAce5utjWuO5U03d1GoctpqCIiMHj8HoHLFI2Mst+7Z9SCjcalvu9HedsCoN3ZYJuYRoBqXm0qK1nsVNM8t0RDY9mBMBEIOIKpy2IYDCc7NVcbRrO5Gw4nxZtLPwY0BPclkJPD4QSMEO6AN5x0itJ0jMo0Juh7Rpd8aXThc5ktywXFmt9HobMJlyMpNqxmlJjxIseb6B7N6wbzkcy+XKLBpqdJSfvLBUG9MtV5FsONkc3imjQJMuGiBxwbuBwE37eRFp9cGl8ZB0SXDm2v9Xt8+sPzizF/WvNnMLx14x9cgZ9F2Bc+J1J+cPqpOtamJ31OsTQZFg4bCTtH287KYR1FgU1iGUp+q2GI5GUs8hDZETUWDZ1RvhZnZEIGco6q+dh1QlHzIVPjzWLE4zRWO/CjMOGBPmF0LARPUV3wGxPdxXRl2agA1D4fOsTbBG/OaQvqE02C39v2sR5TfkYV6gWN6bmfRBHBPLshiRifYeOQjJt9KcPyvb2Om8ZOEQkIBS/jnCq5g3IXMdgBiz2MGy3XYuTl2umavdvt50sYs8LDO6b7vwPYdVgPJ7RgYjqcfJiYeDnuSLYaHQef1ydFejI6GZ0N/ec/lWLqkb1z+/Jt0nVJ3bs/Z+UIPy/fNc9qDGVmSepPQByjTN+LF36G291e/dR5N6prHXOQczyQ2N/JgRQTsXJtuGA75E8kWeBxc8i7zjeJx9mehuSVSlKQjyD7gX3Auy2/pz3eKYQ8b8m/s0KbIZArMBYM35vPsGw2VJqizF+VaSNIvNRqachgG0qa/AgXGJ8E0W5+FD2/qFumM89aqOLxErhmrKiQ5tMuwseJ1SnxEOZhomvBgEVRYWH+Tanidm9HmePmgukMXv64j3TcQedkep3GsqDyrKJd/vK8olt68qyldN5TPGPSpUCcGZqZ81MfZsNYVT2JCcu2e4aEiGIlHeK7o3dwM9arBfpLm2p6W+01fjxeqx03V84XqeVMVsFjNdcES21O8EzIqkt4utQF3idj6ULigrCJ3PDqkOmDch0qUebCLGEOG8g3CSiI7IUszdp4CbLw6+B4EtOpgT1TDCEUzrwGenEzN8iK3QXE8uUdrTo1AF5unfNbhDc4n8F+nXqW3yZ4Tl4HeyOmYNeFi7DC230O/bJtND+kaLaOhCRiShdCastmhKenC+HpQvg8Dpiv560caWehGIjEK+TSF3pQen6aoLl7zgI6Jkpb1MLkyTQtLJK09Kif4ASIE84ipvDi0l7fsg+8pgpH8ghwluMNf0RBsk8UqPJ7+LRdSfHRzxSY5z2dKqQDu9E2vnV3oYrSjOHYlMKvARYXf2kKH/MSCXQtfHQ24UhEvrxwfZqCiI1wR3hWd5s1w+eVou+fc7fX/3F579HB+a7ze+rn1S2uj5bQetB61nrUGrTctf+XaSnfl3sr91e3Vd6vDVJAr16Z2/zUqj2rp/8ClTzq1g=</latexit>

Multidimensional generalization

<latexit sha1_base64="0mD4IxcZpxSnhJTR52bQKdFTdAU=">AKsXiclVbtw2EFXSW6LenPaxL0SdBdogWUhrtG4DBAjgJLCBHDt2EljGS5Fze7Spi4hKXvXhL6hX9PX9jv6Nx1K2rUudrEhsAth5syZ4cxwyDATXGnP+/fW7Y8+/uTz+7cdT/4suvl67982hSnPJ4IClIpVvQ6pA8AQONcC3mYSaBwKeBOebVn9m3OQiqfJaz3P4Dimk4SPOaMaRSdrPwYqDxUw8yoXmkc8hsRiqSATSEBSwS9LZOGerK17Q69cpP/h1x/rTr12T+7dzYIoZTlSaiaoUke+l+ljQ6XmTEDhBrmCjLIzOgFT7qQgAxRFZJxK/CWalNIWjsZKzeMQkTHVU9XVWeG1OiuRaqzaXsMwLtym5CjX41+ODU+yXEPCqoDGuSA6JTZ9JOISmBZzQhnDfeVUY7wthtf+sbHBW/OWQjEqIHoy/HV0bKYgzkGjXmKWL1gaxzSJTDCmMRfzCMYUq1GYQI0X3+7AHZCynIoglIS5EKDVY5QHiMHql7kycT5JRVQYOQkL4z/0hps/PfSKHkbCvMZ4QwRUP4Q1gwlFDpUbjKQdp6AhdpyGmKOqIrXogQyF2CsV5gV7uKj6DBb96sz1+Guyo3b/yBum64VuUOBpfyQlCD8/7jbDHoK1Vksz5shJpgq7B14YB5w5/RpJ1p7HUhyH23ikq9z6mE+1cmI+uD3OTkRj+bV36Wedc8md/saGPhqOmclAc4VkIrpw9MNd1i48E5V/NaVNUZojYA0DW/evpOx5WdL0UhZi6aRbD6x5bQeiLH1lTgwVgjfZrEOn7glL2kTZ1Q2iVeIFUO15UTCMhnlLNjPgHEc03YapqIaBE0vPNkr7D8J7OgLQ7PX3SBP3rUR7qILeyOhXKrq9xrKHvc7xrKHu02FyFIvYDgWDTbXQy8x4la6Ox5yQIE5zjvSE5TSZ4dZgnhZ3bTYOXaVZgNmVMXna5ni1Vz7qnaVqp6uKeKJrXdTz9gLvXw6yuYUXvUKHl9YeqxNVRSKXvR6bdRCzHuKigwjSGCa0F1B41sGd9YdXByF6JDPcsLZCzH1wuO638Kbq5cUakHM8b3gdVej9jJX4RTZq8ICUDVyDXlWU4+u1AJW5jA2UubTC09Ol8HQp3EkizkzdX7Hxl4pdmYaVXDFpdovq4sNqoXlwSOYUm1K1NLk+SyrLNJswWie48lLUsFjrvHB4A4elYvsg8ZROCE4Q0mGNzpR/BIaDbE5QYK8PWVTPQUpyzMNLrDt4jxRwzfGE2lPf9JBLa7vFLXntCL15zSEqktw3DUKaWmYdEY23aQIRHuCN9ofvdF1v84HA39jaH32j96eP6tXbH+c753vnB8Z1N56mz7ew6Bw5z/nT+cv52/nE3N/dP9ywgt6+Vdt867SWe/Yf3oA=</latexit>

Algebra of CPWL functions

30

  • any linear combination of (vector-valued) CPWL functions RN ! RN 0

is CPWL, and,

  • the composition f 2 f1 of any two CPWL functions with compatible

domain and range—i.e., f2 : RN1 ! RN2 and f1 : RN0 ! RN1—is CPWL RN0 ! RN2.

<latexit sha1_base64="JP1zucoK4KqXcGi4s/LFrTWHdk=">AMZ3iclVbdb9s2ELe7r1bN1mYDhgF74RYb64rYsFxs6YoVKJA2SIA0y5ImKRJlBiVRNhPqoySV2CH0p+0P2fOeBmzve9yRkh1JTgZXgAPm7sfHe+Ox3MTRoXs9f5o3vngw48+/uTuPev+0qefPXi4/PmhiFPukQMvZjF/62JBGI3IgaSkbcJzh0GTlyz9e1/uiCcEHj6I2cJOQ0xMOIBtTDEkSD5eah45IhjRSVJKRXJLMcvUI4miDNiTny4tClkcGjOECPLognY965wCwl/vdofdoGwVp5GmAQC1n7cd5MgYwULtfJe1EBUGtAqk/ioqDMgRsYA5iQU1zC3nNQoGfeR4lHvI/GO3kDaoXZGXcd3QJZUj7VsCnsFpkR+HmEbaBuI4GpJOp0O7pLs6ZX6WOzSws2vBn1wT+wcpA9BfXKIDtrabL8FOaABlFh6VoOifzrKA4ervS6PfOh+YVdLFYaxbc7WL73r+PHXhqSHoMC3Fi9xJ5qjCX1GM6L6kgCfbO8ZAok/cMtUHkoyDm8IskMtIKDodCTEIXkCGWI1HXaeGNOi3hIhBVq64bZlZcpLK4OmpolGShJ5uUNByhBERhcb8imHYmEThD0PzpViCf5WGN7Yp0o7r7dXFMLDjPjPuz/1T9WIsAsiQc9JRC4h5yHkTDkBDimb+CTAKZOZckQwXVtq41M8QuTXjdljEjxDOQOYKCuTaxUmA5j5meKD91M2au97toPq71sDsPJpMD0ugDIfwArO+PCbcjNgCdVPxl24X7qygBVTqrRUGI8ZURpq2ScWdNFVmPW5hdnLtxdlBuO/17cOlwLcrsMUvk+IQH4/3FXGaB/iNQVxHSDTCHljATUDnmsOr3uj7ClGmodcZQy8q9Eu9SzEnrektf20C3GbnVztq1nVncJYV+dauhJ1NDZTO5gewE7oJzbeyxuqlabCAwfwpOHSITIaQvAFqxb6avWVjUSNmKScTMSD3ZdKjTqS0gpfPLoWEs4L6OYuE+sgwvqhInmJeJF/AVXNXpBEITDNML9hPiUcyQ7oYxyxtB2QqN9jL9Fzm69bmu2qsfkEbHVcRxHbEO1TFVrteVeyXlHPdxSTlHu1NS7tSVm5S5hMspBHqm2qxjyDtoZzpxEgqSE6acC3hUOIVHEt4V9TzTb28YTtOMg1D9F2nevlTPWyrtqaqbqKp9GstD5c9Y2YJShJePsDFXBe6V3g+p8/MoivgaVch4xpkPFejdQSfQ1zWE4ckiGe89k9r+HO5tfDcHmSMYQE1nbBh7yelx9eDmpnGR5bP38eXIOV+yKAZVfnunXRvmYJnL0fuIZ/DTCBbiNzI0pQK9zyuBaG/4MdTLOtb8koDYEPDSIXEzITEhM2rTw7GwmPJsJtyKfeqo5VDZM8Uuj91cLjyudqfMGzFsdy6pT0ZYKoOabXk1TvIdcTJlVK+gA0QxoyGVMLhY7Y750D6R0JKHeuBECUwWSMCAhoj0uvBZjiAwM0dDOYJuT8YSzMFMpGzbg1mnrNR9KNKpgCHA6KovxXQGF5IDtWbo9muJlNjNSs+HbqhABCeCWdGuT4bzi8N+137S7f3aX3nxtJga7za+bnzbeNSwG2uNF43Nxm7joOE1f2/+2fy7+c/9v5YeLH259FUOvdMs9nzRqHxL3/wHb+F/dA=</latexit>
  • The max (resp. min) pooling of two (or more) CPWL functions is CPWL.
<latexit sha1_base64="XRfy2U80Qh4eWTm4IqJsVPK6Cg=">ALeXiclVZtT9xGEHbSt8S0DWk/9su25CSCktMZ1JWjRSJBIFEU8prBEZobY/vFtYv2V3DHSv/xn7uD6lUqVLbWdt3+AUqstKdVjPjOemZ0dL+VMqsHgj3v3P/r4k08/e/DQnv8iy8fzT/+6kAmfBh3094It5VAJnMewrpji8SwXQyONw6J2vGf3hBQjJknhPTVI4iegwZiHzqULR6TxzPRiyWDMFEbuC3HbNjuyNgER0TBYFyLRPIhY/JWmSoJUhSUKiLhOymAgSJQKekrXtwy0SZrFvKCVhspD0bRfi4Jr4dH5h0B8Ui3Q3TrVZsKq1fr4V9ukPhZBLHyOZXy2Bmk6kRToZjPjauZhJT653QIuohETnoCkiIroVJrEghbeBoJOUk8hAZUTWSbZ0R3qgzEiFD2bTqeVFu1yXHmQpfnGgWp5mC2C8dCjNOVEJM+EnABPiKTwj1fyujCr0t8Gw5xo47w53lBIn3IXvZ/XD7RI+AXoFAvIZLP4kirF2QxoxPgkgpBlXuXZlON3bPbtHinKQBKHEyzgHJX9CuYsYrJ4iVjrKhgkPci2GXq6dZ4P+6vfPBnkHI2BSYQZ9BJQ/hNWd8XgGpRn0pOknpx5WrKkMVJWkBp0TV2QctLEK49yebvIWszF/d+bK3bty4+d/ELcJ1x25PY6p/JCQIPz/uJsMagQy8yQU1zDXRLsjibUDS/r5oP8DHmlGmudc/LELr2S7zMq4Mn1kWVjg9xm5FY7q9d2ZnFXLJ7cbmhlaqhupjSQH+NdcK+NLembqsVBguKv4jQhKiJEzAUgC87N9C0LdzVSt1IkYmaknWw2NOk0Fog2+RXYMO7gvoli5T6xC17SJE6pqBPfwVd01aQTCYtgFL1gNwWfU5MN0x42QjqVli8k5t/4prW53l6p/2BLD5qIo7aiDWsjqlyra3cqSk73Ec1ZYf2bU35tq3cYNwDoaYQ7Jl6o42B9jOTOIUFqQArt0LfFQEo/EQ3xX9MjdNvX5gK0lzDLWIyFab6/VM9bqt2pypNtuqgMWq0gUda+v4uDMQ9U9Y71SBd2XOY+qCMoPkCnl6Tci4BRl3arSNEB3EZQvhJhEMacdn7yFO+82vxaCd0jGBPVOoYeinZcA3w5mZrkZWyD8nlyDxachgFdXp7p6pELvOD4/pXo3dQv8NMIV+AeKW5MBfqlpAyvtdHPWCfjUvtriuqCQEQFohQDzIRQpM0Iz85mwrOZcDMOmK+rUo60M1Nsi8Qr5dIXenvKvJ7gcfeSBTCiSheo2ZE347Q8kaRTRv0GO0CMQ1vEFA4udu95scguKGUGOezlJMXJgkgc0Agov4/LdiXgFBkP1Qi7PYwVmsOZSDuOj7NOXWn6UGxSgUNAoWu+FNOpVCqB1Iahv9xKpKJeXns+TENFIvwinBWd9mTY3Rws952V/uC35YVXL6qp8YH1jfWdtWg51qr1ytqwtq19y7d+t/60/rH+tf+e+3ZucW6phN6/V5352mqsuZX/ABTmM3M=</latexit>

Sketch of proof: The continuity property is preserved through composition. The composition of two affine transforms is an affine transform, including the scenari where the domain is partitioned.

<latexit sha1_base64="o/kE70JpBH/1P3ExOKgsJ2iD1Us=">AL3HiclVZLb9tGEKbSVyK1jdMe1nUEVAEjiAqaJ0GDRDAiWEDaer6lcCmayzJobT28pHdpS15wVtvRa/9Vf0D/Sc9dpakZD7sQlAwmLm2+GM7Oz4yacSTUc/tO589Hn3z62d173d7nX3x5f+XBV4cyToUHB17MY/HOpRI4i+BAMcXhXSKAhi6Ht+75htG/vQAhWRztq1kCJyEdRyxgHlUoOl35WztuQBwIyd45KG9C4oAkIo6D7BnZnwDx4kixKGVqZsQJCNwiXuQIC7AJ2oi4nQ8QWCYxJIZ1kG3OLkQGFJ1GRMaBOgmUYJGMohFKA0VjVryNcIij6c+i8bID0R6EFHByOUEBOQSPw4pi3JPqFC5EfAHpyurw8EwX6S9scvNqlWundMH9/51/NhLQ4iUx6mUx/YwUSfacHocsq6TSkiod07HoPNgZ6SPIp+gm/iLFMmlNRwNpZyFLiJDqiayqTPCG3VGImQg61ZdN8y6VclxqoKnJ5pFSaog8gqHgpQTFROTYeIzAZ7iM0I9jJtKqUJ/awz79ok2zpvjNYX0KAf/+eDH0YmeAL8AhXoBEVxiNkMa+doJaMj4zIeAplxl2pHBfN/td/skrziTVJ+4Keg5DOUO4jBDOex0mE6jrmfaTF2M2vDQfr368NsxZGwKzEDAcIKH4Iqzrj8hQKM+hJ3U9OXbwUCkKGqoLUoDPiJSDNlZhmnXnm6zBbMwvz1y6uyw3fv4HcZtwLcntckzlh4QE4f/HXWfAqydTV2J54YXLNHORGLtwCP9eDj4AY/UI421zjl52C28ku9TKuDh9ZGRsUFuM3KrnfVrO4u4Y4+a3W7oydxQ1UxhIDvGu+BcG3ukb6oWGwnyv5LThCiPEDEXgKzaN9M3LCxrpGolT8TCSDPZbGzSaSwQbfIrsGEs4b6JYuk+6ea8pE6MTbVKvISv6KpJxLmwch7wV4CHqOcmG4Y86IRVK2waDcz/8Qxrc919W7zA1l0VEcNREbWB1z5UZTuVtRtriPKsoW7ZuK8k1TucW4iw/hHI9U281MfAe25lJnMKCFMC1c4GPimA0GuO7op9npqlXD7yOkwxDLULyusn1cqF62VRtL1TbTRW+nqrU+S1rmzg/MBDVT9hsVYF7Zc5j6vwig+QKefp1yLQBmbZqtIkQLcRlA+HEIYxpy2f3vIE7bze/BoK3SKYE9U4h6KZlx9fDlx6MmK2PrF8+Qcrto1A7q4PVJxd4wfH9K9B7iZfj5xEuwX2S35gS9HNBGVxrw5+wTqaF9pcE1TmBCHNEIQZYCFPmxGenS2EZwvhduQzT5elHGp7odgRsVvIpSf0zpx5E8cv7VwyHyZU6Ry1OPJqmhQn4mTOqF9hB4hizkKmcHDp9h/ni+yBUvPpLcHJgkh2BQHzAGuriMB9VorCbY7WGq0BzORNq2PZx1qkrThyKTChwCcl39pZgPvlIJM7siw2DUSKSiblZ5PkxDRSL8IpwV7eZk2N4cjgb2k8Hw19Hqi6fl1HjX+sb61vrOsq164W1Ze1YB5bXGXedWjH7f3W+73R+/PAnqnU5752qt3l/AfhxWa4=</latexit>
slide-16
SLIDE 16

Implication for deep ReLU neural networks

31

Each scalar neuron activation, σn,`(x) = ReLU(x), is CPWL.

<latexit sha1_base64="W0odNsF0eSdSZ8xNA1Mf6Gtc20w=">AMlXiclVbdb9RGEL+kX+Cm5aMPfaAPqyYnBQin86EWihoJNRCBFCANJEHgKFrbY9+S9Qe763CXlR/7R/Z/6HtfO7v2XWwfVMHSndYzv/3NeGZndvycM6mGw7+Xlr/48quv7l02fl25bvr1y9dv1AZoUIYD/IeCZe+1QCZynsK6Y4vM4F0MTncOifbBn94SkIybL0lZrmcJTQOGURC6hC0fG1pb8H2KWamOr1J5l1EkRC5iWhHi4KBdmJT2r3QI0Z7Kcgn5gmxOMQqYQKZNgcDe5CQjymIDHqoVELFo9n+mGQlE7f6skfM3SkKx5wPmxu/aA5IKlAcs5hITmuchoMCZRJojMIWDRlKUxERAXnAp2Zl0nUZEGZkG5U3E+NntkQBFDUigEYigiTi18A21JFif0WKcbxmq5Prm5qT2RkD3Y2TdvaxuESbK1e7gzcPrVd5BXYyDAIYFUTGd20SyA1LzGTLiBUwEBIWRfV/HFdKve8/IYQ1AwW189e3rzePba4HaViFnTjHV1eHg6F9yOLCrRervfrZPb52+V8vzILCeBZwKuVbd5irI02FYgGH0vEKTAINTmgM2ia1JH0UhTaqUZYqYqUtHE2knCY+IhOqxrKrM8KP6oxEyEi2rfo+5rspeVuo6P6RZmleKEiDyqGo4ERlxJxNEjIBgeJTzFuA31VQhf62GF65R9o4b7a3FCbvEG4Ofhsd6THwU1CoF5DChyBLEoqR9iKaMD4NIaIFN0dRrO108eE21qRBKHELzgHJR+g3EMlta8LDIelrEPtbDxnBw75eNYbmAMaVTYDBFQ/hDWd8XkBlRn0pO0npz6WM54+Nq9Hgy7rYjRWsRSd2aLsMBvzF2eVfoFufHzP4vbhOuC3D7HVH5OSBD+f9xtBjUGWfgSbPmWGrvYWOLZgVv6znDwK25pRxrPOhYt1qg1I98XVMDa+ZaRsUE+ZeSTdu6d25nHXbF0+mlDd2eGmYqA+VbrAXv3Ngt/bHTMm/dNacJkY0QMQVAVt2P03csXNRI04pNxNxIN9ksNuk0Fog2+RXYMC7gvoli7T5xLC9pE+d4QTSIL+ArumrSiYQ2GLYXvDS3DuXEdMOMV42gaYWle6X5J5pfb6v97ofyNI3bcSbLmILT8dMudV7jWUC9xvGsoF2ucN5fOu8gnjPg1g2DP1E+6GHiP7cwkTuGBFMC1d4qXimA0jfFe0ZulaerNDTtZXmp7k+50uR7NVY+6qdz1dOuKmSpqnXhgrVtnHwYiOYnbC+cAv/M7MfUhVUGyVlpRo8WZNKBTBbOaBchFhAfOgvSyCmCz7Jx3cyWLz6yD4AskEY6I629BD0Y1riDcnU9Oyim1YXU/ewarbMqCr4pk9fXKBY73X4V+mQcWP4twDe4TWzE16FlFGZ1rk9/xnEwq7Ysc1ZAJBZRiQHmQrBpM8J37+bCd3Ph0zRkga6PcqLduWJXZH4l4HQuzPm7Qy3ex9YCGOqtEXNtzye5NWOLJ8x6sfYAdKMs4QpHFyc/h37kJeglJk2sZeTHCcLItkZzoAqGODjeBJwxE5jNcZuDxOF5nAm0q5rZtum0vSh1KQChwCra98Us5FdKoHUhmEw6iRSUb9sXB+moSJRSXBUdLuD4eLiYDRw7w6Gf45WH96vh8ZLvRu9n3vrPbd3r/ew96S329vBUv/LF9fvrH808qPK5srj1a2K+jyUr3nh17rWXnxH3yHleA=</latexit>

The whole feedforward network f deep : RN0 → RNL is CPWL

<latexit sha1_base64="hkATh4f7LI4PBo63jtxKg7s4a5E=">AM3iclVZLb9tGEJbSV0L3EbfHXha1BaSBI1AyWqdBDQRwbNiA47p+BjZdYUmOpLWXj+wubckL/oH+ml56aP9J0VvRa8+9dpYUZKyC4eADHq+b78ZzszOrhtzJpVt/9F8N7H3z40cNH1tzHn3z62eP5z49klAgPDr2IR+KNSyVwFsKhYorDm1gADVwOx+7FmsGPL0FIFoUHahzDWUAHIeszjyo09eabC4LAxZq4yvVTqaog2QgYJwS4oiEg+7AKM3+pBYh2lFRLCFe7cKIOBz6KqACFVa7WUIiMUBAa2DSzYFjgthekVivDyTr1hkR6lFNBQkhEFBLqKXaZRbVEFh3JBgHt6XDJAc7TJ6OvV7UjArIH24fmv8UlwiRZ2z3eblu5S3IwBHI1jDiQPoDfj8QVFT6Kq6tIXBCUfE36vUzEB4jTF8TZ+0nv9OwUv2fyvp0uFrKWA6GfJ4VYvcLdtvOHjL70pm8LDQmz25v/tG/jh95SQCh8jiV8rRjx+pMU6GYxyG1nARTRL0LOgCdpTwlLT5BOPGX6hIZq3waCDlOHCRGVA1lHXMG/FjEXIvqx6dV2sRtlymqj+8zPNwjhREHp5QP2E0yP6RziMwGe4mOslIflVCF8VYUDjpn2gRvlcAU2nwV9vfdc/0EPglKMQFhHDlRUFAMdNOnwaMj3o04SbRpT94t1qWS2SdbIkSCVuwjko+QLtDnKw8adNG3E/1WLgYrcu2e2Vb5bsdIZjGjvn2G0k5D+klYNxeQK5G4ykGienLm42bDg23S2GnU62ivGKG8UqXtKasnF/f+ViH95TGz/nbRNu6p7XIs5bukBOn/p1VUEOQiSuxvXD7pxpnzFBi78BT/cxuf4tLqpnGXuecLFp5VPJtQgUs3izpGh/kLid3+lm58TPNu2Lh+G5Hy4WjspvcQXqKe8G5cfZU39Yt08E60TQpyjJEzAYgC53b5Wse7uk7CUrxNRJvdhsYMpPBt6itwYNwjfJPFSfjEynRJVTimoix8j1gxVFNOFMySkc2C/Rg8Rjkx0zDi+SAoe2HhXmr+EseMPtfVe/UPZOFJlXFSZ6xhdxTgWh3cK4Ez2iclcEZ2pwTu1MFNxl0QqDgzNSbdQ68xXFmCqewIQVw7VzioSIYDQd4rujV1Az18oLtKE51duxt17VeTaFXdWhrCm3VIZ+FaoL5M9428F7CQJQ/YWOmC9xrsx5L5+cVJNepuRhUKMaZTo3WGmGFc1RhOFMCAzsTsXtR4F7PDr8bgMyIjzImqLcMIRT2vPp6cTI3TPLd+fjw5RwudigOdb57iaZFL3OB4/uXs/djL+EWGJ+QWyXbMhPQ6l+zfoMH32CejHP0hRjgTEHGyM0AUyNkZTPG8/Op8Xxq3Ap95ulJKwe6MwV2ReTmdukJvVsob0S43LliPgyp0hlrumR9FOcrorhQ1Os4AcKIs4ApvLhYrWfZQ/ZB4UgeEJzlJMabBZHsGgor42P5UjAC3A4UEOc9jBS6A7vRLrTMTfPMmjmUGhKgZeADKueFMWFWiqB0kah3a0VUlE3LR0fZqCiUErwqtipXwxnX467c5y2/6xu/Dy+eTS+LDxZeOrxpNGp7HSeNnYbOw2Dhte8+fmL83fmr/P/Tr359xfc3/n1AfNyZovGpVn7p/ALOgd0I=</latexit>

This holds true as well for deep architectures that involve Max pooling for dimension reduction

<latexit sha1_base64="tXFumNqxIhD9ErkmPHsWfoNsiE=">AMRHiclVbNbtGEJbSv4Rum6Q9jKoLSANHEFS2jotaiCA4yAGnNR14iRIaBhLciRtvPzJ7tKWvOCT9Gl6aF9gr5Db0XRW9HZpSTVFI4BCzQM9+M5y/nSATXOle74/2pfe/+Djy5f8VY+/uTq9euf/ZUpbkM8SBMRSqfB0yh4AkeaK4FPs8ksjgQ+Cw43rL6ZycoFU+TJ3qa4WHMRgkf8pBpEh1db3/tBzjibG2CuM7RhPnI4nTAsCXuUDTx0nhfgoPwPg6zRmwOcgC9wqGMmiWFz0L2NMfhcY2zVPauWfDSe63thXHgdp4dtFo5BhUwCQnmMk2AhZqfOK/WYc1XfBSzI5Os+yhEcWPy1abxZQz7uHtg/1tbB65ga+/ZbtcrTcKTMUnGqYgUaJkjMAWndBaGqYQIMQMmwzEhQ51LJMyYaeDJSpOEB6yCWRpSkEclXgeY2JjBhKjPLROeT4mURkl8I6urfa6PfA8kt/9rLamj17R9ev/ONHaZgTrQ4FU+plv5fpQ8Ok5qHAwvNzihkLj9kIjctBAR0SRc6dYZpocNIajsVKTeOAkDHTY9XUWeEbdVYi1VDVrQYBpacqeZnr4Z1Dw5Ms15iEpUPDXIBOwZYSRUlSNMWUhfSd+VMk781hif9Q2Odt8drCpt6jDa73w0OzRgpBZr0EhM8DdM4ZhRpf8hiLqYRDlkubGWq4fzd63gdcKWtgKAQ5EKgVt+T3CcMdcKiqkcCiNHAZXveq+78c16r1jC2EovMb0uAco/glWdCUSOpRnypO6nYAF1H1UgX7SPRez3rFWqXO8+UvRYLbmL848b8wLctPnvxO3DdcFuQNBqXyXkBD8/7jrDHqMKg8UutYrDA2dsaLawZvmVq/7LR2pR5pqnTp9zSu9Uq9zJnHt/MjA2oC3GXmrnY1zO4u4a5M327o9txQ1UxpoHhJveCfG7tp3lQti0k747QhchEC2wCw2n8zfcPCRY1UrbhELIw0k81HNp3WAhibX0kD4wLu2yjO3AfP8UKdOGOySnwBX8lVm04idMFws+BxhiFnAuw0TEU5CKpWeLJf2F/w7egLArPf/ECevKgjXjQRW1Qdc+VWU7lfUS5xv6gol2gfVZSPmsoHXAQo9RxCM9M8aGLwNY0zmzhNBSlRGP+ELhXJWTKie8VsFnaoVw/splh3GW62+S6t1Dda6p2FqdpiriZ7poiVr92lR4Sirn3B/qQqCM3s+sHe3yCcFXZTqEmDchkqUabCLmEOG0g/DTGEVvyOThu4I6Xh18DIZIJhQT3ThGHspmXCO6ObmeFmVso/J68p+u9msGTNk86cDJ9TgdP+V6MdZ6PDzCM/AHXAdMwM9LCmH59r4B6qTSan9MSO1I5CxQ5RixIUQXdqs8NWrhfDVQriTRDw0s1KOTX+h2JNpUMpVKM3enPl+Ssf9Ux4hLWLGoRZHtidZeSLN5oxmyZAQhtazDUtLl7nlnvgMWptza5ZDRZgGKnyGgDrv0eL5C2oiTkR7TtMeJnO0E5l+36iVaWdQ4lNBS0BTle/KeYbtKSqC1Dd9BIpGZBUbk+7EAlogJoVew3F8Pl6eDbv92t/fTYPXundnSeLn1RevL1o1Wv7XRut60NprHbTC9s/tX9q/tX9f+Xlz5W/Vv4uoZfaszOft2rPyr/AToegIo=</latexit>

fdeep(x) = (σL f L σL−1 · · · σ2 f 2 σ1 f 1) (x)

<latexit sha1_base64="lmdC59RS/rQUXJO/q36ipUv5zTY=">AL4HiclVZLbxs3EJbTVyL14bTHXog6ApIgFrRKGqdFDQRwYtiAk7pO7AT2GgJ3d1aizX2E5NqSib3VvTaX9Vr/0mPHe6upH3YhUJAjHz8ftmOcMhnZgzqfr9f1ZufLpZ59/cftOu/PlV19/s3r32yMZJcKFQzfikXjvUAmchXComOLwPhZA4fDO+d8y/jfXYCQLArfqmkMpwEdhcxnLlVoGq7+b8i/lDbIiAeQJzet4/I5MGmzcFXZm5LNgrocI/YLhMu0XYmqR1O3fMU/f5wLy18C7TeW7fmVteLlGxABjcTDpqE1s1oK7UFG43Vgzy9nB1rd/rZ4M0J1YxWsVY394986/the5SQChcjmV8sTqx+pU6GYyFt24mEGPXoCHQmn5IumjziRwJ/oSKZtYKjgZTwEFkQNVY1n3GeK3PWIT0ZVXVcYK0XbacJMp/dqpZGCcKQjcPyE84URExWSYeE+AqPiXUdfG7Eqow3grDW+tUm+DN8opDupSDt9n7aXCqx8AvQKFfQAiXbhQENPS07dOA8akHPk24SrUt/dm83W13SVZ1kiCUOAnoOTPaLcRg0WaJzBIRhH3Ui1GTqtR/3exo+P+mkDI2BaYPo9BOQ/hJWDcXgCuQxGUo2TUwcPhoKApYvKSQDrSyQctFGFSdqeTdIas5FfnrkId1lu/PyP4jbtSR3djo+ZkvMYfof7iqDGoNMHInlhQ0k1XgmxJrBx7q9X7vKS6p7jTWOufkXjuPSn5IqIB7iyUDo0FuErlRZ2OhM93xcLpzUKPZ0JlmVwgPcGzYC/EHurqsVCguyv4DRblO0QMQeArFnX09cUlhUpq2SJmIvUk81GJp1GgWiTX4ENY4nwzS4W4ZN2xkuqxDEVZeIlYsVQTqRMNuMrBe8icFlBPTDSOeN4KyCgsPUvNPbNP6HEcf1D+QhcdVxHEdsYXVMXNu1Z0HJWeD+7jkbNC+Ljlf1507jDsg1AyCPVPv1DHwAduZSZzCghTAtX2Bl4pgNBzhvaI3U9PUywv2ojV2V28V+d6MXe9qLt2567dustjoSp8XkNtG98QDET5E7YbVeBcmfWYOi/PILlCnm4VMqlBJo0arSNEA3FZQ9hRACPaiNk5r+HOm82vhuANkgnuiaotwhFfV89vDmZmqb53nr59WQfrVkVAZ0fntnokgs84Hj/5eg3sZvhZztcgLskOzEF6FVO6S+8wS9YJ5Pc+2uM7oxABkiNwPMjZClzRjPzubGs7lxN/SYq4tSDrQ1d+yLyMnt0hV6f8a8HeFy+5J5MKZKZ6j5kpeTOF8RxTNG/RI7QBhxFjCFD5d2dz0b5A0obMkjgr2cxPiyIJdAQHl9nC0bQn4WA1HaozdHiYK5fBNpK0nLr51yk7Th0KTCnwEZL7qTF7/EolkNow9Aa1RCrqpKXrwzRUJMIvwreiVX8ZNidHg571uNf/7cna82fFq/F26/vWD637Lau10Xre2mntw5b7srTFXsFVvyO0/m980fnzx6a6VY812rMjp/Qd4J1cQ</latexit>

Each layer function σ` f `(x) = (W`x + b`)+ is CPWL

<latexit sha1_base64="27IyL7CR10zDSCz5sNXRmZDehYc=">AMRHiclVZb9s2FLa7Wyv0m6PeyGWGEjb1LDdbe2GBSiQNkiBtMuSJikSBQElHdtsqEtJKrVL6Jfs1+xlD9sv2H/Y2zDsbdghJTmSnAypABvUOR+/c3RupJdwJlW/0f72nvf/DhR9dvOJ2P/n0s5u3Pt+XcSp82PNjHouXHpXAWQR7ikOLxMBNPQ4Hin60Z/cAZCsjh6oWYJHId0HLER86lC0cmt9teuB2MWaWMr065l1GE6FjDLCHFykEPYJrZv8whRLsqTiQka0OYEpfDSIVUIMPasHcfQuIyBaFR941asPGk1Pf9MHO6Vk+eUH9CpE85FSCVMQRob5iZ9arVbLsSjYO6YmOVl3gPFuZ3l7TrgjJDmztmbflVcIkWd8+2Oo5pELJ6QwEGaWRb4iQZ58UVIaHuD4TPkHhyL6v4AqZV9xn5KAoOAuvnr29fbJ3eXSjuNCFORIs7JzaV+r28fsrgYFIulVvFsn9y68Y8bxH4aQqR8TqU8GvQTdaypUMznkDluijGj/ikdg7Y5yEgXRQEZxfg9caSIldZwNJRyFnqIDKmayKbOC/UGYmQI1m36nmYnqrkKFWjh8eaRUmqIPJzh0YpJyompRIwAT4is8wdT5+V0oV+ltjeDE41sZ5s72mMKmHYK3fBYT4CfgUK9gAje+HEYUoy0O6Ih47MARjTlpjLlqFw7XadLbGlLglDipZyDkt+j3EUMdsK8imMeZFqMPSzf1X7vwTer/WwBYyo9x/R7CMh/CKs64/EUcjPoSd1PTj3sPqxANm8fg86K3jFWsXOcpE1mI35qzOXjXlFbvz8d+I24boit8cxle8SEoT/H3edQU1Ap4E28aZxqEzkVg7cEf6/e+xS31SGOtY/MuO7lX8nVKBSyfbxkaG+QyI5faeXBuZx53xaLZ5Ybul4aqZnID2RH2gntu7I6+qFrmk7bgNCGyESKmAcjS4GL6hoWrGqlasYmYG2km41NOo0Fok1+BQ6MK7hvoli4TxzLS+rECRV4iv4iq6adCKhDYadBbsJ+IxyYqZhzPNBULXCop3M/BPXjD7P0zvND2TRYR1x2ESsY3WUyvWmcqeiXOA+rCgXaJ9XlM+byk3GPRCqhODM1JtNDLzGcWYSp7AgBXDtnuGhIhiNxniu6LXMDPXqhq04ybQ9TLeaXI/nqsdN1dO56mlTFbBIFbpgwdoGXlQYiOonbCxUgfW7MfUBXkGydvM3BRqkGkDMl2o0SZCLCDeNBuHMKYLvjsnTZwp4vDr4HgCyRTjIlqbEMPRTOuAZ6cTM2yPLZBfjy5+0uDmgGdN0/5dMkZNjiefzl6N/EtvoxwAe4S2zEF6FlOTrXhj9gnUxz7Y8Jqi2BC0iFwPMhWDTZoSvXs2Fr+bCp1HAfF2UcqgHc8W2iL1cLn2ht0vmjRi3u29YABOqtEXNtzyZJvmOCkZ9ROcAFHMWcgUXlyc7j37kF1QOJLHBGc5SfBmQSR7CwSU38PHcSXgjTgaqwlOe5gqNId3Ij0YmKtoVWnmUGRSgZcAq6ufFOUNWyqB1IahN2wkUlEvqxwfZqAiUbwqjhoXgwXF/vD3uB+r/TcOnRw+LSeL31Zeur1kpr0HrQetTabG239lp+f2L+3f2r93fu382fmr83cOvdYu9nzRqj2df/8Dqb573g=</latexit>

The CPWL also remains valid for more complicated neuronal responses as long as they are CPWL; that is, linear splines.

<latexit sha1_base64="u+fXuHrC58VaCj40arp3MNCRr+M=">AMYXiclVbdb9RGEL+jX2DSFugjL6MmJ1EUTneH2tCPSEgBFJKQyCAwFG0tud8m6zXZncd7lj5/+q/0pe+9KH9D/raWfvuYvugCpaSODO/c14vnaCTHBtBoM/uhc+fSz7+4eMm7vPblV19fuXrtuU5zFeJBmIpUvQyYRsElHhuBL7MFLIkEPgiONlx+henqDRP5TMzy/AwYbHkYx4yQ6Kjq90nfoAxl9bZKqxfMtokjxXOCgBf5QLtEKdF+avwAKxv0kxjtj3CKfgCxyZhihi2R/3bmIDPDSZOPXBqxePJQj8Ik8LrlXq4z8IJ6JAJpkBirlIJLDT8tPRqEzZ8zeOEHVm56aMQxY3pd9vWVwns4+6B+29jE7iGnb0Xu32vMgnPJlgKgAmdgsKEcanhlAkewThVkKQKIUwTip9PEZzu0wQVmep1KiBaRCpjN1fM8EZMFVx/kz/MkMmNz3rB2Nw0SbPdeZedNH3fJRFULwjq6sD/qD8oHVl+H8Zb0zf/aOrl7614/SME9QmlAwrV8PB5k5tEwZHgosPD+ngLwhMVoywQV0CNR9V3jVBopQ0cS7SeJQEhE2Ymuq1zwvfqnETpsW5aDQLKXV3yOjfjO4eWyw3KMPKoXEuwKTg6gwirjA0gmIYhvRdeRnyBsOz4aF1zrvjDYWrC4y2+z+ODu0ExSka0iuU+JbSlzCKtD9mCRezCMcsF65s9Xjx7vW8HpR1T+mUEQS5EGj0TyT3CUP5WpZ4KqLCqjig2t4c9Le+3xwUKxjXBhVm0CdA9UOwujOByLEyQ540/RQsoNak8uTL3nLoYt5Yziq1lbd4KVrMzvz5mRde05u+vyP4nbhOid3ICiVHxMSgv8fd5OBWlPngabyomFRWJpIE021gzftrUH/BzrSjDTVuhCw4Ve6Tc5dfXG2ZGRswEfMvJBO1tndpZxN1zOPmzo9sJQ3UxloHhNveCfGbtp31ctyzE853QhKiMErgFgfh+paF8xqpWykTsTSTjaPXTqdBbAuv4oGxjncd1Gcuw9eyQtN4oypOvE5fCVXTqJsAxGOQueZhymvJuGqaiGgR1K1zuF+43+G70BYHdb38gl6+aiFdtxA5Vx0K501bu15Qr3K9qyhXaxzXl47byIRcBKrOA0My0D9sYfEPjzCXOUEqFNY/pUtFcSZjulfsduGev3AbpoVtrxpd9tc95aqe23Vo6XqUVsVcWnmumjF2gPaYjiq+ic8WKmC4J07T6mLqgzCu8KtEQ3ItAWZrtRoG6FWEG9bCD9NMGYrPgcnLdzJ6vBrIcQKyZRiYlrHyEPVjmtENyc3s6KbVRdT/7z9WHDgK2aZ/H04JQanO6/Cv0C0v8IsJzcA/KjpmDfq0ox2fa5Beqk2ml/S0jdUmgkhJRiRGXQizT5oTHx0vh8VL4SEY8tPNSTuxwqdhTaVDJdajs3oL5QUrH/bc8Qlq7bIlaHrk/zaoTabZgtPdpAshU8IQbWly83q3ygadoaCTHbpeDjDYL0PwdApqwT4/na6R1WcZmQtMep4bM0U5kh0O3p9aVbg5JlwpaAkpd86ZYrN/aKJ2DP1RK5GBUXt+nADlYgKoFVx2F4MV1+ej/rD2/3Bk9H63TvzpfFi53rn286NzrCz1bnbedjZ6x0wu7v3T+7f3f/ufzX2qW1K2vXKuiF7vzMN53Gs3b9P27dhlo=</latexit>

CPWL functions: further properties

32

The CPWL model has universal approximation properties (as one increases the number of regions)

<latexit sha1_base64="9GN898VMvRVgh54ZdD1NLuNpGPU=">AMpHiclVbtGEJXdW8K6bdI+9mVaW0UujiApaJMWNZDCSZCgbur6kgQJXWNJjqiNl7vM7tKWsuBf9Of6JX3oS2dJSZaopHAIyKBnzp4Zzm0nygU3tv9e2X1gw8/+viTS5eDT9c+/yLK1e/fGpUoWM8jJVQ+nEDAou8dByK/B5rpFlkcBn0cm21z87RW24kgd2nONRxlLJBzxmlkTHV1f+CiNMuXTeVunCitFlRapxXAKEuhDoejgqz9lAOBCq3KD+VYfRxAKHNiMaWLY6nduYwYht5h5derNU+HU303zsqgXenhAYuHYGImAaJhVYSWGz5aeXVJmyEhqcZO3ZyM0Qhymuj61su1Bns4c6h/29jE7iB7d1nO52gNgkHQ6wEkKkEBQyZgUJy/+1MAMtzrUY8q/iB3nPUlqMJwCuEVJBC5jipxBA5aoZJFqENQFN8lDTXg/bE0i9yXFuKmYSIDma5wAylhVPOgEGCmFe0qfZM6VP4IzbIUhFrmkdjq2EQqVHvfh1w0QbExOBiHKpM4CBMdX1rudbvXA8ktv8rLemjy7x1cv/xMmKi68F7FgxrzsdXN75Bh9ZSywDMKCcsLiE5aiq3JcQptECQyUph/5XkXcCwzZpxFhKS4DU1T54Vv1XmJNgOzaDWKP3zkpeFHdw9clzmhUZ1w4NCgFWgS9VSLjG2IoxlUZM31UwS/4uMBz0jpx3h9fUPjSwmSr82P/yA1RnKIlvUaJZ7HKMkaRDgcs42Kc4IAVwle+GUzfgzalumodAwSFqBACrfmJ5CFhqNmXaJEUjqdRtQem93One83u+USxndSjel2CFD/CDbvTCQKrM2QJ4t+ChZRd1Pd8Vl7enQ56U1vlTozmL6UDWZv/uLM08a/IDd9/ntx+3BdkDsSlMr3CQnB/497kYE63BSRofKi1i4dDbWhodrBG+5Wt/MDHVmMNW6ELAR1F6Z1wXTuHF+pO9twLuMvNPOnXM7s7hbTsPlnYZuTw3Nm6kNlC+pF8JzYzfc26plNsknD5EVYTANwCs95O37BwUSPzVqpEzIw0k81Tn05vAZzPr6aBcQH3fRQn7kNQ8cIic70PEFfCVXfTqJsApGNQv2c4w53SJ+GipRD4J5K1zulf4vhH70RZHba34gly8WES+aiG2qjqlyu6ncm1Mucb+YUy7RPplTPmkqH3FBd5ydQmhmukdNDL6mceYTZ6kgNQoXntKlojmTKd0rbqv0Q3+wI7KS1d1jtNrvsz1f2m6vFM9bipSri0E12yZO0hLUIc9fwnPFyqguiNP0+pS+oMwpvSbyILkFEDMlq0SZCLyHOGohQZiyJZ+jkwbuZHn4NRBiWREMbGNY+ShbsY1oZuT23FZxzapr6fw6XpvwYCrm2f6tOGUGlz5rcSj9/O4wk8jPAG3oeqYCei3mnJwrs1+pjoZ1drfc1JXBDqrELUYcSbEKm1e+OrVTPhqJnwsEx67SlnrjdT7GoV1XITa7c7ZX6o6Hh4xhMcMusq1OzIg1Fen1D5lNE9oAkgleAZt7S4BO1b1QP7aGkp9VKmNmAYa/QUAbd+gJQoO0cvUDmna48iSOdqJXK/nV915pZ9D0qeCloBKt3hTDd4YzVRe4ZOv5FIy6Jy7vrwA5WISqBVsdcDJdfnvY7vdud7h/9Xt3J0vjpdbXrW9b1q91p3Wvdaj1m7rsBWv/Lv6zeqN1Ztr363trO2vHdbQ1ZXJma9aC8/an/8BUh2biA=</latexit>

(Arora ICLR 2018)

Any CPWL function RN → R can be implement via a deep ReLU net- work with no more than log2(N + 1) + 1 layers

<latexit sha1_base64="fwOT4vPK+Clre0USCUO9JxD2Nzs=">AMuHiclVZtc9NGEDbpG6hpgfZjv+w09kyA4LHMUCjTtHQCDMykaRoIMKA0PUlr+cjpJO5Ojs2N/k/Un9Jv3ZPsh1bhk7QjD3S7XP7u3bZgLrk2v98+FtU8+/ezLy5e8r5c/+ry1eufvNcZ4WK8DKRKZehkyj4BIPDTcCX+YKWRoKfBGe7Dj5ixEqzTP5zExyPEpZIvmAR8zQ0vHVC38HISZcWqertEHFaNMiUTgpAQJVCLQ+jsvqr/QAbGCyXGO+3cxBAIHJmWKGLb73VuYQsANpk7c2LFk+FM3ovS0utUcnjIoiHoiAmQGKhMgksMnxUWbUF7UDzJGXHVm4FKES5Ob62bQOVwgHuHrqv9hZwDTv7L3a7XqfWX3BkI0QCsndmZkAlucqG/O04gV6z1EZjoIPOhsMg2ZROAyIpdp1GCGCLJIQ1SQDUCRYzKpr3lTBb/KSa1kUMioImwHB3/uATkEgoM2RExCSHRpLjBFaWDEGTCIEfPKcDqpOc3UCZxyMwSZQZopJ3MEYksOe5v7t3wr93w2yDYhA7gBSjOjLgHV/Z6HV71QOrL/70ZaM1faPr176N4izqHCmRIJp/drv5ebIMvJAJLD0goLixKITlqCt4l5Ch5ZiGSKfnSAanUJx1KtJ2lISPLpUDdlbvG9Mrei9EAvaw1DSonFldeFGdw9slzmhUEZ1QYNCgHkYpe+EHOFkRETSpeIzlUwQ/YuMTzj6wz3m1fErh0w3i7+2P/yA5RjNCQXKHE0yhLU0aeDgYs5WIS4AVwlWDHszevQ7lWVOGgKYSEGn2P1gPCUPXNKycTcWlVElLJbPW6d25v9coVjKuGtPrEqD+EWzRmFAUWKshS5btFCykiqec5POSdehyWq9OK1WrN3spG8xO/fmZ83gnNx0/I/idu46J3coKJQf4xKC/x/3MgMVvy5CjVpl5Ya3VBT7uB1e7PX/YG2LHuacl0IaHu1VfptwRS2z7b0nQ74kJIP6rlzpmfud8Op8XxQ0a2ZokU1tYLyNdVCcKbsun1ftsy7+5TuajyELgCgA3/fQNDedVsqilCsRcSTPYPHhdBrAuvgqahjnMN95cWo+eBUvLBPnTC0Sn8NWMtWFkwgrZ1S94GmOEacbxnXDTNSNYFELlwel+4fAtb4wtAfNA3L5ahnxqonYoeyYCXeawoMF4Qr3qwXhCu3egnCvKXzMBV1/ZgahnmkfNzH4ltqZC5yhFQobDCiS0VxJhO6V+x26Zr64obdLC9tdYHvNrkezEUPmqInc9GTpijm0kxl8Yq2RzQcVSLR3i0kgXhO7efQhfXEYR3pZtOliDjBmS8kqNhFpBnDYQZiwlZsDk8auJPV5tdAiBWSMfnENLaRharp15huTm4mZe3buL6egucb/pICWxfP7OnAiAo8c1OJQz/Nowo/8/AU3IGqYqag32rKwZk0/YnyZFxLf89JXBGotELUy4jzRazC5hbfvJkvpkvPpExj+w0lVPrzwX7KgvrdR0puz9jfpTR9uCUxzhkxlao+ZaH47zekeUzRvuQOoDMBE+5ocHF69ysHniKhlpyUk2LOU0WoPk7BDRlx4v0EhTuEzMkLo9jg2po5nI+r4bfxeFrg9JFwoaAirZ8k0xm+q1UTtGLr9RiANC8uF68M1VCIqgUZFvzkYr4873f9W93eH/2N+3enQ+PF1net71ubLb91p3W/9bi13zpsRWuX126v/bz2y/q9b/Wk3VeQ9cuTPd821p61tV/f/ugzQ=</latexit>
slide-17
SLIDE 17

Refinement: free-form activation functions

33

layers nodes (n, `) …. ….

neuron

(n − 1, `)

zn,` = σn,`

  • wT

n,`z`−1 + bn,`

  • Linear step: RN`−1 ! RN`

f ` : x 7! f `(x) = W`x + b`

Nonlinear step: RN` ! RN`

σ` : x 7! σ`(x) =

  • σn,`(x1), . . . , σN`,`(xN`)
  • Joint learning / training ?

fdeep(x) = (σL f L σL−1 · · · σ2 f 2 σ1 f 1) (x)

Layers: ` = 1, . . . , L Deep structure descriptor: (N0, N1, · · · , NL) Neuron or node index: (n, `), n = 1, · · · , N` Activation function: : R → R (ReLU)

Constraining activation functions

34

Native space for

  • M(R), D2

BV(2)(R) = {f : R ! R : kD2fkM < 1}

Second total-variation of σ : R ! R

TV(2)(σ)

M

= kD2σkM = supϕ2S(R):kϕk∞1hD2σ, ϕi

Regularization functional

Should not penalize simple solutions (e.g., identity or linear scaling) Should impose diffentiability (for DNN to be trainable via backpropagation) Should favor simplest CPWL solutions; i.e., with “sparse 2nd derivatives”

slide-18
SLIDE 18

Representer theorem for deep neural networks

35

(U. JMLR 2019)

Theorem (TV(2)-optimality of deep spline networks) neural network f : RN0 ! RNL with deep structure (N0, N1, . . . , NL)

x 7! f(x) = (L `L L−1 · · · `2 1 `1) (x)

normalized linear transformations `` : RN`−1 ! RN`, x 7! U`x with weights

U` = [u1,` · · · uN`,`]T 2 RN`×N`−1 such that kun,`k = 1

free-form activations ` =

  • σ1,`, . . . , σN`,`
  • : RN` ! RN` with σ1,`, . . . , σN`,` 2 BV(2)(R)

Given a series data points (xm, ym) m = 1, . . . , M, we then define the training problem

arg min

(U`),(n,`∈BV(2)(R))

M X

m=1

E

  • ym, f(xm)

N

X

`=1

R`(U`) + λ

L

X

`=1 N`

X

n=1

TV(2)(σn,`) !

(1)

E : RNL ⇥ RNL ! R+: arbitrary convex error function R` : RN`×N`−1 ! R+: convex cost

If solution of (1) exists, then it is achieved by a deep spline network with activations of the form

σn,`(x) = b1,n,` + b2,n,`x +

Kn,`

X

k=1

ak,n,`(x − τk,n,`)+,

with adaptive parameters Kn,` ≤ M − 2, τ1,n,`, . . . , τKn,`,n,` ∈ R, and b1,n,`, b2,n,`, a1,n,`, . . . , aKn,`,n,` ∈ R.

Outcome of representer theorem

36

Link with `1 minimization techniques

TV(2){σn,`} =

Kn,`

X

k=1

|ak,n,`| = kan,`k1

Each neuron

  • fixed index (n, `)
  • is characterized by
  • its number 0 ≤ Kn,` of knots (ideally, much smaller than M);
  • the location {⌧k = ⌧k,n,`}Kn,`

k=1 of these knots (ReLU biases);

  • the expansion coefficients bn,` = (b1,n,`, b2,n,`) ∈ R2,

an,` = (a1,n,`, . . . , aK,n,`) ∈ RK.

These parameters (including the number of knots) are data-dependent and adjusted automatically during training.

<latexit sha1_base64="07zaAV79momeL2YeyW4oNDsgW3A=">ANLXiclVdb9s2FHa9W+tlW7M9DgOIxQaSwjUsF1u7S4ACvaBFL8vSJi0apQElHduMKUolqdQuobf9mr3sYfsrexgw7HU/Ya871MWR5LZIBdiheb7zfUfnkIeMF3Om9HD417n2e+9/8OFH5y90Pl75NPLq5/vq+iRPqw50c8k89qoAzAXuaQ5PYwk09Dg8WY3rP3JCUjFIvFYL2I4DOlEsDHzqcapo/X2V7eoPyUCEhkJ0nU9NtnsjtkcAsJEAHPS3R94gLnW9a21SVMEX9KJfU1SPYKcd6i03E9mDBhmIYQ59KOa0eEaUVEnogSXfociD3jozoW7K0S6IxmYkIEZsAMr5ok/CBENRIf5ADz2lGNCD7tYPBZueAuFRHjhGalxNk6PZdvbHzPoFsZvij20nfW5OxXI19FdQau7C/T3iMcycqgnAPKbCZov4EYwxTwE4rvuA+It+bY3ceyUin38MSp/bBGXid3no27fdTtd5/QihetehGXBxhK3yLuNdzNvbQ76LgtOUPs7CjzHzIWDqbd6Ez5OAiUkWeZHoMq1bhEogxvXGJKCaXg4gRjp8mZRQERAaHCdKY/VoqMQc+rbEpAgkRmfpEzgYHB0cWM4GYPWR04xWCjVTw7R+sX/nODyE9CVPI5VerAGcb60FCJEtyujERBTP0ZnYDJlm9KejgVkHEk8SM0yWZrOBoqtQg9RGKkU9W02cnX2uyMVGNV/W8MO1UZw4SPb52aJiIEw3CzwMaJ5zoiNg9QwImwdeYHur7+F4JtYmrMTx2Do0N3rXDArzCsH24LvRoZkCPwGNdgkCXvpRGIljDumIeOLAMY04To1rhqX406v0yPZHlZ0bwEd4ZW3+O8ixjc8lmuTJhMIh6kRk68FfYcHD1m/4wXcFIWBSY4QAB+Qdh1WA8nkAug5HU4+TUwzZjlyOaclKLTokrEw7GqsI87ZSDtMFs5c/OXIR7Vm58/Xfituk6I7fHsZTvkhKEv427zmCbUuIpXF7Yc1KDG3aqcO3AJXN5OPgWXeqZzpoj6XbyqNSLBDd59RlZDXIm0TeqHP1VGeZd83E4s1CV0qhqkwukB7gXnBPxS6Z160WBwmyr4LTpijLELEbgGw4r6dvKJxVpKqSFWIp0iw2m9hyWgVibH0lNowzhG+zWIRPOhkvqRNj364SnyFWDNWEwmzZGS94FEMPqOc2G4Y8bwRVFXw7EjtN3Ft6/M8s9t8QSae1RHPmogbuDpK42mcbdiXOF+VjGu0D6sGB82jXcYx7NLlxDsmeZOEwMvsJ3ZwmlckBK4cU/wUJGMigmeK2Y7tU296nA/ilNMtQzJ/SbXzaXpZtN0d2m62zThUasLW7CidhtvZAxk9RVur6wC75X1x9IFeQXJK+Tp1SHzBmS+skabCLmCeNlAuFEIE7oSszdr4Garza+B4Cskc8yJbrhLKZV3v3YHqR5rkN8uPJ3d9wagIm3zl0yMnuMHx/MvRj2I/w5cZLsA9ku2YAvQgpxyfWsMfcZ3Mc+tPMZozAhlmiHwaYDkJWdns5PHxcvJ4OXlXBMw3xVIOjbM07MjIy+eVL81OyXw7Qnf3JV5yp1SbDLV0uTWPc48oLhnNLewAIuIsxNsz5rp3OXvI9C6vOrFeLMgCm+FBLQ/wKfjKsCrv5joKXZ7mGuUwzuRcRwf7zpVo+1D+TXQDNb/aQo/5VQWiK1ZRiMGoXU1Esrx4dtqEiUErwqOs2L4epgfzRwrgyGP482rl8rLo3nW1+2vm5tpzW1db1p3WTmuv5bd/af/a/r39x9pva3+u/b32Tw5tnyt8vmjVnrV/weD/84c</latexit>
slide-19
SLIDE 19

Deep spline networks: Discussion

37

(Kn,` = 1, bn,` = 0)

Linear regression: λ → ∞ ⇒ Kn,` = 0

<latexit sha1_base64="WEN5hvLrvHbFqA5So5vlFIrXKFY=">AMZXiclVZb9s2FLa7W+umW7MNe9nDiCUGusIxLGdtmABiqUNEizrstxaJAoCSjq2mVCXklRsh9Av2y/Z454GbH9gjzuUbEWSk8EVYFs+38fvHB4eHtKJOJOq0/mjfu+9z/48KP7DxoPFx59/MnjxU+PZRgLF47ckIfijUMlcBbAkWKw5tIAPUdDq+dy02Dv74CIVkYHKpxBGc+7Qesx1yq0HS+WD9s2jJ2JLj6J+peDqnwiBv6EcIO40yNk4btQJ8F2gSTaDt1qf24L2CcEGKLmIO2YJSkX0mjSVKGXCcOR8EWEeC1CLIhaBF0DoGiLeKOaZBbx8B5OEQ07ofcS39RHJWMy5WICuqDwjmQJ26vTXapH4VC4a+CEYnaxLJWv2loW4WRhGiji0abQ0/5VGDY+N8nNlPgG7DTfmZgwfqDKd5xfZyiIZBdzCEVGDC6lyZj62TZ5phLj6K6zYKeGhN73wymQoRD8vO5Dlo2Rp9sdJYbxIbAy9J0/nip0+6kD5l9sSYvS7XJs3e+OBf2wvd2Mf0uJxKeWp1InWmqVDM5YARxhg/5hMTqNMEJ6SJo/0QoGfQGVpL/GoL+XYd5DpUzWQVcwYb8WMRcieLHt1HExU0XIaq973Z5oFUawgcLOAejEnKiSm1IjHBLiKjwl1XZxXjAvmkZLCoXWmTfBmeAmQLuXgbR/6J7pAfArUIgLCGCIxelTzLPdoz7jYw96NOamMGVv+t5oYu2kpS8JUokTcw5KrqPdRg6ucl7EWG+JFn0Hq7fVa89a3WSGY4p9IyD5dOafJBWDMbhMWRuMJynJw6uDuxvli+ew7mWwd4xU3TmP6klSUjfv5laf7ck5tnP47aZt0zamd7v53SQnS/0+7rKAGkHUt08USTbQ9kFg78FSvdNrf4ZByprHWOSfLjSwq+TamApZvhnSND3KXkzv9rN34yfOuWDC+29Hq1FHRTeYgOcW9YN84e6pvq5a80U40TYrSDBGzAciSdbt8xcO8Tope0oXInVQXm/XNchoPRJv1Fdgw5gjfZHESPmkuqQsjM2/KDxHrBiqWU4UTJOR9oKDCFxGOTHdMORZIyh6YcF+Yr6JbVqf4+j96gRZcFJmnFQZm1gdU3CzCu4XwBntkwI4I/uqAL6qgtuMOyDUlI9U29XOfAW25lZOIUFKYBr+woPFcFo0MdzRW8kpqkXB+yGUYKpFngeVrVe5NCLKrSTQztVyGOBmDejLctvMgwEMUpbM1UgXNtxuPSedkKkmtz0ShTRhXKaKZGqwxwxhWGHboQ5/OxOxcVniXs82vwuAzIiPMiaoMwhFNa8enpzmGpbl1suOJ/t4ySo50NnmT5NcoUbHM+/jH0QuSl/muEJuUnSHTMh/ZJ9m5Q/0esk1G/hohnAoIP2VkZoDcCOmyGePFRW68yI07gcdcPSlX1s5sCdCJ7NLV+i9qfJWiMPtIfNgQJVOWfmQl6MoGxFGU0X9EjtAEHLmM4UXl0ZzJX3IAShsyX2CvZxEeLMgkl0DAeW28WnYEvDGHPTVALs9jBS6wzuRtr5NL4UF0PShwCwFXgJSrHxSTG/gUgmUNgrtbmUhFXWSwvFhGioK4YzwrmhVb4azL8fdtrXa7vzWXq+Prk13q9Wfu69qRm1dZqz2vbtb3aUc2t/17/s/53/Z+Hfy08Wvh84YuMeq8+GfNZrfQsfPUfOamI5Q=</latexit>

(Agostinelli et al. 2015) Adaptive-piecewise linear (APL) networks

<latexit sha1_base64="HBSIoFcNqpFlMl6Baew2fFycldY=">AMT3iclVZbxNHF7TGzi0QPvYl1GDpYAcy5uU0kZNBQUikFKaAgFENopmd4/tIbMXZmZjO6P9Of01farUPvSXtE9Vz+wte4EqrGR7fb5vnPmzJkz48acSTUe/9W78MGH38ycVL/ZXLn3525eq1z5/LKBEe7HsRj8RLl0rgLIR9xRSHl7EAGrgcXrjH9wz+4gSEZFH4TC1jOAzoNGQT5lGFpqNrvR8GjkxcCZ7+kXrHcyp84kVBjLDLOFPLtO+4MGWhNsGk2slc6iCZClimhDgi4aBtWKTZV9ofkIwht4jLUXBIBPhDgmwIhwSdQ6jokHhLGlbWJXAezRFNphH3s18URyXjcj2mgagcA5kzZuMyC4N4kgo/FWwIPGI2Pbmjb52VBRLiLc30OhwmKiACgwb/wfEYQoCA45Htws2HRW4mMvwCkaArnr01ixE1iPGXgwZxKISsVZO3u3u4NEoKaR+JY9okDoZ8n5Ojq6ng0zh7SfbGLl1WrePaOrl36x/EjLwkwER6nUh7Y41gdaioU8zhgLAlGipnDVOkslSkZoMknk0jgJ1R5ghs8Gki5DFxkBlTNZBszxrdixiLkRDa9ui6mpG45SNTk20PNwjhREHp5QJOExURU1TEZwI8xZeEeh7OK8Gl8UlD4Zl9qE3wZngDkB7l4G+Pvts41DPgJ6AQFxDCHMswoJhnZ0IDxpc+TGjCTQnKSfneH2CVZEUuCVKJm3AOSm6h3UEOLl5VrlhZqRZTF+t0OB7dvjUcpx2OKemcg4UyLD5Iqwfj8gRyNxhJM05OXdyHWEms2ieGnRabxHjFLdIvX9KWsnF/fuVyB5TG6f/XtomXefUzvb5+6QE6f+n3VRQM8j7k+lXqSbamUmsHbip18ejb3BIM9NY65yT6/08KvkmoQKunw3ZMD7Iu5y808/tMz9V3hULl+92tFk6qrvJHaQHuBecM2c39duqpWqphaZJUZYhYjYAWbXfLt/ycF4ndS/ZQlRO2ovNpmY5jQeizfoKbBjnCN9ksQif9DNd0hTGNl8XPkesGKpZThTMkpH1gqcxeIxyYrphxPNGUPfCwiep+SaOaX2uq5+0J8jCV03GqzbjHlZHCd5rg09qYEf7VQ3syD6ugY/b4EPGXRCqpGDP1A/bHiD7cwsnMKCFMC1c4KHimA0nOK5ordT09TrA3ajOMVUi4DstrXuV9D9NvSogh61IZ+FqsD8jrcdvLIwEPUp7HSqwD0143Hp/HwFyam5UjQpixZl0anRNkN0GPMWw4kCmNJOzO5xi3fcbX4tBu+ILDAnqjUMIxTtvPp4cpoLV5bPz+enOerdsOBzjdP+QzICW5wP9y9tPYy/hlhgvygGQ7piD9lEtOztDge6yTRY7+HCOcCYgY+RmgMoI2bIZ4+vXlfF1ZXwU+szTRSkH2q6APRG5uV16Qu+VyjsRDnfmzIcZVTpjVUMeLOJ8RBSXivoBdoAw4ixgCi8u/cF69pCnoLAlTwn2chLjzYJIdgoElDfCp+9IwLtxOFUz7PawUOgO70Ta/jq7/tVA04dCsxR4Cciw5klR3rWlEihtFEYbrYVU1E1rx4dpqCiEM8K7ot2+GXZfnm+M7M3R+JeN1Ttbxa3xovWl9ZW1ZtnWbeuO9dDas/Ytr/dr7feH70/V35f+Xvl38sF9UKvePnCajyXL/0HxuyBvw=</latexit>

(Kn,` = 5 or 7, bn,` = 0)

<latexit sha1_base64="YsBoYURyo9beUtDnCqQI4/Rj52k=">ALQHiclVbtw2EFXSW6Je4rSPfSHqXSANnIXkNHVa1EAJ4aNpqlrx05gyzAoaXaXNnUJSdm7JvQb/Zo+FWj/oX/Qt6JvRZ86lLRrXezCIeCFPOfwzHCGHNJPOZPKcf64cfOd97/4Nbt+0P/r4kzsLdz/dk0kmAtgNEp6I1z6VwFkMu4opDq9TATyObzyT9YM/uoUhGRJ/FJNUziM6ChmQxZQhajBede7/sjHS95wHm+qhHEkF6K70l0vOIt0f8Cwz/c3pf2kcLi87AKQbpfrjVx6JVja2ju7f/8cIkyCKIVcCplAeuk6pDTYViAYfc9jIJKQ1O6Ah0saCc9NEUkiHGMkxiRQprg0cjKaeRj8yIqrFsY8Z4KWYsQg5l06vR7ldtxkavj4ULM4zRTEQRnQMONEJcRkYRMQKD4lNAgwHVlVG8DYWX7qE2wZvpDUAGlEO4Ovhm+VCPgZ+CQlxADGdBEkU0DrU3pBHj0xCGNOMq154czr7tvt0nRVUlQSrxM85ByW/R7iEHN0GRKx1lo4SHuRYjP9fukjNYebTk5B2OgGnFcQZIKP+QVg/G5xmUbjCSZpyc+rjxFEQMoVLUsHPiYyDNl5hktuzj7ylbNxfX7kK97rauPy30jbpuqa2z7GUb5MSpP+fdlNBjUFmvsTthQc010R7Y4l7B+7rB87ga5zSzDTudc5Jzy6jkm8yKqB3MWXZ+CBXObnSz8qFn3neFYunVzt6OHNUd1M6yA/wLHgXzu7ry3aLiwLFT6VpUlRkiJgDQBbdy+VbHq7rpO6lKMTcSbvYbGTKaTwQbeorsGFcI3yTxSp8Yhe6pCmcUlEXvkasGKopJwoWySh6wU4KAaOcmG6Y8LIR1L2weDs3v8Qzrc/39XZ7gSzebzL24w13B0zcK0NbtfAjvZ+DezIvqiBL9rgBuM+CDWjYM/UG20OvMF2ZgqncEMK4No7xUtFMBqP8F7Rq7lp6vUJz5M0x1SLiDxvaz2dQ0/b0OYc2mxDIYtVhYUdb+t4RzMQ9SWsd3aBf27mY+nCsoLkHX6TcqkRZl09mibITqMsxbDSyIY0U7M/kmLd9Jtfi0G74hMCeqNQ0jFO28hnhzMjXNy9yG5fXk7S26DQe6PDyz0SeneMDx/ivZO2lQ8GcZrsh9UpyYivRDKTm8QKPvcJ9MSvTHFOFCQEQFozQDzI1QlM0Yj4/nxuO5cTMOWaCrRxpdw5sicQv7TIQemumvJ7gdO+MhTCmShes+ZRnk7SckaQzRf0MO0CcBYxhQ8Xu/+gGQHFLbkEcFeTlJ8WRDJzoGACgY4bE8CPgbjkRpjt4eJQnf4JtLuVwG+deqg6UOxKQU+AgqseVPMHpdSCZQ2CoPlViEV9fPa9WEaKgrhivCt6LZfht2PveWB+3Dg/LS8+ORx9Wq8ZX1ufWHds1xrxXpibVhb1q4VWD9bv1i/Wb/bv9p/2n/Zf5fUmzeqOZ9ZjWH/+x+Nx31</latexit>

(Kn,` = 1)

<latexit sha1_base64="8N8UzKw9qR1YdaNKWYW/yACNf4w=">ALIXiclVZb9s2Fa7W6tdmyPeyGWGOiK1LDcbemGBSiQNkiwrsuSJi0SGQElHdtMqEtJKrFD6KfsacD2W/Y27G3YH9njDiXZ0SUZXAI2iPN9/M7ROeQhvYQzqXq9v2/dfufd97/4M5d+8OPv7k3tLyp4cyToUPB37MY/HaoxI4i+BAMcXhdSKAh6HV97ZpsFfnYOQLI5eqmkCg5COIjZkPlVoOlavr/6w4mO1lzgPNtwVr+0T5ZWet1ePkh74pSTFascuyfLd/91g9hPQ4iUz6mUx04vUQNhWI+h8x2UwkJ9c/oCHQecUY6aArIMBb4ixTJrTUeDaWch4yQ6rGsokZ47WYsQg5lHWvnhdmdtVynKrh4FmUZIqiPwioGHKiYqJSRMJmABf8Smhvo/flVKF8dYUXjoDbYI3y2uA9CmHYKP7bX+gx8DPQSEuILPw5DGgXaHdKQ8WkAQ5pylWlXDmdzu2N3SF42SZBKvJRzUPI7tLvIwSrnudJhOop5kGkx8jLtrPW61+v9bIWR8C05PS6SCh+SKsG4/EUCjcYST1OTj3cWQpChlAhatgZcUXKQRuvMns2SRrKBv3iyuX4S6qjZ/VtomXQtqexL+TYpQfr/adcV1Bhk6kncXngCM020O5a4d+CBftjrfoNL6pnGvc45WbWLqOSblApYvVrSNz7ITU5u9LN+5Wed8Wi6c2OHs0cVd0UDrJjPAvulbMH+rd4qBA/ldqmhTlGSLmAJAV53r5hodFnVS95IWYO2kWm41MOY0Hok19BTaMBcI3WSzDJ3auS+rCRV4QVixVBNOVEwT0beC/YT8BnlxHTDmBeNoOqFRXuZ+SeuaX2ep/eaH8iozrjqMnYxN0xAzeb4F4FbGkfVcCW7IsK+KIJbjPugVAzCvZMvd3kwBtsZ6ZwCjekAK7dc7xUBKPRCO8VvZGZpl5d8DxOMky1CMnzptbTOfS0Ce3MoZ0mFLBIlVjQ8raFlzADUf2ErdYu8C7NeixdUFSQXKJOp06ZNCiT1h5tMkSLcdFguHEI9qK2Ttr8M7aza/B4C2RCeZENZhKZ1wBvTqamWZHboLie3MVp+ZAF4dnNjrkHA843n8Fez/xc/4swyW5Q/ITU5J+LCSHV2j4Pe6TSYH+lCcC4gwZxRmgLkR8rIZ4+np3Hg6N+5EAfN1uZVD7cyBXRF7hV36Qu/OlLdiXO5esADGVOmcNV/ybJIUK+JkpqifYQeIYs5CpvDhYnce5oPsg8KWPCLYy0mCLwsi2SUQUH4Xh+1KwNdeNFJj7PYwUegO30Ta+crHt04VNH0oMqXAR0CO1W+K2etRKoHSRqHbxRSUS+rXB+moaIQfhG+FZ3my7A9Oex3nUfd3s/9lSePy1fjHetz6wvrvuVY69YTa9vatQ4s37qwfrF+s363f7X/sP+0/yqot2+Vaz6zasP+5z8SWxQO</latexit>

(He et al. CVPR 2015)

Justification of popular schemes / Backward compatibility

<latexit sha1_base64="1sNw3U2LtZhDEYeEMXeGw+Bi70=">ALUniclVZtbxw1EN60vLSbAil85ItFehKq2utCqQgKlVNGyWolJA2aZVsFHm9c3dOvC+1vcldrP0/Bo+IQH/BPGJ8e7eZV8SlFq602qex8+MZ+yxg1RwpQeDvxeuXf/gw48+vnHTXbz1yaefLd3+fFclmWSwxKRyLcBVSB4DuawFvUwk0CgS8CY7XLP7mBKTiSfxaT1M4iOgo5kPOqEbT4dJTX2WBAmZ+ypSe20kyJGmSZoJKotgYIlDkAXlK2fEplSFhSZQiL+C62l+uLQ86A+KQbofXvWx7FRj6/D2zX/8MGFZBLFmgiq17w1SfWCo1JwJyF0/U5CiMzoCUywxJz0hWSYSPzFmhTWBo9GSk2jAJkR1WPVxqzxQsxapBqptcgiHK3btnP9PDRgeFxmIWRnQMBNEJ8TmlYRcAtNiSihjuK6Maoy3ofDaOzA2eDu9AShGBYSP+9+vHJgxiBPQiEuI4RQTHdE4NP6QRlxMQxjSTOjc+Go4+3Z7bo8UdVYEqSTIhACtfkC7jxzcFkWuTJSNEhHmRo6C3Hj3Bv3Vb+8N8g5HwrTiDPpIKH9IqwcTiAxKNxhJM05BA9yKGiKOUClq2TnxZSbAWK8wyd3ZR95Stu6vrlyFe1VtXP57adt0XVE7EFjK90kJ0v9Pu6mgx1AeUns0c0OMP1a4d+CuT/of4dTmpnGvS4EueOWUal3GZVw53zKivVBLnNyqZ/Vcz/zvGseTy939HDmqO6mdJDv41nwz53dNRftFg8Fir9K06aoyBCxB4AsexfLtzxc1UndS1GIuZN2sfnIltN6IMbWV2LDuEL4NotV+MQtdElTOKWyLnyFWDFUW04ULJR9IJXKTBOBbHdMBFlI6h74fF2bv+Jb1tfEJjt9gJ5vNdk7LUZa7g7ZuBaG9yugR3tvRrYkX1ZA1+2wQ0uApB6RsGeaTbaHiH7cwWTuOGlCMf4KXiuQ0HuG9Yh7ntqnXJ7xI0hxTLSPyoq31bA49a0Obc2izDYU81hUWdryt463NQdaXsN7ZBcGZnY+lC8sKkjPU6TUpkxZl0tmjbYbsME5bD+JYEQ7MQfHLd5xt/m1GKIjMsGc6NY0jFC28xrizWmfFGVuw/J68neXvYDUx6e2eiREzgeP+V7FcpK/izDFfkHilOTEX6uZQcnqPRj7hPJiX6S4pwISCjglGaAeZGKMpmjUdHc+PR3LgZh5yZaitHxpsDWzIJSrti0mzNlNcTnO6f8hDGVJuCNZ/yfJKWM5J0pmieYweIE8EjrvHh4vbuF4O8Ao0teUSwl5MUXxZE8TMgoFkfh+srwOdhPNJj7PYw0egO30TG+4bhW6cO2j4U21LgI6DAmjfF7LmptERpq9BfaRVS0yCvXR+2oaIQrgjfil7Zdj92F3pew/7g19Xlp8ql6N5wvna+crx3PWXWeOBvOlrPjMOc353fnT+cv9w/38WFxesl9dpCNecLpzEWb/0HfZYnkg=</latexit>

Global optimality achieved with spline activations

<latexit sha1_base64="iogvmyZVER1ymYu5nIU+oPley3s=">ALXHiclVZtbxw1EN6WAu0dhRQkvDFIj0JVe3pNgVSEJUqpQ2JFEpImrRKNors3bk7J96X2t7LXa39U/wa+IEv4KPjHf3LvuSoKulO1kzj59ndsYemyWCKz0Y/Hnj5ge3Pvzo49t3Ot1P7n762cq9zw9VnEofDvxYxPINowoEj+BAcy3gTSKBhkzAa3a+Yf2vJyAVj6NXepbASUhHER9yn2o0na7seCplCnzs4gZFSRONA+p4HpGqD/mMIGAXHA9JsZjQ+LlgkZCkKnESiJI80nOpbLsdGV10B/kg7QnbjlZdcqxe3rvzr9eEPtpCJH2BVXq2B0k+sRQqbkvIOt4qYKE+ud0BCbXzkgPTQEZxhJ/kSa5tYajoVKzkCEypHqsmj5rvNJnLVINV2VsTDrVC3HqR4+OTE8SlINkV8ENEwF0TGxGSYBl+BrYRPo43elVGO8NYZX7omxwdvlNYfyqYDgaf+HtRMzBjEBjX4JEVz4cRjSKDekIZczAIY0lTozHhqOJ93ep0eySuCEIJS4UArX5Eu4cYrFZRvTAdxSLIjByxzLgPB/317x4OshZGwqzEDPoIKH4IqwbDRAqFDEZSj1NQhptSQ8jRVZBadEY8mQowVhWmWc+yRrMVn5jLcZbnx89+L26ZrSW4msJTvkxKE/x93nUGPoTiu9sBlBg/lWOHegQfm0aD/PS6pZxr3uhDkfqeISr1NqYT7l0vWrAa5TuRanfVLnUXeNY9m1ws9ngtVZQqB7BjPgncp9sBctVtcJMj/Sk6bojxDxB4AsupeTd9QWFakqpIXYiHSLDYf2XJaBWJsfSU2jCXCt1kswyednJfUiRMq8RLxIqh2nIiYZ6MvBfsJ+Bz7Om2G8aiaARVFR7tZfafeLb1MWb2mh/Io6M64qiJ2MDdMXduNJ17FWeL+6jibNG+rDhfNp1bXDCQeg7Bnm2mh4i+3MFk7jhpQgjDfBS0VyGo3wXjFPM9vUqwt24iTDVMuQ7DS5ni9cz5u7YVru+kKeKRLX9BS28T7m4OsfsJmaxewd3Y9li4oKkjeIU+vDpk2INPWHm0iZAtx0UB4cQgj2oqZnTdw5+3m10CIFskUc6IbyzBC2cxrgDcnPkKyIrdBcT15h6tuTcAUh2c+emSCBxzvwK9n/g5fp7hEtwj+YkpQb8UlMNLb/gT7pNp4f01QXdOIMcUZgBFkbIy2aNZ2cL49nCuB0F3DflVg6Nu3DsypgVduVLsztn3oxuXfBAxhTbXLUYsmLaVKsiJM5o3mBHSCKBQ+5xodLp/coH2QfNLbkEcFeThJ8WRDF3wEB7fdxdDwF+FCMRnqM3R6mGuXwTWTcb31861Sdtg9FthT4CMh9Zti/vBUWiK1ZeivNQqpKcsq14dtqEiEX4RvRbf5MmxPDtf67uP+4Le1WdPylfjbecr52vnG8d1p1nzpaz6xw4vO784fzt/NP56/urW63e7eA3rxRrvnCqY3ul/8B9hcrnw=</latexit>

State-of-the-art Parametric ReLU networks 1 ReLU + linear term (per neuron)

<latexit sha1_base64="RKMNleUNEw2Y8RJVQTx2fi5D1s=">AMdHiclVZtb9s2EJa7t9bNtmb7MmD7QCw10BfHsJx13YoVKJa2SIGsy5ImLRIFAUWdbSbUS0kqsUPoD+4f7F/s07Dt046SrUhyOqQCbMv3PHzueDwe6SeCK93v/9G69sGH38yfUb7ZtLn372+a3lL/ZUnEoGuywWsXzjUwWCR7CruRbwJpFAQ1/Aa/9k3eKvT0EqHkev9DSBw5COIj7kjGo0HS23WMdTqa+AmZ8pOzmjMiAsDhOEfS64nmZtz4cRj4wNJjNe7tKE6UjCNCPEk6kA48Iky7+ydofkDPWI+AIFu0RC0CXIhqhL0DlEmnYJm9KotE5BiPgM0XQUiyD/RXFUsi5XEypCBrnQO6wY9s0jCJpcZfDROS9Ijrt1tG0/HiYLk8QCNnoChDqnEsPF/SDyuIbRgv/fAwpKPxnO8z0KcoiWQHY2Sq/FwVY9hlaKLrcK15Ixsw+YuiUCfxfJEeV7bLSz3iU08lQTjC8mdBCSUhlHd9vEgygosnZ0a6Xf6+cPWXxZy8rT75y8mfraPnG314QszTEbDFBlTpw+4k+NBgUZwIw4BSng+nFfJo83xnpoCkgw1jiJ9LFKtR4NFRqGvrIDKkeqyZmjZdi1iLVUNW9+j7mrWo5SPXwh0PDoyTVELEioGEqiI6JrTwScAlMiymhjOG8Ukx2QGoKr9xDY4O3w2uAYlRA8Lj34+DQjEGcgkZcQgRnWKshxTx7QxpyMQ1gSFNh61QN5+/tDpZSvhMUQSrxUyFAq0do95CD61fWNJZfZuTIx2Lu9nsPH3T72QLH1n3BwWrqzj5IqwbjixQKNxhJPU5BfdysWG683EyWnc12kvWK+6g9f8kaytb91ZXn2/SK2j9K26bqidt4M3iclSP8/7boCbtidmlhlivLHC2oF7ZrXf+x6H1DONtS4Eud0uolJvUyrh9sWQgfVB3uXknX4eXvgp865NH23o7W5o6qbwkF2gHvBu3B2z1xWLWXfnWnaFOUZInYDkBX3cvmGh6s6qXrJF6J0lxsPrLaT0QY9dXYsO4Qvg2i7PwSTvXJXVhPAuqwleIFUO1y4mCeTLyXrCTAONUENsNY1E0gqoXHm1n9pt4tvX5vtluTpBH+3XGfpOxjtUxB9eb4HYFXNDer4ALsi8r4MsmuMGFD1LPKdgzUaTA2+xndmF01iQEoTxTvFQkZxGIzxXzOPMNvXqgM04yTDVeLptNrWeltDTJvSihF40oYBHeoYFC96e472Gg6xO4flCFfjndjwuXVCsIDm3946ZdKgTBZqtMmQC4yzBsOLQxjRhZj9kwbvZLH5NRhiQWSCOdGNYRihbOY1wJPT3sqK3AbF8eTtrbg1B6bYPOnQ05xg+P5V7B3Epbz5xmekTsk3zEz0i+F5PACDX/COpkU6K8JwrmADHNGYQYojZAvmzUeH5fG49L4Igo4M7NSDo1bAlsy9gu7YtJszZWfxzjcO+MBjKk2Oasc8mySFCPiZK5onmEHiGLBQ67x4tLurOYP2QGNLXlEsJeTBG8WRPFzIKBZD5+2pwAv0NFIj7Hbw0SjO7wTGfe7/I5YAW0fiuxS4CUgx+onxfxCrREavQGzQWUlM/qxwftqGiEM4I74pu82a4+LI36LlrvcFvg5Unj4pLo3Pd+dr51rnjuM5D54mz4Ww5uw5r/d76s/VP69+bfy19s7Sy1Cmo1qzMV86tWep9x98TI2o</latexit>

(Glorot ICAIS 2011) (LeCun-Bengio-Hinton Nature 2015)

Standard ReLU networks

ReLU(x; x1) = (x − x1)+ x1 x

Comparison of linear interpolators

38 0.5 1.0 1.5 2.0 2.5 3.0 0.5 1.0 1.5 0.5 1.0 1.5 2.0 2.5 3.0 0.5 1.0 1.5

zm

<latexit sha1_base64="24JOALiYbOXS/l1FTCkcZ7uZyZ0=">ALEXiclVbtw2EFXSW6LekvSxL0Q3CxRBslg5bZwWDRDAiWEDaerasRPYMgxKmt2lTV1CUvauCX1Cnwq039K3oq/9gv5JHzuUtGuJsosNgV0Qcw7PjGbIYOM6mGw3+uX/v/Q8+/OjGTfjTz797PNbt+/syTQXIeyGKU/Fm4BK4CyBXcUhzeZABoHF4HJ2sGf30KQrI0eaVmGRzGdJywEQupQtPO+VF8dKs3HAzLQboTr570nHpsHd2+a8fpWEeQ6JCTqU8IaZOtRUKBZyKFw/l5DR8ISOQZcRFqSPpoiMUoG/RJHS2uLRWMpZHCAzpmoibcwYL8WMRciRbHsNgrhwm5aDXI0eH2qWZLmCJKwCGuWcqJSYtJCICQgVnxEahvhdOVUYb0vhlXeoTfBmeQuQIeUQPRl8t3KoJ8BPQSEuIGzMI1jmkTaH9GY8VkEI5pzVWhfjuZzt+/2SVkmSZBKgpxzUPJ7tPvIwaqWudJxPk5VGgxDgrt3R8OVr+9Pyw6HAGzmjMcIKH6Ia0ZTMBzqNxgJO04OQ1wJymIGUKVqGEXxBc5B28wrRw5PCUjbul1euw1WGz/nbRNupbUDjiW8l1SgvT/024rqAnIPJC4vfDEFZpofyJx78A9/WA4eIRL2pnGvc45uetWUcm3ORVw92LJivFBrnJypZ/VCz+LvCuWzK529HDuqOmclAc4FnwL5zd05ftFg8Fyr9a06SozBAxB4D0vMvlLQ/LOml6KQuxcGIXm41NOY0Hok19BTaMJcI3WazDJ26pS9rCGRVN4SVixVBNOVGwTEbZC3YyCBnlxHTDlFeNoOmFJduF+Se+aX1BoLftD2TJfpuxbzPWcHfMwTUb3G6AHe39BtiRfdkAX9rgBuMBCDWnYM/UGzYH3mI7M4VTuCEFcO2f4qUiGE3GeK/oJ4Vp6s0FL9KswFSLmLywtZ4toGc2tLmANm0oYomqsajbR0vXQai+QnrnV0QnJv1WLqoqiA5R51+mzK1KNPOHrUZosM4sxh+GsOYdmIOTizeSbf5WQzeEZliTpS1DCMUdl4jvDmZmhVbqPqevL3el7Lga4Oz3z0ySkecLz/KvZOFpb8eYZrcp+UJ6Ym/VhJji7Q+AfcJ9MK/SlDuBQcmozALI5RlM8bj4XxeGHcTCIW6norx9pbAFsiDSq7DIXemiuvp7jcP2MRTKjSJWux5Pk0q1ak2VxRP8cOkKScxUzhw8XtPygH2QGFLXlMsJeTDF8WRLJzIKDCAQ7Xl4Cvu2SsJtjtYarQHb6JtLcS4lunCZo+lJhS4COgxNo3xfy1KJVAaMwWLEKqWhQNK4P01BRCL8I34qe/TLsTvZWBt7DwfDnb3pPH9evxhvOl85XzteO56w6T50NZ8vZdUJn7Pzi/Ob87v7q/uH+6f5VUa9fq9d84bSG+/d/INkPsQ=</latexit>

˜ ym

<latexit sha1_base64="IWIuZRUfqQxQR2SfjmUiEwEbuo=">ALGHiclVZLb9w2EFbSV6K+kvTYC9HNAkWQLFZOG6dFAwRwYthAmrp27AS2DIOSZndpU4+QlL1rQv+ipwLtb+mt6LW3/pMeO5S0a4myiw2BXRDzfxmNEMOGWScSTUc/nPt+nvf/DhRzduh9/8uln9+6fWdPprkIYTdMeSreBFQCZwnsKqY4vMkE0Djg8Do4WTP461MQkqXJKzXL4DCm4SNWEgVmvZ9xXgEZHYUH93qDQfDcpDuxKsnPaceW0e3b/7rR2mYx5CokFMpD7xhpg41FYqFHArXzyVkNDyhY9BloAXpoykio1TgL1GktLZ4NJZyFgfIjKmaSBszxksxYxFyJNtegyAu3KblIFejx4eaJVmuIAmrgEY5JyolJjskYgJCxWeEhiF+V04VxtSeOUdahO8Wd4CZEg5RE8G360c6gnwU1CIC0jgLEzjmCaR9kc0ZnwWwYjmXBXal6P53O27fVJWSxKkiDnHJT8Hu0+crC4Za50nI9THhVajINCe/eHg9Vv7w+LDkfArOYMB0iofkhrBhPwHCo3GEk7Tk4D3FAKYoZQJWrYBfFzkEbrzAt3PmksJSN+WV63CX1cbPfydtk64ltQOpXyXlCD9/7TbCmoCMg8kbi8eIUm2p9I3DtwTz8YDh7hknamca9zTu6VTybU4F3L1YsmJ8kKucXOln9cLPIu+KJbOrHT2cO2q6qRwUB3gW/Atn9/Rlu8VDgfKv1jQpKjNEzAEgPe9yecvDsk6aXspCLJzYxWZjU07jgWhTX4ENY4nwTRbr8Ilb6pK2cEZFU3iJWDFU04ULJNR9oKdDEJGOTHdMOVI2h6Ycl2Yf6Jb1pfEOht+wNZst9m7NuMNdwdc3DNBrcbYEd7vwF2ZF82wJc2uMF4AELNKdgz9YbNgbfYzkzhFG5IAVz7p3ipCEaTMd4r+klhmnpzwYs0KzDVIiYvbK1nC+iZDW0uoE0biliaizqeFvHu5eBaH7CemcXBOdmPZYuqipIzlGn36ZMLcq0s0dthugwziyGn8Ywp2YgxOLd9JtfhaDd0SmBNlLcMIhZ3XCG9OpmZFlduoup78vZ7XcqCrwzMfXKBxzv4q9k4Ulf57hmtwn5YmpST9WkqMLNP4B98m0Qn/KEC4FRFwyKjPAwghl2Yzx+HhPF4YN5OIhbreyrH2FsCWSIPKLkOht+bK6yku989YBOqdMlaLHk+zaoVaTZX1M+xAyQpZzFT+HBx+w/KQXZAYUseE+zlJMOXBZHsHAiocID9SXgIy8Zqwl2e5gqdIdvIu2thPjWaYKmDyWmFPgIKLH2TF/NEolUNoDFasQioaFI3rwzRUFMIvwreiZ78Mu5O9lYH3cD8+Zve08f1q/G86XzlfO14zmrzlNnw9lydp3QSZxfnN+c391f3T/cP92/Kur1a/WaL5zWcP/+D6toEoQ=</latexit>

arg min

f∈H1(R)

Z

R

|Df(x)|2dx

s.t.

f(xm) = ym, m = 1, . . . , M arg min

f∈BV(2)(R) kD2fkM

s.t.

f(xm) = ym, m = 1, . . . , M

(de Boor 1966) (U. JMLR 2019; Lemma 2)

slide-20
SLIDE 20

Deep spline networks (Cont’d)

39

Key features

Direct control of complexity (number of knots): adjustment of λ Ability to suppress unnecessary layers

Generalizations

Broad family of cost functionals Cases where a subset of network components is fixed Generalized forms of regularization: ψ

  • TV(2)(σn,`)
  • Challenges

Adaptive knots: more difficult optimization problem Optimal allocation of knots

`1-minimization with knot deletion mechanism (even for single layer)

Finding the tradeoff: more complex activations vs. deeper architectures

<latexit sha1_base64="GgXKr8vpnLDfg9nGCe6bJwl2814=">ANMHiclVdb9xEFN4ut3YJ0MAjD4xIVmqrZLVOgRZEpK0USuVNrRNW7WOorF9vDvJjMedGW92M/Irv4YnJPgt8IR45RfwyBlfNmtvi1JLWVtzvnOmXObkyDlTJvh8M8L3Xfefe/9Dy5e6n248tHn1xe/fSplpkKYT+UXKrnAdXAWQL7hkOz1MFVAQcngXHO07+bAJKM5k8MbMUDgQdJSxmITW4dLja/cLXWaAhtDtjyjkI9C5H8CIJdapz61fKLEiGymY5YT4KuNgh4MbX8M0r95r08KnP6OBJyGxtEQbRBcA8kGwSVQmLoBglnNJmvzoBzeYLSbCR5VLxRBTI5xZspVSAQdvJlTAekPtUpFIZfBuYknRAPO/61Z71jUw1pNseLvocYiOoQuO3t0AQnxkQTjh0QsVG41o6DEXeK8Tkh4imhk2AHCfSoPlCKiARi9FHGTdEolCw08JdJFUSHSuqnQ+diHKCfpOlP4mMSxrf7637eLxDb31TsOSM4YSZcQEhEXAolgSEY5owLcgVmEBCYqmIZsmIA+F0Bupqr19ZGjDOzIwYSXSWYpi1LhG6MmiXJRFuJGYMxCgagYzj6kChFClHL9AQz1qYoslED3w0A1JQhKpwjByhyZC2R3xIojL+h5fXhoNh8ZDlD6/6WOtUz97h6qV/UiGmcCIh5xq/dIbpubAUmVYyAHdnmFIMEUwJ2yRMznp41JUHDyWiSkzqYGjQuZCBApqBnrtswtvlbmVpSOdVNrEGD0F1deZia+eWBZkmYGkrA0KM64c7WrGswHhb7hM/RfiOfKMAcj0mB4h1YZ7zb3hDokHKItgfbh3YMfAJGJQrSOAEYyIo+tmPqWB8FkFMeOw4nRcf/f6WA5FWuCUBJkWKOYpbjuIwaLfl6dWEK5VaMgt96Gq8qNYb6EcRVcYoYDBJR/CFs0JuAZlGrQkqadnAbYaDT2LwtOHTe7Am9eVNoMTv152euG845ufH4b8Xt3HVO7qKhvY1LEP5/3E0GLNayA7uizC2x/lhj7sA1uzkcfINbmp7GXOecrPdKq/SrjCpYP9uy5XSQNyl5o54bZ3rmfjcsmb1Z0fVa0aKaUkH+EmvBP1N2zb4uWzx3fbifitO5qPAQcQVA1rzX07c0nFfJopby5qVtIPNRi6cTgOxLr4KG8Y5zHderMwnvYKXNInxPlskPoetaKoLJxIWzih6weMUQoZ3juGkpeNYFELSx7l7pf4rvUFgX3UPiBLXjQRL9qIHcyOWrjTFj5aEC5xv1gQLtE+WBA+aAvMh6AMjUEe6a928bAK2xnLnAGE1IBt/4ELxXFqLsrc7udu6a+uOG+THN0tRLkfpvr9lx0uy26Nxfda4vwdjWVLFrStoszGQO1eITdpSwITt1+DF1URpCcutmpCZm2INOlHG0j1BLipIXwpYARXbI5OG7hjpebXwvBl0im6BPT2oYWqrZfI7w5cX7JS9G5fXkP13zGgpsWTz10ycTLHDpZhyHfpyGBb72cAXuk6JiKtCPJWV8JhXfY5MS+nDFMUFgRIFolwGmC9CETa3eHQ0XzyaL97DKSu0VSoL680Fezgelus6VHavZt6VuN0/YRGMqbEFar7lzjQtd8i0ZrR3sAMkuPQaHBw6fU3i4c8BmPq6S7FyQKHxFMgYMIBPj1fg3Hjuxljt4epQXU4E1nvq2LSXRC6PpS4UOAQUMiaN0X9z4Q2Cqkdw2CrFUhDg3zh+nANFYnwRDgreu3JcPnj6dbAuz4Y/rS1dutmNTVe7Hze+bJzpeN1bnRude529jr7nbD7c/eX7m/d31d+Xflj5a+Vv0to90K157NO41n5z9J9Xj</latexit>

⇒ In need for more powerful training algorithms

<latexit sha1_base64="sZDk0ztnbRnR0MRExWVtldXN9dQ=">ALS3iclVZfb9xEHcLhdYFmsIjLyuSk1CVnuyENi20UqW0USKVEpKmrRJH0dqeu9tk/ae76+QuK38YPg1PSPDI5+AFIR6Y9Z+L7UvQdaU7WTO/c3szOzs+ClnUjnOn9euf/TxjU8+vXnLv3Z51/cWbj75RuZCKAvSDhiXjnUwmcxbCnmOLwLhVAI5/DW/9k3ejfnoKQLIlfq0kKhxEdxmzAqpQdLTw5K3w4YjRYVIzpbIVkxigJAMEkGiRABJkzMQg4wTJSiLWTwklA8TwdQokcLi07/geM+fugQp+8Uq/h45K6xK0ki1a1to/u3vrbC5MgiyBWAadSHrhOqg41FYoFHLbySkNDihQ9DF2XLSQ1HpzyCJFSmkLRyNpJxEPiIjqkayqzPCS3VGIuRAtq36fpTbTclBpgaPDjWL0xBHJQOFQFJiAkoCZmAQPEJoUGA58qoQn9bDK/dQ2cN9tbChlQDuHT/uOVQz0CfgoK9QJiOAuSKJxqL0BjRifhDCgGVe59uSg/rZ7do8UCZYEocTPOAclv0e5hxishyJWOsqGCQ9zLYZ+rt1lp7/2YNnJZzACJhUGc7pc/RDWdMbnGZRm0JO2n5z6WIMKIoaqktSgc+KJjIM2VmGc2/VH3mE25udnrtydlxuP/0HcJlxzcvscU/khIUH4/3G3GdQIZOZLC+8q7km2htJrB24p+87/Ye4pR1prHXOyZJdeiXfZ1TA0sWFWODXGXkSjtrF3amcVcsnlxtaLU21DRTGsgP8C54F8bu6cuqxUWC4q/iNCEqIkTMBSCL7uX0HQvzGmlaKRIxNdJNhuadBoLRJv8CmwYc7hvoli5T+yCl7SJUyqaxHP4iq6adCJhEYyiF+ymEDKiemGCS8bQdMKi3dy8080/p8X+90D8ji/TZiv4tYx+qoletd5U5DOcO931DO0L5qKF91lZuM+yBUDcGeqTe7GHiP7cwkTmFBCuDaO8VHRTAaD/Fd0U9z09SbG14maY6hFhF52eV6PlU976q2pqtripksap04Yy1DXyuGYjmETZmqsA/N/sxdWGZQXKOPL02ZNyBjGdqtIsQM4izDsJLIhjSGZ/9kw7uZLb5dRB8hmSMVGdbeih6MY1xJeTqUlexjYsnyfvzaLbMqDLy1OvHjnFC47vX4neTYMCX0e4AvdIcWMq0I8l5eBCGz3BOhmX2p9SVBcEIioQpRhgKoQibUZ4fDwVHk+FW3HIAl2VcqTdqWJbJH4pl4HQ2zXzRoLbvTMWwogqXaCmW16M03JHktaM+gV2gDjhLGIKBxe7d79YZBeUMrMZ9nKS4mRBJDsHAiro47I9CTgXxkM1wm4PY4XmcCbS7ncBzjpNpelDsUkFDgGFrv1S1HOmVAKpDUN/pZNIRf28XyYhopEeCKcFeuBkFz98Wal7672nZ9XFp89qabGm9bX1jfWt5ZrVnPrE1r29qzAusX61frd+sP+zf7L/sf+98Sev1atecrq7Vu3/gP/ZYl9w=</latexit>

CONCLUSION: The return of the spline

40

■ Continuous-domain formulation of compressed sensing

■ gTV regularization ⇒ global optimizer is a L-spline ■ Sparsifying effect: few adaptive knots ■ Discretization consistent with standard paradigm: minimization

■ Favorable properties of splines

■ Simplicity (e.g., piecewise polynomial) ■ (higher-order) continuity: the difficult part in high dimensions ■ Adaptivity/sparsity: the fewest possible pieces = Occam’s razor ■ Efficiency: B-spline calculus

■ Foundations of machine learning

■ Traditional kernel methods are closely related to splines
 (with one knot/kernel per data point) ■ Free-form activations with TV-regularization ⇒ Deep splines
 ■ Deep ReLU neural nets are high-dimensional piecewise-linear splines

1732

172

slide-21
SLIDE 21

ACKNOWLEDGMENTS

Many thanks to (former) members of EPFL’s Biomedical Imaging Group

■ Dr. Julien Fageot ■ Prof. John Paul Ward ■ Dr. Mike McCann ■ Dr. Kyong Jin ■ Harshit Gupta ■ Dr. Ha Nguyen ■ Dr. Emrah Bostan ■ Prof. Ulugbek Kamilov ■ Prof. Matthieu Guerquin-Kern


....

41

■ Prof. Demetri Psaltis ■ Prof. Marco Stampanoni ■ Prof. Carlos-Oscar Sorzano ■ ....

and collaborators ...

2

References

42

Deep neural networks Image reconstruction with sparsity constraints (CS)

<latexit sha1_base64="qBCtjnwOiLUjU+6cUwE4ER6m7w=">AJjniclVZdb9s2FW7r9j7aLq9DNgLsdRAVzSG5G7JVixYgBRFAvShS5a2SGQEFHVtc6YolaTieIT+zH7NXre3/ZtdSnZqScngEpBN8B6ec8l7eckoE1wb3/3zt0Pvzo4082Ot1P/v8i3ub9798pdNcMThlqUjVm4hqEFzCqeFGwJtMAU0iAa+j6YGzv74EpXkqfzPzDIYJHUs+4owaHLrY/DnUeaSB2SM0AFHAUqmNypkzkxk3E6IzivPNnFQmyqXR5OHByXfFxeaW3/fLRtqdYNHZ2v/aK9vLi/sbT8M4ZXkC0jBtT4P/MwMLVWGMwFN8w1ZJRN0RdbLq4gPRyKyShV+ElDytEajiZaz5MIkQk1E920ucGbOe5Gf04tFxmuQHJKqFRLohJidspEnPcDSPmhDKG/ubUoB81Bs2ogHiv/9NgaCcgLsHgChRImLE0SaiMbTiCRfzGEY0F6awoR4t+91et0fKAGmCUBLlQoDRT3E8RAzGs1yqTfJxKuLCqnFU2OCx39/94bFftDAK5guM30dA9SFs1ZlI5FDJoCd1PwWNMIcMJBxNFalDFyRUuQDrVOGqWP4XDWKnvj7xwts1qXHx70XtNms96khgHN9nPxD+P9R1AjOB6mS5g1RYsMJHiQGj+y239/BKfVdxjQVgjyofNJvc6rgwbsZAydBbtO4VWb3ncz1nhsu57fqPFnqrKpU/MV5MFz2ndYje1OeBG5P3M+C0yVjyekyn2wFN7M3BNbVWBGpgrHUaMaZj10knQCxLrQK8M6zm9fe1/Rkjov1sV3vU8daFc8JGyBJxkwDgVxNWwVFTnf1WFy+PC/ZLQlbAosfN9XF5VkecNREHmBlL40HTeLxibHGfrRhbtIdcRKDMEoLV0B42MfAWC5ULjcGEUyBseInVXnEqx1jw7R7i6xNepFmBu6kS8qLJ9eza9KxpOro2HTVNMV5YC1vcUnuOFykHtbqE561AR1duPkYnroJErlopNm0gpi3ErIEI0wTG1DnU2y4bOQGDp3NM8FiTzN3Gmv8BAzrY+uGvB2l2MzqTJPxuDW5bOkVReWDwK8rXH2BHOwP2i4Y2hUrBQLTHTkKWoiBq7MjMfGzf8erZ0e4SMyT3MiAW9CvCZp/HuTelvCSTpCDsyTmfdXsNdPeWZq5aDGtGM4o2ORBghlcY5A/fYG4PIjAzAIn7oOhY0QzvbnxtBM23RbvzatAPnvQHvw629neqZ4e34X3jfes9AJv19v3Dr2X3qnHvD+9v7y/vX86m52dzl7nlwp6985izlderXUO/wPCbXvq</latexit>

Optimality of splines

<latexit sha1_base64="QzIPrcPtj2xzr1QvCd7gcG4Xexc=">AJcHiclVZtb9s2EFa7bou9l6blwIFNm6pgSFoDMndmrVAgQApihQosC5d2iKREVDUWeZMvZSk4niEvX7Ov2Z/Y39gt2lOxUL8ngEpBN6B4+z/HueFSQCa606/5z7fpHNz7+5NONXv+z7/48ubmra9eqTSXDI5YKlL5JqAKBE/gSHMt4E0mgcaBgNfBbN/aX5+BVDxNftOLDMYxjRI+4YxqfHW6+Y2v8kABM79kmsdUcL0g6YSozBKq4nRzyx265SDdibecbO3dsrx4vTWxiM/TFkeQ6KZoEqdeG6mx4ZKzZmAou/nCjLKZjQCU7pfkAG+CsklfgkmpRvGzgaK7WIA0TGVE9V2ZfXmY7yfXk57HhSZrSFglNMkF0SmxsSAhl8C0WBDKGPqbU41+NBgUowLCx8OHo7GZgjgDjTuQkMCcpXFMk9D4ExpzsQhQnOhC+OryWreH/QHpEyBIglQS4EaPUI3/uIwQCXWzVxHqUiLIyMgsJ49zh7k/3KDkbBYtwhAqoHYXVnApFDJYOeNP0UNMAq0RBzNFWkFl0QX+YCjFWF82L1X7SIrfr6xEtv16TGzX8QtQ3WetSBwDx+SDwQ/j/UTQI9hers2JNUGL8qcK6gW2z4w4f4JmlLFMhSB3K5/U25xKuPt+xchKkKs0rpTZfS9zEXPNk8WVOvdXOnWVir848carudXaNpfViWdjYn+WnLYS05b+WTLu5y9JbCuRk2kSsZKo51nHtlMWgFibGoldoZ1nN+58L6iJU3ejMo673qe2lQu+UjZAl5mwDgVxPawVFTnv67Ck8PC/hLftrAgMIft/fHkuIk4biP2sTJWxv28bBm7HAf14wd2gMuApB6BcFuaA7aGHiLjcqmRmPBSRDGP8NuLzlNImz45jHimwuep1mB0ZQxed7menJhetI2PbswPWubQp7opS3sqD3Fq5KDrG/haSfRwbldj9kJqySR806JzVqIWQcxbyH8NIaIWocGO+UgL0Hj6YwIHmuS4f1CFP8DCGg2xNH3FeD9nUR6WlVeEoLdl8viTl9YXflKS1w9xRocjlruaBoUtWaBhY48RUNEw7me81Db9T+itTcgfEIWaU4SwJsQr0ka/p4rXfpbAu23wRwdS+f9QctdNeOZ7ZajBtGc4o2ORJghmY5wy2jPzYGAeg5QIJxkDSNMO7G782vPa3RXfyajT07g9Hv4629h5Unx3OhnPH+d75wfGcXWfPOXBeOEcOc945fzp/OX9v/Nu73fu2910FvX5tueZrpzF62/8BR+5wnQ=</latexit>
  • M. Unser, J. Fageot, J.P

. Ward, “Splines Are Universal Solutions of Linear Inverse Problems with Generalized-TV Regularization,” SIAM Review, vol. 59, No. 4, pp. 769-793, 2017.

  • M. Guerquin-Kern, M. H¨

aberlin, K.P . Pruessmann, M. Unser, ”A Fast Wavelet-Based Reconstruction Method for Mag- netic Resonance Imaging,” IEEE Transactions on Medical Imaging, vol. 30, no. 9, pp. 1649-1660, 2011.

  • E. Bostan, U.S. Kamilov, M. Nilchian, M. Unser, “Sparse Stochastic Processes and Discretization of Linear Inverse

Problems,” IEEE Trans. Image Processing, vol. 22, no. 7, pp. 2699-2710, 2013. K.H. Jin, M.T. McCann, E. Froustey, M. Unser, ”Deep Convolutional Neural Network for Inverse Problems in Imaging,” IEEE Trans. Image Processing, vol. 26, no. 9, pp. 4509-4522, Sep. 2017.

  • H. Gupta, K.H. Jin, H.Q. Nguyen, M.T. McCann, M. Unser, “CNN-Based Projected Gradient Descent for Consistent CT

Image Reconstruction,” IEEE Trans. Medical Imaging, vol. 37, no. 6, pp. 1440-1453, 2018.

  • M. Unser, “A representer theorem for deep neural networks,” J. Machine Learning Research, vol. 20, pp. 1-30, Jul.

2019.

■ Preprints and demos: http://bigwww.epfl.ch/

slide-22
SLIDE 22

Sketch of proof

43

⇒ ˜ f(xm) = ˜ ym,L

<latexit sha1_base64="Spnut0MkNbtcVFL0V1mRDaWc8=">ALS3iclVbtw2EFXSpk2UtnHSx74QtRdIA2ex2qBxejEQwIlhA07q2rET2DIMSprdpU1dQlL2rgl9TL+mTwXax35HX4qiDx1K2rUudrEhsAthzuGZ4Qw5pJdwJlWv9+eNmx9fOuT2/fse9+9vkX9xbuP9iXcSp82PNjHot3HpXAWQR7ikO7xIBNPQ4vPVO1wz+9gyEZH0Rk0SOArpMGID5lOFpuOFH5bcHTYcKSpEfE7c9ykNiKsYD0C7r8ge+juk/Fx+M3q1LhPJtmxDpe3sqXjhcVet5cP0v5wyo9Fqxzbx/fv/O0GsZ+GECmfUykPnV6ijQVivkcMtNJSTUP6VD0PnaMtJBU0AGscBfpEhurfFoKOUk9JAZUjWSTcwYr8SMRciBrHv1vDCzq5bDVA2eHWkWJamCyC8CGqScqJiYhJKACfAVnxDq+7iulCqMt6bwxjnSJngzvQZIn3IVrvf9Y/0CPgZKMQFRHDux2FIo0C7AxoyPglgQFOuMu3KwfTb7tgdkhdYEqQSL+UclPwe7S5ycD/kudJhOox5kGkx9DLtLPe6K98u97IWR8Ck5PS6SCh+SKsG4/EUCjcYST1OTj3cgwpChlAhatgZcUXKQRuvM7s6UfWUDbu51cuw51XG5f/QdomXNqexL+SEpQfr/adcV1Ahk6kncXnhWM020O5K4d+CRftzrPsUp9UzjXuecLNlFVBLPsoClyl94Nc5+RaPyuXfmZ5VyaXO/oydR1U3hIDvEs+BeOnukr9otDgrkf6WmSVGeIWIOAFl0rpZveJjXSdVLXoiZk2ax2dCU03g2tRXYMOYI3yTxTJ8Yue6pC6cUFEVniNWDNWUEwXzZOS9YDcBn1FOTDeMedEIql5YtJOZf+Ka1ud5eqe5QBYd1BkHTcYa7o4puNYEdypgS/ugArZkX1fA101wg3EPhJpSsGfqjSYH3mM7M4VTuCEFcO2e4aUiGI2GeK/o1cw09eqErTjJMNUiJFtNrRcz6EUT2pxBm0oYJEqsaDlbR2vawaiuoT1i7wLsx8LF1QVJBcoE6nThk3KOPWHm0yRItx3mC4cQhD2orZO23wTtvNr8HgLZEx5kQ1pmGEopnXAG9OpiZkduguJ7c/UWn5kAXh2c6OuQMDzjefwV7N/Fz/jTDJblD8hNTkl4VkoNLNPwR98m4QH9KEM4FRJgzCjPAzAh52Yzx5GRmPJkZN6OA+brcyqF2ZsC2iL3CLn2ht6fK6zFOd89ZACOqdM6aTXk5ToZcTJV1C+xA0QxZyFT+HCxO4/zQXZBYUseEuzlJMGXBZHsAgov4vDdiXguzAaqhF2exgrdIdvIu30fXzrVEHThyJTCnwE5Fj9pi+M6USKG0Uuv1GIRX1sr1YRoqCuGK8K3oNF+G7Y/9ftd50u393F98/qx8Nd62vrK+th5ajrViPbc2rG1rz/KtX6xfrd+tP+zf7L/sf+x/C+rNG+WcL63auHvrPy0gJQU=</latexit>

min

(U`),(n,`∈BV(2)(R))

M X

m=1

E

  • ym, f(xm)

N

X

`=1

R`(U`) + λ

L

X

`=1 N`

X

n=1

TV(2)(σn,`) !

<latexit sha1_base64="OS7zBy/Feh6hdmPvGkTWx0PnqM=">AMZniclVZfb9s2ELe7f62bdc2GYQ97IZYiNvUsNt6YFKJYmSIEmy5ImLRKmBiXRNhNSUkqsUPwi+2b7G2P2z7CnaUZEeSmyIlYIO6+/F3x7vjkV7MmdKdzp/1Wx9/Mmn92+07g79/m9L+7Pf3mokT69MCPeCRfe0RzkJ6oJnm9HUsKREep6+8s3Wnf3VOpWJR+FKPY3oiyCBkfeYTDaLefH1/EeGAqZiTsdJjTpHBgoU9s2Rwym48nlCLt9FBD1PObWsZTVWSBhYrNhCkZ8LlVG0RZiFwSIF+PbRvzNJKy7hvVbL2gbmtK+XEFaJ6Bmx1rVvtEG9thgCR+icU8sg5W+m496ouXkLYQlGwx1GzVxGIWJ8KhEGdEbfQYZHkdM64Y9xBe6mfZfdRwX9YxSE+AamsfJF/h+7D7GTobCcvpzt5385bmbOtxmLv/kKn3UkHmp108lCLR+7vfk7/+Eg8hNBQ+1zotRxtxPrE0OkZj6nELpE0Zj4Z2RATeqBRU0QBagfSfiFGqXSEo4IpcbCA6QgeqiqOid8p85JpOqrslXPE7ZRlBwnuv/kxLAwTjQN/cyhfsKRjpCrNBQwSX3Nx4j4PuwrIRr8LTG87J4Y57xbXlIon3AarLV/WjkxQ8rPqQa9pCG98CMhSBgY3CeC8XFA+yTh2hqs+pN5o9lorTyFQIo8hLOqVY/gxwDBg5KlkORDCIeWCMHnjXd5U579Yfljp3BSDrOMZ02ALIfwIrOuCLzIAnZT858eBwaiqYrZSkTDg1ziod2cZkYivMzvzNmXN3b8oN2/8gbheuG3J7HFL5ISEB+Pu4ywx6SFXiKSgvaGLWwCkdKqgd+sA86rR/hCXlSEOtc4WG5lX6m1CJF28WrLibKDrjFxrZ/XKzjTumoXj6w09nhgqmskM2GM4C/jK2APzrmrpAkH6l3O6EKURQu4AoIXu+krFm5qpGglTcTUSDXZbODS6Swg4/IroWHcwH0Xxdx91Eh5UZk4JrJIfANfwVWXTiBMg5H2gv2Y+oxw5LphxLNGULTCwj3r/uE6gdbneWavukEWHpUR1XEOlTHRLleVe4VlDPcRwXlDO1OQblTVW4xDheinkCgZ5qtKoa+hXbmEqehICXlBp/DpSIZCQdwr5g165p6cGLKLYmvfVeVLmeTVXPqrnU9Xzqipgoc51wYy1TXjHMCqLW9icqQLv0q2H1AVZBtEl8DTLkFEFMpqp0SpCziAuKgcCTogMz57ZxXc2WzqyD4DMkIYqIry8BDWY1rADcn02ObxTbIrid8uNAtGTDZ4ZmMJjqHAw73X4bej/0UP4lwDm6i9MTkoO2Msn+lFb9AnYwy7W8xqFMCKVJEJqZ0KqRp2pzw9HQqPJ0Kn4cB801eysJ0p4pdGXmZXPnS7E6YNyNYji9YQIdEmxQ1XbIxirMVUTxhNBvQAcKIM8E0PFwazUfpQPtUQ0seIOjlKIaXBVLskiKq/TaMBlYUHszhQA+h29ORBnPwJjLdx+0nPrx2imrXiUKXDHgGpLryXTF5gistgdxtFcqdTEs4ULxLVUI9wWuxW30bzk4OV9rgV+f3lYWn3+fvxtu1b2vf1Zq3dpq7Wltq7ZbO6j59T/qf9X/qf979+5e3Nfz32TQW/V8zVf1UpjDv0PtrOHNw=</latexit>

Apply “optimal” network ˜

f to each data point xm:

  • Initialization (input): ˜

ym,0 = xm.

  • For ` = 1, . . . , L

zm,` = (z1,m,`, . . . , zN`,m,`) = ˜ U` ˜ ym,`−1 ˜ ym,` = (˜ y1,m,`, . . . , ˜ yN`,m,`) ∈ RN`

with ˜

yn,m,` = ˜ n,`(zn,m,`) n = 1, . . . , N`.

<latexit sha1_base64="Hpt3jLgfAn9zJ92HvScHn1qr0=">ANMniclVdLbxs3EFbUV7J127g9sLWEvKAImgdpEmDCkiRB2Igcd0kdgJ7XYW7S0mMuY9wubZkgvf+mp4KtH+lvRW9gf02CH3oX0khbNABGbm4zfDmeFw7MaMJmI0+uNc973P/jwo/MXrI/XPvn0s4vrn+8lUco9sutFLOIvXJwQRkOyK6hg5EXMCQ5cRp67R3e1/vkx4QmNwmdiGZPDAM9COqUeFiCarHe/+j6O2RK9fBnFgaYXbqEQiJOIn6Eeo6gzCfSeYzQVPWQiBDB3hz5WGAURzQUANlDi0nQu205LpnRUFJBAnpKlOXoFdoKqaCY0VNjDl2mYZyK7etknoPLdVEBoORGudUQyvf+yDiwE8YG9sDx49EMnjUQ45jOQK7xvCp3qgBanwZ1vYg/98AZXAN2J7ITJbroyRdEzYpMtSolZH3FUZ1BkohZruacU1W/Wswj5qnSB3JBOjZdWfzJ2VZnui5aVLyKHhk59yqdJnPBgNb5DFYWZq9XNCxbwMnbYaFhzj4lCc+OWZEjoLsEZjNJBKndcV6n2EdhGdzM/NDqQfxJ6K8yObm4MRqOzIfaCztfbHTyb2eyfuFfoPTSgITCYzhJDuxRLA4l5oJ6TNdGmpAYe0d4RqRxW6E+iHw0hZRPI6grI63hcJAky8AFZIDFPGnqtPCNOi3hyTSpW3XdQFlVyUEqprcOpalPEnqZQ9OU6aLX9wb5lBNPwE3BngfnSrEAf2sMz+xDqZ3X2uKxMOM+OPht5uHck7YMRGg5yQkJ14UBhi7UxQNnSJ1OcMqEgdNibfWtPjL3OEARW7KGBHJbZA7gIFrnyU+SGcR85XkM1dB6Y2GN2/ApWphOFnmGCixQf4PYFVn9MXIzIAndT8ZdqHV6MpQtWuEHJ4yIrVslBWsVANZm3+7My5u2flhuO/E7cO1xm5XQapfJeQAPz/uOsMYk6S1E2gvKBHKgkNap5A7ZCr8tpo+A1sqUcap0xBK3ImEngGnPSW23Z1DbQ24y81c7NlZ0y7oKGy7cbul4YqprJDKgDuAvOythV+aZqsYHA/OScOkQmQkhfALRhv5m+YeGsRqpWTCJKI81k05lOp7aApM4vh4ZxBvd1FHP3kWV4UZ04xrxKfAZfwVWdTiA0wTC94GlMPHhSke6GEcsaQdUKvCVK/yJHtz7XlU+aB6Thfh2x30TcheolHebyicVZYt7v6Js0W5XlNtN5UPKXMJFAYGeKR82MeQ1tDOdOAEFyQmTzjE8KpzicAbvihwr3dSrGx5FsYJQ8wA9anLdK1X3mqtUrXVPkw9eQ6v2XtAUxlPDqER60qsA91fshdX6WQXQKP06ZNGALFo12kTwFuKkgXCigMxwy2f3qIE7aje/BoK1SBYQE9HYBh7yZlx9eDmpWKostn72PDl7G3bNgMwuT/H10TFcHj/MvT2DP4IsI5uI/MjclBjzPK6UobfAd1si0P8SgNgQ8MIhMTEgpJCZtWvjqVSl8VQq3Qp96Mi/lQNqlYodHbiZPC53CmYZQPpnFCfzLGQBlVub+Isx1RXDK+9ABwojRgAoYXKz+NfOhp0RAS54h6OUohskCJTCgISK8IXyWkxAY/8OZmEO3JwsB5mAmkvamB7NOVan7UKhTAUOA0dVfiuLPiURwoNYMw81GImEYVZXnQzdUITwaxoNyfD9mJvc2hfH45+3Ny4cyufGs93vux83bncsTs3O3c6Dzs7nd2O1/25+0v3t+7va7+u/bn219rfGbR7Lt/zRaf2rf3zH6yK0EY=</latexit>

This fixes two terms of minimal criterion: PM

m=1 E

  • ym, ˜

ym,L) and PL

`=1 R`( ˜

U`).

<latexit sha1_base64="gSDa4AunrB0Biqn9pNTkFniec+0=">ALq3iclVZtTxtHED6nb4ndNqT92C+rglUageVz1JGRYpEQCBSiGQCI5a9zK2F/ZesrsHNqv7Jf1l/Sf92Nl7MfcClbOS0TLzDNzM7Oz60SMCtnv/9N69NnX3z51eMn7c7X3z7dOnZd6cijLkLJ27IQv7BsQUwGsCJpJLBh4iD7TsM3jtXW1r/hq4oGHwTs4iuPDtcUBH1LUlioZLf7+bUEFGdAqCyJuQSOC+IOGI+DSgvs2IynKEPuqTVYsEftD5W+ayV8HZNty6HjVOiWzob9mSco8UPq/BFr+8nPK8QOPFIYWcCYtsnR0O9X1VWGr1yWAxJYX5ATpJUjea94dJyv9dPF2luzHyzbOTrcPjsyb+WF7qxD4F0mS3EudmP5IWyuaQug6RtxQIi272yx6BS7wnposgjo5DjL5AklVZwti/EzHcQ6dtyIuo6LbxXpyVcjETVq+P4SbsOY/l6OWFokEUSwjcLKBRzIjEYmDBiEc5uJLNiO26+F2xLTHeCsM780Lp4LV5RSFcm4G32ftcKEmwK5Bop5DADdu6PtYHWNbJ+ymQcjO2YyUZYFft2t90laQOJtJBOzBhI8QrlFmKw37L6+fE4ZF6i+NhJlLnW7238stZPGhgOsxzT7yEg+yGsHIxuhcwNRlKNk9kO9rgEnyaVxiEWjxko7RWmSbvYJDVm7X5x5jzcRbnx8z+JW6drQW6HYSk/JSUI/z/uKoOcgIgdge2F5ztRFkTgb0Dz9V6v/crmlQzjb3OGFlpZ1GJj7HNYeXOZKB9kIecPOhn487PO+SBrOHb0oHJXdZA6SczwL1p2z5+q+bjGRIP2Tc+oUpRki+gCQZfN+pqHRZ2UvaSFmDupF5uOdTm1B6J0fTkOjAXC1nMwyftlJdUiSObl4kXiBVD1eVEwjQZ6Sw4jsCleCPoaRiybBCUvdDgKNF/iaVHn+Oo/oH0uCsijirI7awOwrlVl15VFI2uM9Kygbt25LybV25S5kDXBYQnJlqt46BjzjOdOEkNiQHpqxrvFQ4tYMx3itqM9FDvWywH0YJpr7ZL/O9WauelNX7c1Ve3WVRwOZ67yGtx18DlDg5U/YaXSBc6vtsXReVkFyizdKmRag0wbPVpH8AbipoawQh/GdiNm56qGu2oOvxqCNUimBNZM8MIeT2vHt6cVM6SLdedj1Zp8tmxYHKDk+xuQaDzjefxn6OHJTfJHhHNwl6YnJQcZ5ehO6/+OfTLNtH9EqE4JuJ8iMjHAXAhp2bTw8nIuvJwL9wKPuipvZV+Zc8UhD51MLlyuDgvmnRDNrRvqwcSWKkXNTbanUWYRgWj2sYJEISM+lTiw6XdXU8XOQaJI3lMcJaTCF8WRNBbICDdHq62JQDfncFYTnDaw1SiO3wTKXPg4lunrNRzKNClwEdAqveFMU7VkiO1JqhN6gVUtpOUro+9EBFIvwifCua9Zdhc3M6Jkvev0/B8uvX+avxsfGD8aPxqphGhvGa2PXODRODLfVav3U6rfMznrnuHPWsTLo1Zu871RWR34D9FzQtQ=</latexit>

˜ f achieves global optimum , ˜ σn,` = arg min

f∈BV(2)(R) kD2fkM

s.t.

f(zn,m,`) = ˜ yn,m,`, m = 1, . . . , M

<latexit sha1_base64="WoH25LpINGrVpMxIcYIAutNTp4=">AMHniclVZb9s2FLa7W+usW7s97oVYiAtUsNysbUbFqBYL2iBpsvSNi0apgElHclsqEtJKrHD8o/sacD2W/Y27HX7J3vcoSQ7stwOqQzbwjkfv+/w8PCQfi640sPhP91zH3z40cefnL/QW/n04mefX7r8xa7KChnA0yATmXzuMwWCp/BUcy3geS6BJb6AZ/7hbed/dgRS8Sx9oqc57CcsTnEA6bRdHC5e3GNai5CMHSLRHaNsGDM4QgUiUXmM0GyXPOkSCjd82Cy31ujIVe5YFOlpwIfQiRljweayZldkzo64KFxNAyMCMhtBU7oYrHCTsw6QYFIazdpEzGhCY8PTARoTzFQTIhP+7al2Z9dMWu050rltA3hN7J8pcjEtE3B4Y+zgOyZUsRmvjZxBC1EAP8M/W2tH6iVNJKp0rm7X+tGHcIBQ/ya3QcNMq40tnFU9vYNLq8PBsHzI8otXv6x26mf74PKFf5EkKBJIdSCYUnveMNf7hknNAwG2RwsFOQsOWQymzIolfTRhoJnEb6pJaV3AsUSpaeIjMmF6rNo+Z3yrz1mkitSiqu8nte07BU6urlveJoXGtKgCigqBNEZcTVCQi4h0GKxRDgvAqmMd4FhifevnHBu+ELDhUwAeHm4LvRvhmDOAKNfgkpHAdZkrA0NDRiCRfTECJWCG0NVdHsvdfv9UlZs4oglPiFEKDV92iniMESr+oqKeJMhNbI2LfG2xgObnyzMbRLGAnTGjMcIKD6IqwZjC8KqGQwksU4BfNxW2lIuJ0VtENjoclCgHGqMLG92YtMTv5szPX4Z6VG6f/XtwuXWfk9gUu5fukBOH/x73IoMegCl9heWH7sQZ3/Vh7cBVc204+BaHLGYa10IstarolK4wyWsnQ4ZOQ3yLpF36tw41ZnXfN0+m6h6zOhpkwlYPdwL9BTsavmbdWC3cWPzWnS1GZIeI2AFn13k7fUjirSFOlXIi5SHuxeyW0ykQ49ZXYsM4Q/gui3X4pFfykXinMkm8RlixVDdciJhmYyFzOIeB4CLlumImqETRVeLpj3S8eJNj6fN/stCfI0xeLiBdtxG2sjpnzdtu503Aucb9oOJdoHzWcj9rO+1z4IPUMgj3T3G9j4DW2M7dwGgtSgjD0CA8VyVka47liNq1r6s0BD7PcmvIUfdjmujN3Wm7HsxdD9qukKe69oVLavfwBsJBNqdwb6kK/BM3HpcurFaQnCBPfxEyaUEmSzXaRsglxHELQbMEYrYUs3/Ywh0uN78WQiyRTDAnujUMI5TtvIZ4cnI9tVuw+p4orur3oKAqTbP7OmTI9zgeP5VaLzvlPhZhmtwn5Q7pgZtVZTRqTf5AetkUnl/ytFdEsikRFRmgLkRymVzxlev5sZXc+ODNOSBqUs5Md7csS0zv7KrQJrtGfO9DIfTYx7CmGlTouZD7k7yakSWzxjNXewAaSZ4wjVeXHr9a+VDHoPGlhwT7OUkx5sFUfwECOhgE+PKsCrbhrMXZ7mGiUwzuR8UYB3nWaTteHUrcUeAkofYsnxezqrLREascwGLUWUjPfNo4P1CRCGeEd0WvfTNcftkdDbzrg+HPo9VbN+tb4/nOV52vO+sdr3Ojc6tzv7PdedoJur7S/e37u8rv678sfLnyl8V9Fy3HvNlZ+FZ+fs/oMFt+w=</latexit>

Optimal solution ˜

f = ˜ L ˜ `L ˜ L−1 · · · ˜ `2 ˜ 1 ˜ `1 with optimized weights ˜ U` and neuronal activations ˜ σn,`.