Nonparametric regression using deep neural networks with ReLU activation function
Johannes Schmidt-Hieber February 2018 Caltech
1 / 20
Nonparametric regression using deep neural networks with ReLU - - PowerPoint PPT Presentation
Nonparametric regression using deep neural networks with ReLU activation function Johannes Schmidt-Hieber February 2018 Caltech 1 / 20 Many impressive results in applications . . . Lack of theoretical understanding . . . 2 / 20
1 / 20
2 / 20
3 / 20
4 / 20
◮ version of ResNet with 152 hidden layers ◮ networks become deeper
◮ AlexNet uses 60 million parameters for 1.2 million training
5 / 20
6 / 20
◮ Xi ∈ Rd, Yi ∈ R, ◮ goal is to reconstruct the function f : Rd → R
7 / 20
◮ architecture (L, p) ◮ number of active (e.g. non-zero) parameters is s
8 / 20
9 / 20
◮ few letters in one word ◮ few word in one sentence 10 / 20
◮ gi : Rdi → Rdi+1. ◮ each of the di+1 components of gi is βi-smooth and depends
◮ ti can be much smaller than di ◮ we show that the rate depends on the pairs
11 / 20
◮ one function that can be non-smooth but every component is
◮ one function that has high-dimensional input but the function
12 / 20
13 / 20
ti 2β∗ i +ti log n
2β∗ i 2β∗ i +ti log2 n. 14 / 20
2β∗ i 2β∗ i +ti log2 n.
15 / 20
◮ here the number of active parameters 16 / 20
17 / 20
◮ shows the trade-off between approximation and the number of
◮ builds on work by Telgarsky (2016), Liang and Srikant (2016),
◮ network parameters bounded by one ◮ explicit bounds on network architecture and sparsity 18 / 20
2β 2β+1 log2 n.
19 / 20
20 / 20