CNN Case Studies
- M. Soleymani
Sharif University of Technology Fall 2017 Slides are based on Fei Fei Li and colleagues lectures, cs231n, Stanford 2017, and some are adopted from Kaiming He, ICML tutorial 2016.
CNN Case Studies M. Soleymani Sharif University of Technology Fall - - PowerPoint PPT Presentation
CNN Case Studies M. Soleymani Sharif University of Technology Fall 2017 Slides are based on Fei Fei Li and colleagues lectures, cs231n, Stanford 2017, and some are adopted from Kaiming He, ICML tutorial 2016. AlexNet [Krizhevsky, Sutskever,
Sharif University of Technology Fall 2017 Slides are based on Fei Fei Li and colleagues lectures, cs231n, Stanford 2017, and some are adopted from Kaiming He, ICML tutorial 2016.
[Krizhevsky, Sutskever, Hinton, 2012]
– AlexNet – VGG – GoogLeNet – ResNet
– Wide ResNet – ResNeXT – Stochastic Depth – FractalNet – DenseNet – SqueezeNet
Input: 227x227x3 First layer (CONV1): 11x11x3 stride 4 => Output: (227-11)/4+1 = 55 Parameters: (11*11*3)*96 = 35K
Second layer (POOL1): 3x3 filters stride 2 Output volume: 27x27x96 #Parameters: 0!
Input: 227x227x3 After CONV1: 55x55x96 After POOL1: 27x27x96
Details/Retrospectives:
val accuracy plateaus
Historical note: Trained on GTX 580 GPU with
2 GPUs, half the neurons (feature maps) on each GPU.
[Zeiler and Fergus, 2013]
[Simonyan and Zisserman, 2014]
– 8 layers (AlexNet) -> 16 - 19 layers (VGG16Net)
– stride 1, pad 1
layers has same effective receptive field as one 7x7 conv layer
– 3 ∗ (32𝐷2) vs. 72𝐷2 for C channels per layer
[Simonyan and Zisserman, 2014]
– ILSVRC’14 2nd in classification, 1st in localization – Similar training procedure as Krizhevsky 2012 – No Local Response Normalisation (LRN) – Use VGG16 or VGG19 (VGG19 only slightly better, more memory) – Use ensembles for best results – FC7 features generalize well to other tasks
– 22 layers – Efficient “Inception” module – No FC layers – Only 5 million parameters!
– ILSVRC’14 classification winner (6.7% top 5 error)
[Szegedy et al., 2014]
Inception module: a good local network topology (network within a network) GoogLeNet stack these modules
[Szegedy et al., 2014]
– Multiple receptive field sizes for convolution (1x1, 3x3, 5x5) – Pooling operation (3x3)
[Hint: Computational complexity] Example
[Hint: Computational complexity]
– Q1: What is the output size of the 1x1 conv, with 128 filters?
Example
[Hint: Computational complexity]
– Q1: What is the output size of the 1x1 conv, with 128 filters? – Q2: What are the output sizes of all different filter operations?
Example
[Hint: Computational complexity]
– Q1: What is the output size of the 1x1 conv, with 128 filters? – Q2: What are the output sizes of all different filter operations? – Q3:What is output size after filter concatenation?
Example
Example
– [1x1 conv, 128] 28x28x128x1x1x256 – [3x3 conv, 192] 28x28x192x3x3x256 – [5x5 conv, 96] 28x28x96x5x5x256 – Total: 854M ops
– Pooling layer also preserves feature depth, which means total depth after concatenation can only grow at every layer!
use 1x1 convolutions to reduce feature depth Example
– [1x1 conv, 64] 28x28x64x1x1x256 – [1x1 conv, 64] 28x28x64x1x1x256 – [1x1 conv, 128] 28x28x128x1x1x256 – [3x3 conv, 192] 28x28x192x3x3x64 – [5x5 conv, 96] 28x28x96x5x5x64 – [1x1 conv, 64] 28x28x64x1x1x256
version Bottleneck can also reduce depth after pooling layer
[Szegedy et al., 2014]
[Szegedy et al., 2014]
– 22 layers – Efficient “Inception” module – No FC layers – Only 5 million parameters!
– ILSVRC’14 classification winner (6.7% top 5 error)
[Szegedy et al., 2014]
[He et al., 2015]
– 152-layer model for ImageNet – ILSVRC’15 classification winner (3.57% top 5 error) – Swept all classification and detection competitions in ILSVRC’15 and COCO’15!
convolutional neural network?
[He et al., 2015]
convolutional neural network?
– A deeper model should not have higher training error – The deeper model performs worse, but it’s not caused by overfitting!
[He et al., 2015]
are harder to optimize
shallower model.
– A solution by construction is copying the learned layers from the shallower model and setting additional layers to identity mapping.
[He et al., 2015]
directly trying to fit a desired underlying mapping H(x)
[He et al., 2015]
H(x)=F(x)+x Use layers to fit residual F(x) = H(x) - x instead of H(x) directly
– Stack residual blocks – Every residual block has two 3x3 conv layers – Periodically, double # of filters and downsample spatially using stride 2 (/2 in each dimension)
[He et al., 2015] 128 filters, spatially with stride 2 64 filters, spatially with stride 1
– Stack residual blocks – Every residual block has two 3x3 conv layers – Periodically, double # of filters and downsample spatially using stride 2 (/2 in each dimension) – Additional conv layer at the beginning
[He et al., 2015] Beginning conv layer
– Stack residual blocks – Every residual block has two 3x3 conv layers – Periodically, double # of filters and downsample spatially using stride 2 (/2 in each dimension) – Additional conv layer at the beginning – No FC layers at the end (only FC 1000 to output classes)
[He et al., 2015]
No FC layers besides FC 1000 to output classes Global average pooling layer after last conv layer
[He et al., 2015]
Total depths of 34, 50, 101, or 152 layers for ImageNet For deeper networks (ResNet-50+), use “bottleneck” layer to improve efficiency (similar to GoogLeNet)
[He et al., 2015]
For deeper networks (ResNet-50+), use “bottleneck” layer to improve efficiency (similar to GoogLeNet)
[He et al., 2015]
– Batch Normalization after every CONV layer – Xavier/2 initialization from He et al. – SGD + Momentum (0.9) – Learning rate: 0.1, divided by 10 when validation error plateaus – Mini-batch size 256 – Weight decay of 1e-5 – No dropout used
– Not explicitly address generalization, but deeper+thinner shows good generalization
Kaiming He, Xiangyu Zhang, Shaoqing Ren, & Jian Sun. “Deep Residual Learning for Image Recognition”. CVPR 2016.
Kaiming He, Xiangyu Zhang, Shaoqing Ren, & Jian Sun. “Deep Residual Learning for Image Recognition”. CVPR 2016.
[He et al., 2015]
ILSVRC 2015 classification winner (3.6% top 5 error) better than “human performance”! (Russakovsky 2014)
– Able to train very deep networks without degrading (152 layers on ImageNet, 1202 on Cifar) – Deeper networks now achieve lowing training error as expected – Swept 1st place in all ILSVRC and COCO 2015 competitions
Inception-v4: Resnet + Inception! VGG: Highest memory, most operations GoogLeNet: most efficient AlexNet: Smaller ops, still memory heavy, lower accuracy
Inception-v4: Resnet + Inception! ResNet: Moderate efficiency depending on model, highest accuracy
[He et al. 2016]
Identity Mappings in Deep Residual Networks
throughout network (moves activation to residual mapping pathway)
𝑦[𝑚] 𝑦[𝑚+1] 𝑦[𝑚] 𝑦[𝑚+1] ReLU could block back prop for very deep networks
Pre-activation ResNet
Kaiming He, Xiangyu Zhang, Shaoqing Ren, & Jian Sun. “Deep Residual Learning for Image Recognition”. CVPR 2016.
+ 𝐺 𝑦 𝑚+1
𝑀−1 𝐺 𝑦 𝑗
𝜖𝑦[𝑚] = 𝜖𝐹 𝜖𝑦[𝑀] 𝜖𝑦[𝑀] 𝜖𝑦[𝑚] = 𝜖𝐹 𝜖𝑦 𝑀
1 +
𝜖 𝜖𝑦 𝑚 𝑗=𝑚 𝑀−1 𝐺 𝑦 𝑗
Any
𝜖𝐹 𝜖𝑦[𝑀] is directly back-prop to any 𝜖𝐹 𝜖𝑦[𝑚], plus residual.
Any
𝜖𝐹 𝜖𝑦[𝑚] is additive; unlikely to vanish.
Kaiming He, Xiangyu Zhang, Shaoqing Ren, & Jian Sun. “Deep Residual Learning for Image Recognition”. CVPR 2016.
[Zagoruyko et al. 2016]
Wide Residual Networks
(parallelizable)
[Xie et al. 2016]
Aggregated Residual Transformations for Deep Neural Networks (ResNeXt)
multiple parallel pathways (“cardinality”)
[Huang et al. 2016]
Deep Networks with Stochastic Depth
reduce vanishing gradients and training time through short networks during training
during each training pass
[Larsson et al. 2017]
is transitioning effectively from shallow to deep
– residual representations are not necessary
deep paths to output
FractalNet: Ultra-Deep Neural Networks without Residuals
[Huang et al. 2017]
Densely Connected Convolutional Networks
connected to every other layer in feedforward fashion
vanishing gradient, strengthens feature propagation, encourages feature reuse
[Iandola et al. 2017]
SqueezeNet:
“expand” layer with 1x1 and 3x3 filters
– AlexNet – VGG – GoogLeNet – ResNet
– Improvement of ResNet
– DenseNet – SqueezeNet
and improving gradient flow
width and residual connections
task (using a large amount of data) in another related task (for which you don’t have enough data)
Donahue et al, “DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition”, ICML 2014. Razavian et al, “CNN Features Off-the-Shelf: An Astounding Baseline for Recognition”, CVPR Workshops 2014.
CVPR, 2016.