NAS NAS - - PowerPoint PPT Presentation

nas nas
SMART_READER_LITE
LIVE PREVIEW

NAS NAS - - PowerPoint PPT Presentation

NAS NAS NAS NAS Deep learning vs non-deep learning Automatically learn features from data Achieved


slide-1
SLIDE 1

快速有效的NAS和基于NAS启发的模型压缩

欧阳万里

slide-2
SLIDE 2

大纲

  • 简介
  • 快速有效的NAS
  • 基于NAS启发的模型压缩
  • 结论
slide-3
SLIDE 3

Deep learning vs non-deep learning

  • Automatically learn features from data

Achieved by deep learning

slide-4
SLIDE 4

Deep learning – not fully automatic

  • Automatically learn features from data
  • Number of layers?
  • Number of channels at each layer?
  • What kind of operation in each layer?
  • How one layer is connected to another layer?
  • Data preparation?
  • Objective/Loss function?

Manual tuning is required Automatically learning them is possible by AutoML Achieved by deep learning

slide-5
SLIDE 5

AutoML

  • The problem of automatically (without human input) producing test set

predictions for a new dataset within a fixed computational budget [a].

  • Target: low error rate with low computational budget (高精度+高效率)

[a] Feurer, Matthias, Aaron Klein, Katharina Eggensperger, Jost Springenberg, Manuel Blum, and Frank Hutter. "Efficient and robust automated machine learning." In Advances in neural information processing systems, pp. 2962-2970. 2015.

slide-6
SLIDE 6

AutoML – Our works

  • NAS:
  • Dongzhan Zhou*, Xinchi Zhou*, Wenwei Zhang, Chen Change Loy, Shuai YI, Xuesen

Zhang, W. Ouyang, "EcoNAS: Finding Proxies for Economical Neural Architecture Search", CVPR, 2020.

  • Xiang Li, Chen Lin, Chuming Li, Ming Sun, Wei Wu, Junjie Yan, W. Ouyang, "Improving

One-shot NAS by Suppressing the Posterior Fading", CVPR, 2020.

  • Liang F, Lin C, Guo R, Sun M, Wu W, Yan J, Ouyang W. “Computation Reallocation for

Object Detection", ICLR, 2020.

  • Data Augmentation:
  • Chen Lin, Minghao Guo, Chuming Li, Xin Yuan, Wei Wu, Junjie Yan, Dahua Lin, W.
  • Ouyang. "Online Hyper-parameter Learning for Auto-Augmentation Strategy", Proc.

ICCV, 2019.

  • Loss:
  • Chuming Li, Xin Yuan, Chen Lin, Minghao Guo, Wei Wu, Junjie Yan, W. Ouyang. "AM-LFS:

AutoML for Loss Function Search", Proc. ICCV, 2019.

slide-7
SLIDE 7

AutoML – Our works

  • NAS:
  • Dongzhan Zhou*, Xinchi Zhou*, Wenwei Zhang, Chen Change Loy, Shuai YI, Xuesen

Zhang, W. Ouyang, "EcoNAS: Finding Proxies for Economical Neural Architecture Search", CVPR, 2020.

  • Xiang Li, Chen Lin, Chuming Li, Ming Sun, Wei Wu, Junjie Yan, W. Ouyang, "Improving

One-shot NAS by Suppressing the Posterior Fading", CVPR, 2020.

  • Liang F, Lin C, Guo R, Sun M, Wu W, Yan J, Ouyang W. “Computation Reallocation for

Object Detection", ICLR, 2020.

  • Data Augmentation:
  • Chen Lin, Minghao Guo, Chuming Li, Xin Yuan, Wei Wu, Junjie Yan, Dahua Lin, W.
  • Ouyang. "Online Hyper-parameter Learning for Auto-Augmentation Strategy", Proc.

ICCV, 2019.

  • Loss:
  • Chuming Li, Xin Yuan, Chen Lin, Minghao Guo, Wei Wu, Junjie Yan, W. Ouyang. "AM-LFS:

AutoML for Loss Function Search", Proc. ICCV, 2019.

slide-8
SLIDE 8

Network Architecture Search (NAS)

  • Automatically search the suitable network architecture for specific

tasks

  • Time consuming
slide-9
SLIDE 9

Ops:

Search Space

Network Structure (from DARTS [b])

3x3 avg pooling 3x3 Separable Conv Identity 5x5 Separable Conv 3x3 max pooling 3x3 Dilated Conv zero 5x5 Dilated Conv

[b] Liu, H., Simonyan, K., & Yang, Y. Darts: Differentiable architecture search. ICLR 2019.

248= 110,075,314,176 ~ 1 × 1011

slide-10
SLIDE 10

Search Space

  • Possible choices for 24 layers with 8 operations per layer:
  • 248= 110,075,314,176 ~ 1 × 1011
  • Suppose each choice requires 1 hour:
  • About 12,000,000 ~ 12 million years

Architecture GPU Days Method NASNet-A [c] 1800 Reinforcement Learning AmoebaNet-A [d] 3150 Evolution

[c] Zoph, B., Vasudevan, V., Shlens, J., Le, Q.V.: Learning transferable architectures for scalable image recognition. In: CVPR (2018)

[d] Real, Esteban, et al. “Regularized evolution for image classifier architecture search.” In: AAAI. 2019.

3x3 avg pooling 3x3 Separable Conv Identity 5x5 Separable Conv 3x3 max pooling 3x3 Dilated Conv zero 5x5 Dilated Conv

slide-11
SLIDE 11

提纲

  • 简介
  • 快速有效的NAS(高效率搜索)
  • 基于NAS启发的模型压缩(高效率部署)
slide-12
SLIDE 12

EcoNAS: Finding Proxies for Economical Neural Architecture Search

Dongzhan Zhou, Xinchi Zhou, Wenwei Zhang, Chen Change Loy, Shuai Yi, Xuesen Zhang, Wanli Ouyang CVPR 2020

slide-13
SLIDE 13

Motivation

  • Too time consuming

Architecture GPU Days Method NASNet-A [b] 1800 Reinforcement Learning AmoebaNet-A [c] 3150 Evolution

[b] Zoph, B., Vasudevan, V., Shlens, J., Le, Q.V.: Learning transferable architectures for scalable image recognition. In: CVPR (2018)

[c] Real, Esteban, et al. “Regularized evolution for image classifier architecture search.” In: AAAI. 2019.

EcoNAS: Finding Proxies for Economical Neural Architecture Search

slide-14
SLIDE 14

Proxy

  • A proxy is a computationally reduced setting, e.g.
  • Reduced number of training epochs
  • Compared with the original network, the proxy has the same
  • Operation
  • Number of layers
  • Relative ratio for the numbers of channels between two layers

Dongzhan Zhou, Xinchi Zhou, Wenwei Zhang, Chen Change Loy, Shuai Yi, Xuesen Zhang, and Wanli Ouyang. "EcoNAS: Finding Proxies for Economical Neural Architecture Search." CVPR 2020.

EcoNAS: Finding Proxies for Economical Neural Architecture Search

3x3 avg pooling 3x3 Separable Conv Identity 5x5 Separable Conv 3x3 max pooling 3x3 Dilated Conv zero 5x5 Dilated Conv

Computation 1 Training Epochs (e) 600 300 150 75

slide-15
SLIDE 15

Proxy

  • A proxy is a computationally reduced setting, e.g.
  • Reduced number of training epochs
  • Reduced input resolution
  • Reduced number of channels
  • Reduced number of samples
  • Compared with the original network, the proxy has the same
  • Operation
  • Number of layers
  • Relative ratio for the numbers of channels between two layers

Dongzhan Zhou, Xinchi Zhou, Wenwei Zhang, Chen Change Loy, Shuai Yi, Xuesen Zhang, and Wanli Ouyang. "EcoNAS: Finding Proxies for Economical Neural Architecture Search." CVPR 2020.

EcoNAS: Finding Proxies for Economical Neural Architecture Search

3x3 avg pooling 3x3 Separable Conv Identity 5x5 Separable Conv 3x3 max pooling 3x3 Dilated Conv zero 5x5 Dilated Conv

slide-16
SLIDE 16

Proxy

  • A proxy is a computationally reduced setting, e.g.
  • Reduced number of training epochs
  • Reduced input resolution
  • Reduced number of channels
  • Reduced number of samples
  • Compared with the original network, the proxy has the same
  • Operation
  • Number of layers
  • Relative ratio for the numbers of channels between two layers

EcoNAS: Finding Proxies for Economical Neural Architecture Search

3x3 avg pooling 3x3 Separable Conv Identity 5x5 Separable Conv 3x3 max pooling 3x3 Dilated Conv zero 5x5 Dilated Conv

[19] [23] [17, 19 31]

[7] Boyang Deng, Junjie Yan, and Dahua Lin. Peephole: redicting network performance before training. CoRR, abs/1712.03351, 2017. [17] Dmytro Mishkin, Nikolay Sergievskiy, and Jiri Matasa. Systematic evaluation of cnn advances on the imagenet. CVIU, 2017. [23] Kailas Vodrahalli, Ke Li, and Jitendra Malik. Are all training examples created equal? an empirical study. CoRR, abs/1811.12569, 2018 [19] Esteban Real, Alok Aggarwal, Yanping Huang, and Quoc V. Le. Regularized evolution for image classifier architecture search. In AAAI, 2019. [31] Barret Zoph, Vijay Vasudevan, Jonathon Shlens, and Quoc V. Le. Learning transferable architectures for scalable image recognition. In CVPR, 2018.

slide-17
SLIDE 17

What is a good proxy?

Reliable Fast

slide-18
SLIDE 18

This paper

  • A systematic and empirical study on the proxy
  • Appropriate use of proxy can
  • Make NAS fast
  • Get architectures with better accuracy
slide-19
SLIDE 19

Existing proxies behave differently in maintaining rank consistency.

Real Ranking Ranking in Proxy 1 Ranking in Proxy 2 Network A 1 1 3 Network B 2 2 4 Network C 3 3 1 Network D 4 4 2 Good Proxy Bad Proxy

Example: Finding reliable proxies is important for Neural Architecture Search.

Dongzhan Zhou, Xinchi Zhou, Wenwei Zhang, Chen Change Loy, Shuai Yi, Xuesen Zhang, and Wanli Ouyang. "EcoNAS: Finding Proxies for Economical Neural Architecture Search." CVPR 2020.

Proxy – reliability

slide-20
SLIDE 20

Spearman Coefficient of original ranking (Ground-Truth Setting) and proxy ranking (reduced setting). ⚫ Value range [-1, 1], higher absolute value indicates stronger correlation. ⚫ Positive value for positive correlation, vice versa.

How to evaluate the reliability of Proxies?

A model sampled from the search space

slide-21
SLIDE 21

With the same iteration numbers, using more training samples with fewer training epochs could be more effective than using more training epochs and fewer training samples.

Influence of sample ratio (s) and epochs (e)

slide-22
SLIDE 22

With the same iteration numbers, using more training samples with fewer training epochs could be more effective than using more training epochs and fewer training samples.

Influence of sample ratio (s) and epochs (e)

60 epochs, 100 iters per epoch 120 epochs, 50 iters per epoch

slide-23
SLIDE 23

Reducing the resolution of input images is sometimes feasible Reducing the number of channels of networks is more reliable than reducing the resolution.

cxrys0e60 c0rxs0ey cxr0s0ey

Influence of channels (c) and resolution (r)

Dongzhan Zhou, Xinchi Zhou, Wenwei Zhang, Chen Change Loy, Shuai Yi, Xuesen Zhang, and Wanli Ouyang. "EcoNAS: Finding Proxies for Economical Neural Architecture Search." CVPR 2020.

slide-24
SLIDE 24

An efficient proxy does not necessarily have a poor rank consistency.

Efficient proxies

Dongzhan Zhou, Xinchi Zhou, Wenwei Zhang, Chen Change Loy, Shuai Yi, Xuesen Zhang, and Wanli Ouyang. "EcoNAS: Finding Proxies for Economical Neural Architecture Search." CVPR 2020.

slide-25
SLIDE 25

Use reduction factors for training

Original setting: Conv(l): 36 × c channels Conv(l+1): 36 × c channels Input resolution: 32 ×32 Training data: 50000 Training epochs: 600 Proxy: 9 × c 9 × c 8×8 50000 60 r4 e60 c4 s0

Proxy – Example

slide-26
SLIDE 26
  • 1. Select a fast and reliable proxy.
  • 2. Hierarchical proxy strategy: train networks with different proxies based on their accuracy.

EcoNAS: Economical evolutionary-based NAS

Dongzhan Zhou, Xinchi Zhou, Wenwei Zhang, Chen Change Loy, Shuai Yi, Xuesen Zhang, and Wanli Ouyang. "EcoNAS: Finding Proxies for Economical Neural Architecture Search." CVPR 2020.

slide-27
SLIDE 27

Continue training for E epochs

Models trained for E epochs ... ... ...

Continue training for E epochs

Models trained for 2E epochs Models trained for 3E epochs Models randomly initialized Train for E epochs

slide-28
SLIDE 28

Continue training for E epochs

Models trained for E epochs

Generate child networks

... ... ...

Train for E epochs Continue training for E epochs

Models trained for 2E epochs Models trained for 3E epochs

slide-29
SLIDE 29
  • 1. Select a more efficient and consistent reduced setting as proxy.
  • 2. Hierarchical proxy strategy: train networks with different proxies based on their accuracy.

EcoNAS: Economical evolutionary-based NAS

Setting: Three population sets PE, P2E, P3E, which store networks trained for E, 2E, 3E epochs, respectively. For each cycle: Step 1. A batch of networks are randomly sampled from PE, P2E, P3E and mutated. Networks with higher accuracy are more likely to be chosen. Train the mutated networks for E epochs and add them to PE.

Zhou, Dongzhan, Xinchi Zhou, Wenwei Zhang, Chen Change Loy, Shuai Yi, Xuesen Zhang, and Wanli Ouyang. "EcoNAS: Finding Proxies for Economical Neural Architecture Search." CVPR 2020.

Step 2. Choose top networks from PE, P2E, load from checkpoints and train E more epochs, then add to P2E, P3E. Step 3. Remove dead networks from all populations.

slide-30
SLIDE 30

Experimental results on CIFAR-10

slide-31
SLIDE 31

Not only save searching costs. But also save re-training costs. Reliable proxies need not many networks to be re-trained.

Method Number of Re-training Networks BlockQNN 100 NASNet 250 AmoebaNet 20 EcoNAS (ours) 5

EcoNAS analysis

Zhou, Dongzhan, Xinchi Zhou, Wenwei Zhang, Chen Change Loy, Shuai Yi, Xuesen Zhang, and Wanli Ouyang. "EcoNAS: Finding Proxies for Economical Neural Architecture Search." CVPR 2020.

slide-32
SLIDE 32

Provides more diverse structures, which allows searching algorithms to find accurate structures with fewer costs.

Method Network numbers BlockQNN 11k NASNet 45k AmoebaNet 20k EcoNAS (ours) 1k

EcoNAS analysis

Zhou, Dongzhan, Xinchi Zhou, Wenwei Zhang, Chen Change Loy, Shuai Yi, Xuesen Zhang, and Wanli Ouyang. "EcoNAS: Finding Proxies for Economical Neural Architecture Search." CVPR 2020.

slide-33
SLIDE 33
  • 1. Reliable proxy and hierarchical proxy strategy will reduce both searching cost and error rate.

Reduced Setting (w/o hierarchical proxy) Cost (GPU days) Spearman Coefficient Params. (M) Error Rate (%) AmoebaNet 3150 0.70 3.20 3.34 ± 0.06 C4r4s0e35 (ours) 12 0.74 3.18 2.94

EcoNAS ablation study on CIFAR-10

Reduced Setting (w. hierarchical proxy) Cost (GPU days) Spearman Coefficient Params. (M) Error Rate (%) NASNet Proxy 21 0.65 2.89 3.20 C3r2s1e60 12 0.79 2.56 2.85 C4r4s0e60 (ours) 8 0.85 3.40 2.60

slide-34
SLIDE 34
  • 2. Reliable proxy settings can be adopted in other NAS methods.

EcoNAS ablation study

Method Setup Cost (GPU days) Params. (M) Error Rate (%) DARTS (on CIFAR-10) c2r0s0 1.5 3.2 3.0 c4r2s0 (ours) 0.3 4.5 2.8 ProxylessNAS (on ImageNet) c0r0s0-S 8 4.1 25.4 c0r0s0-L 8 6.9 23.3 c2r2s0 (ours) 4 5.3 23.2

Zhou, Dongzhan, Xinchi Zhou, Wenwei Zhang, Chen Change Loy, Shuai Yi, Xuesen Zhang, and Wanli Ouyang. "EcoNAS: Finding Proxies for Economical Neural Architecture Search." CVPR 2020.

slide-35
SLIDE 35

EcoNAS results on CIFAR-10

slide-36
SLIDE 36

EcoNAS results on ImageNet

slide-37
SLIDE 37

提纲

  • 简介
  • 快速有效的NAS (高效率搜索)
  • 基于NAS启发的模型压缩
slide-38
SLIDE 38

提纲

  • 简介
  • 快速有效的NAS (高效率搜索)
  • 基于NAS启发的模型压缩(高效率部署)
slide-39
SLIDE 39

Jinyang Guo, Wanli Ouyang, Dong Xu

CVPR2020 Oral

Model Compression Multi-Dimensi sional Pruning: A Unified Framework for

slide-40
SLIDE 40

Mod

  • del Com
  • mpression

Memory: 16GB / 32GB Computation: TFLOPs/s Memory: 8GB Computation: GFLOPs/s Memory: 100KB – 1MB Computation: MFLOPs/s

slide-41
SLIDE 41

Mod

  • del Com
  • mpression

Channel Pruning Quantization Tensor Factorization Compact Network Design XNOR-Net, Rastegari et al., ECCV’16 HAQ, Want et al., CVPR’19 MobileNet, Howard et al., arXiv ShuffleNet, Zhang et al., CVPR’18 Learning…, Liu et al., CVPR’17 Channel Pruning…, He et al., ICCV’17

Model Compression

Accelerating…, Zhang et al., T-PAMI

slide-42
SLIDE 42

➢ Two redundancies are not explored:

  • Temporal-wise redundancy
  • Spatial-wise redundancy

Mot

  • tivation
slide-43
SLIDE 43

➢ Two redundancies are not explored:

  • Temporal-wise redundancy
  • Spatial-wise redundancy

Mot

  • tivation
slide-44
SLIDE 44

➢ Two redundancies are not explored:

  • Temporal-wise redundancy
  • Spatial-wise redundancy

Mot

  • tivation

Downsampling

slide-45
SLIDE 45

➢ We proposed Multi-Dimensional Pruning (MDP):

  • Simultaneously reduce spatial/spatial-temporal and channel

redundancies

  • A unified framework that can prune both 2D CNNs and 3D CNNs

Con

  • ntributions
slide-46
SLIDE 46

Mul ulti-Dimensional Pruning (MDP)

The Searching Stage The Pruning Stage The Fine-tuning Stage

Guo, J., Ouyang, W. and Xu, D., Multi-Dimensional Pruning: A Unified Framework for Model Compression. CVPR 2020.

slide-47
SLIDE 47

➢ The Searching Stage

  • Construct an over-parameterized network

Mul ulti-Dimensional Pruning (MDP)

… …

The corresponding layer in the over-parameterized network

𝑐𝑠𝑏𝑜𝑑ℎ1 𝑐𝑠𝑏𝑜𝑑ℎ2 𝑐𝑠𝑏𝑜𝑑ℎ𝑗 Input tensor (three channels) Gates 𝑕1,1 𝑕1,2 𝑕2,2 𝑕2,2 𝑕𝑗,1 𝑕𝑗,2 Average pooling Average pooling Conv Conv Conv Upsampling Upsampling … Output tensor (two channels) 𝑇(𝜇1) … 𝑇(𝜇2) 𝑇(𝜇𝑗) The original layer Input tensor (three channels) Output tensor (two channels) width height temporal Conv Spatial downsampling ratio = 2 Temporal downsampling ratio = 1 Spatial downsampling ratio = 4 Temporal downsampling ratio = 2 Spatial downsampling ratio = 1 Temporal downsampling ratio = 1

slide-48
SLIDE 48

… … 𝑐𝑠𝑏𝑜𝑑ℎ1 𝑐𝑠𝑏𝑜𝑑ℎ2 𝑐𝑠𝑏𝑜𝑑ℎ𝑗 Input tensor (three channels) Gates 𝑕1,1 𝑕1,2 𝑕2,2 𝑕2,2 𝑕𝑗,1 𝑕𝑗,2 Average pooling Average pooling Conv Conv Conv Upsampling Upsampling … Output tensor (two channels) 𝑇(𝜇1) … 𝑇(𝜇2) 𝑇(𝜇𝑗)

➢ The Searching Stage

  • How to find the best setting of spatial-temporal resolution?
  • Mul

ulti-Dimensional Pruning (MDP)

Inspired by DARTS

Learn differentialble choice 𝑇(𝜇𝑗)

slide-49
SLIDE 49

➢ The Searching Stage

  • Training objective function

Mul ulti-Dimensional Pruning (MDP)

arg min

𝛊,𝛍,𝐇

𝑀 = 𝑀𝑑 + 𝛽𝑀𝑡𝑢 + 𝜃𝑀𝑕𝑏𝑢𝑓 𝑀𝑑: Cross-entropy loss for classification task 𝑀𝑡𝑢: Penalty to introduce sparsity on branch importance for resolution selection 𝑀𝑕𝑏𝑢𝑓: Penalty to introduce sparsity on gates for channel pruning

Guo, J., Ouyang, W. and Xu, D., Multi-Dimensional Pruning: A Unified Framework for Model Compression. CVPR 2020.

slide-50
SLIDE 50

➢ The Pruning Stage

  • Select branch with the largest branch importance score
  • Prune channels with small gate values

Mul ulti-Dimensional Pruning (MDP)

… … 𝑐𝑠𝑏𝑜𝑑ℎ1 𝑐𝑠𝑏𝑜𝑑ℎ2 𝑐𝑠𝑏𝑜𝑑ℎ𝑗 Input tensor (three channels) Gates 𝑕1,1 𝑕1,2 𝑕2,2 𝑕2,2 𝑕𝑗,1 𝑕𝑗,2 Average pooling Average pooling Conv Conv Conv Upsampling Upsampling … Output tensor (two channels) 𝑇(𝜇1) … 𝑇(𝜇2) 𝑇(𝜇𝑗)

slide-51
SLIDE 51

➢ Image Classification (2D CNNs)

Expe xperiments

ThiNet CP Slimming WM DCP Ours 93.4 93.6 93.8 94 94.2 94.4 94.6 94.8 10 20 30 40 50 60 Top-1 Accuracies (%) FLOPs (%)

Top-1 Accuracies (%) on CIFAR-10

VGGNet Guo, J., Ouyang, W. and Xu, D., Multi-Dimensional Pruning: A Unified Framework for Model Compression. CVPR 2020.

slide-52
SLIDE 52

➢ Image Classification (2D CNNs)

Expe xperiments

ThiNet CP WM DCP GAL Ours 89.5 90 90.5 91 91.5 92 92.5 93 30 35 40 45 50 55 Top-5 Accuracies (%) FLOPs (%)

Top-5 Accuracies (%) on ImageNet

ResNet-50 ThiNet CP WM DCP Ours 91.5 92 92.5 93 93.5 94 94.5 44 45 46 47 48 49 50 51 Top-1 Accuracies (%) FLOPs (%)

Top-1 Accuracies (%) on CIFAR-10

ResNet-56 Guo, J., Ouyang, W. and Xu, D., Multi-Dimensional Pruning: A Unified Framework for Model Compression. CVPR 2020.

slide-53
SLIDE 53

➢ Image Classification (2D CNNs)

Expe xperiments

WM DCP Ours 93.8 94 94.2 94.4 94.6 94.8 95 95.2 65 67 69 71 73 75 Top-1 Accuracies (%) FLOPs (%)

Top-1 Accuracies (%) on CIFAR-10

MobileNet-V2 ThiNet WM DCP Ours 85 85.5 86 86.5 87 87.5 88 88.5 89 89.5 30 35 40 45 50 55 60 Top-5 Accuracies (%) FLOPs (%)

Top-5 Accuracies (%) on ImageNet

MobileNet-V2 Guo, J., Ouyang, W. and Xu, D., Multi-Dimensional Pruning: A Unified Framework for Model Compression. CVPR 2020.

slide-54
SLIDE 54

➢ Video Classification (3D CNNs) ➢ C3D

Expe xperiments

TP FP DCP Ours 70 72 74 76 78 80 82 40 42 44 46 48 50 52 Video accuracies (%) FLOPs (%)

Video accuracies (%) on UCF-101

DCP TP FP Ours 40 41 42 43 44 45 46 40 42 44 46 48 50 Video accuracies (%) FLOPs (%)

Video accuracies (%) on HMDB-51

Guo, J., Ouyang, W. and Xu, D., Multi-Dimensional Pruning: A Unified Framework for Model Compression. CVPR 2020.

slide-55
SLIDE 55

➢ Video Classification (3D CNNs) ➢ I3D

Expe xperiments

TP FP DCP Ours 83 84 85 86 87 88 89 46 47 48 49 50 51 Video accuracies (%) FLOPs (%)

Video accuracies (%) on UCF-101

TP FP DCP Ours 56 57 58 59 60 61 62 63 64 46 47 48 49 50 51 Video accuracies (%) FLOPs (%)

Video accuracies (%) on HMDB-51

Guo, J., Ouyang, W. and Xu, D., Multi-Dimensional Pruning: A Unified Framework for Model Compression. CVPR 2020.

slide-56
SLIDE 56

总结

  • ECO-NAS
  • 可靠的搜索环境(Proxy)至关重要
  • 存在快且可靠的搜索环境(Proxy)
  • 利用环境的切换可以设计新的网络结构搜索算法
slide-57
SLIDE 57

总结

  • ECO-NAS
  • 可靠的搜索环境(Proxy)至关重要
  • 存在快且可靠的搜索环境(Proxy)
  • 利用环境的切换可以设计新的网络结构搜索算法
  • Multi-dimensional pruning
  • 网络模型存在空间和时间的冗余
  • MDP:自动学习去掉这些冗余
  • 模型压缩可以从网络结构搜索方法得到启发
slide-58
SLIDE 58