! LIT: !
Learned Intermediate representation Training for Model Compression
Animesh Koratana*, Daniel Kang*, Peter Bailis, Matei Zaharia DAWN Project, Stanford InfoLab
http://dawn.cs.stanford.edu/
! LIT: ! L earned I ntermediate representation T raining for Model - - PowerPoint PPT Presentation
! LIT: ! L earned I ntermediate representation T raining for Model Compression Animesh Koratana*, Daniel Kang* , Peter Bailis, Matei Zaharia DAWN Project, Stanford InfoLab http://dawn.cs.stanford.edu/ LIT can compress models up to 4x on CIFAR10:
Animesh Koratana*, Daniel Kang*, Peter Bailis, Matei Zaharia DAWN Project, Stanford InfoLab
http://dawn.cs.stanford.edu/
This s talk: achieving higher compression
networks
Deep compression Knowledge distillation
18 residual blocks 9 residual blocks 18 residual blocks 18 residual blocks 9 residual blocks 9 residual blocks
IR comparison KD comparison
Teacher model: ResNet-110 Student model: ResNet-56
FC layer FC layer
KD loss
Losses
18 residual blocks 9 residual blocks 18 residual blocks 18 residual blocks 9 residual blocks 9 residual blocks
IR comparison KD comparison
Teacher model: ResNet-110 Student model: ResNet-56
FC layer FC layer
IR loss IR loss IR loss KD loss
Losses
18 residual blocks 9 residual blocks 18 residual blocks 18 residual blocks 9 residual blocks 9 residual blocks
IR comparison KD comparison
Teacher model: ResNet-110 Student model: ResNet-56
FC layer FC layer
Training only
IR loss IR loss IR loss KD loss
Losses
Teacher (18) Student (10) Scratch (10)
Original Black hair Blond hair Brown hair Gender Age