In Deep Learning Anima Anandkumar & Zachary Lipton DATA - - PowerPoint PPT Presentation
In Deep Learning Anima Anandkumar & Zachary Lipton DATA - - PowerPoint PPT Presentation
Addressing Data Scarcity In Deep Learning Anima Anandkumar & Zachary Lipton DATA AUGMENTATION To improve generalization, augment training data. In computer vision. Simple techniques: rotation, cropping, noise In speech
DATA AUGMENTATION
- To improve generalization, augment training data.
- In computer vision. Simple techniques: rotation, cropping, noise
- In speech recognition: Additive background noise and spectral
transform
- More sophisticated approaches ?
PREDICTIVE VS GENERATIVE MODELS
y x y x
P(y | x) P(x | y)
- Impressive gains with deep learning
- Information loss (domain of y << x)
- Far more challenging
- Need to model latent variations
DATA AUGMENTATION 1: MIXED REALITY GAN
Merits
- Captures statistics of
natural images
- Learnable
Peril
- Quality of generated
images not high
- Introduces artifacts
Our GAN-based framework – Mr.GANs – narrows gap between synthetic and real data
GAN Merits
- High-quality rendering
- Full annotation for free
- Generate infinite data
Peril Synthetic Data
- Domain mis-match
- Rendering for visual
appeal and not classification
MIXED-REALITY GENERATIVE ADVERSARIAL NETWORKS (MR-GAN)
Tan Nguyen, Hao Chen, Zachary Lipton, Leo Dirac, Stefano Soatto, A.
MIXED-REALITY GENERATIVE ADVERSARIAL NETWORKS (MR-GAN)
- Two domains X and Y.
- CycleGAN: Transforms from
domain X to Y and viceversa.
- Enforcing Cycle consistency:
F(G(X)) ~ X.
- MR-GAN: Progressive
CycleGAN
MIXED-REALITY GENERATIVE ADVERSARIAL NETWORKS (MR-GAN)
CLASSIFICATION RESULTS ON CIFAR-DA4
1million synthetic from 3D
- models. 0.25% real from
- CIFAR100. 4 classes.
Improvement over training
- n real data:
- Real + Synthetic:
5.43% (Stage 0)
- Real + CycleGAN:
5.09% (Stage 1)
- Real + Mr.GAN:
- 8.85% (Stage 2)
THE REAL, THE SYNTHETIC AND THE REFINED
Real Synthetic Refined Real Refined Synthetic
- Mr. GAN pushes both real and synthetic images
closer to one another
(R-S): SYNTHETIC -> REFINED SYNTHETIC
Refined Synthetic Synthetic
(R-S): REAL -> REFINED REAL
Refined Real Real
PREDICTIVE VS GENERATIVE MODELS
y x y x
P(y | x) P(x | y) One model to do both?
- SOTA prediction from CNN models.
- What class of p(x|y) yield CNN models for p(y|x)?
LATENT-DEPENDENT DEEP RENDERING MODEL (LD-DRM)
Nhat Ho, Tan Nguyen, Ankit Patel, A. , Michael Jordan, Richard Baraniuk
LATENT-DEPENDENT DEEP RENDERING MODEL (LD-DRM)
- bject
category intermediate rendering image latent variables
Design joint priors for latent variables based on reverse-engineering CNN predictive architectures
LATENT-DEPENDENT DEEP RENDERING MODEL (LD-DRM)
LATENT-DEPENDENT DEEP RENDERING MODEL (LD-DRM)
LATENT-DEPENDENT DEEP RENDERING MODEL (LD-DRM)
STATISTICAL GUARANTEES FOR THE LD-DRM
- Generalization for prediction depends on the generative model
- Better generalization when no. of active rendering paths minimized
- Rendering path normalization: new form of regularization
- Improves performance significantly.
Training loss in the CNNs equivalent to likelihood in LD-DRM
SEMI-SUPERVISED LEARNING RESULTS
Error rate percentage on CIFAR-10 Error rate percentage on CIFAR-100
LD-DRM achieves comparable results to state-of-the-art SSL methods
DATA AUGMENTATION 3: SYMBOLIC EXPRESSIONS
Goal: Learn a domain of functions (sin, cos, log, add…)
- Training on numerical input-output does not generalize.
Data Augmentation with Symbolic Expressions
- Efficiently encode relationships between functions.
Solution:
- Design networks to use both: symbolic + numerical
COMMON STRUCTURE: TREES
- Symbolic expression trees. Function evaluation tree.
- Decimal trees: encode numbers with decimal representation (numerical).
- Can encode any expression, function evaluation and number.
STRUCTURE : TREE LSTM
RESULTS: EQUATION COMPLETION & FUNCTION EVAL
RESULTS: EQUATION VERIFICATION
Generalization to unseen depth
RESULTS SUMMARIZED
- Vastly Improved numerical evaluation: 90% over function-fitting baseline.
- Generalization to verifying symbolic equations of higher depth
- Combining symbolic + numerical data helps in better generalization
for both tasks: symbolic and numerical evaluation.
LSTM: Symbolic TreeLSTM: Symbolic TreeLSTM: symbolic + numeric 76.40 % 93.27 % 96.17 %
- Collection: Active learning and partial feedback
- Aggregation: Crowdsourcing models
- Augmentation:
- Graphics rendering + GANs
- Semi-supervised learning
- Symbolic expressions