for Data-Efficient GAN Training Shengyu Zhao 1,2 Ji Lin 1 Jun-Yan Zhu - - PowerPoint PPT Presentation

for data efficient gan training
SMART_READER_LITE
LIVE PREVIEW

for Data-Efficient GAN Training Shengyu Zhao 1,2 Ji Lin 1 Jun-Yan Zhu - - PowerPoint PPT Presentation

Differentiable Augmentation for Data-Efficient GAN Training Shengyu Zhao 1,2 Ji Lin 1 Jun-Yan Zhu 3,4 Song Han 1 Zhijian Liu 1 1 MIT 2 IIIS, Tsinghua University 3 Adobe Research 4 CMU NeurIPS 2020 Data Is Expensive Computation Algorithm


slide-1
SLIDE 1

1MIT 2IIIS, Tsinghua University 3Adobe Research 4CMU

Differentiable Augmentation for Data-Efficient GAN Training

NeurIPS 2020

Shengyu Zhao1,2 Zhijian Liu1 Ji Lin1 Song Han1 Jun-Yan Zhu3,4

slide-2
SLIDE 2

Computation Algorithm Computation Algorithm

Data Is Expensive

FFHQ dataset: 70,000 selective post-processed human faces

Months or even years to collect the data, along with prohibitive annotation costs.

ImageNet dataset: millions of images from diverse categories

Big Data

slide-3
SLIDE 3

Obama 100 images

Generated samples of StyleGAN2 (Karras et al.) using only hundreds of images

Cat (Simard et al.) 160 images Dog (Simard et al.) 389 images

GANs Heavily Deteriorate Given Limited Data

slide-4
SLIDE 4

GANs Heavily Deteriorate Given Limited Data

11.1 23.1 36.0 5 10 15 20 25 30 35 40

100% training data 20% training data 10% training data

FID ↓ StyleGAN2 (baseline) + DiffAugment (ours)

CIFAR-10

slide-5
SLIDE 5

Discriminator Overfitting

slide-6
SLIDE 6

#1 Approach: Augment reals only

Augment reals only: the same artifacts appear on the generated images.

Artifacts from Color jittering Artifacts from Translation Artifacts from Cutout (DeVries et al.)

Generated images

slide-7
SLIDE 7

Augment 𝑬 only: the unbalanced optimization cripples training.

#2 Approach: Augment reals & fakes for 𝑬 only

slide-8
SLIDE 8

Our approach (DiffAugment): Augment reals + fakes for both 𝐸 and 𝐻

#3 Approach: Differentiable Augmentation (Ours)

Color Translation Cutout Color Translation Cutout

fakes reals

slide-9
SLIDE 9

11.1 23.1 36.0 9.9 12.2 14.5 5 10 15 20 25 30 35 40

100% training data 20% training data 10% training data

FID ↓ StyleGAN2 (baseline) + DiffAugment (ours)

11.1 23.1 36.0 9.9 12.2 14.5 5 10 15 20 25 30 35 40

100% training data 20% training data 10% training data

FID ↓ StyleGAN2 (baseline) + DiffAugment (ours)

Our Results

CIFAR-10

slide-10
SLIDE 10

Low-Shot Generation

Obama 100 images Cat (Simard et al.) 160 images Dog (Simard et al.) 389 images

slide-11
SLIDE 11

10 20 30 40 50 60 Performance FID↓

Scale/Shift (Noguchi et al.) MineGAN (Wang et al.) TransferGAN (Wang et al.) FreezeD (Mo et al.) Ours

1 10 100 1000 10000 100000 Data # Training Images

Fine-Tuning vs. Ours

No pre-training

100-shot Obama

slide-12
SLIDE 12

100-Shot Interpolation

Our code, datasets, and models are publicly available at https://github.com/mit-han-lab/data-efficient-gans.