Long-Tailed Sources & Open Compound Targets Boqing Gong CVPR - - PowerPoint PPT Presentation

long tailed sources open compound targets
SMART_READER_LITE
LIVE PREVIEW

Long-Tailed Sources & Open Compound Targets Boqing Gong CVPR - - PowerPoint PPT Presentation

Towards Visual Recognition in the Wild : Long-Tailed Sources & Open Compound Targets Boqing Gong CVPR 2009 50 classes 85 attributes Kernel Methods for Unsupervised Domain Adaptation 10~100 classes 2011-2015 ILSVRC 2010-2017 ~1000


slide-1
SLIDE 1

Towards Visual Recognition in the Wild: Long-Tailed Sources & Open Compound Targets

Boqing Gong

slide-2
SLIDE 2

CVPR 2009 50 classes 85 attributes

slide-3
SLIDE 3

2011-2015 Kernel Methods for Unsupervised Domain Adaptation 10~100 classes

slide-4
SLIDE 4

ILSVRC 2010-2017 ~1000 classes

Bottom image credit: http://www.thegreenmedium.com/blog/2019/5/24/why-robots-will-help-you-rather-than-t ry-to-take-over-the-world-a-brief-history-of-ai

slide-5
SLIDE 5

ICML 2014 Deep features!

slide-6
SLIDE 6

Object recognition in the wild 5k~8k classes

slide-7
SLIDE 7

in the wild

slide-8
SLIDE 8

Right image credit: https://natureneedsmore.org/the-elephant-in-the-room/

in the wild

slide-9
SLIDE 9

CVPR 2019 (oral), improving neural architectures

slide-10
SLIDE 10

Long-tailed ImageNet (1000 classes) Long-tailed Places-365 Long-tailed MS1M ArcFace (74.5k ids) A memory bank to enhance tail classes

CVPR 2019 (oral), improving neural architectures

slide-11
SLIDE 11

An old AI problem A new AI problem (meta-learning, transfer learning, zero-shot learning)

Acknowledgement: Matthew Brown @Google

slide-12
SLIDE 12

Existing work

Class-wise weighting,

  • ver/under-sampling, etc.

[CVPR’18] Large Scale Fine-Grained Categorization and Domain-Specifjc Transfer Learning [CVPR’19] Class-Balanced Loss Based on Efgective Number of Samples [NeurIPS’19] Learning Imbalanced Datasets with Label-Distribution-Aware Margin Loss [ICLR’20] Decoupling Representation and Classifjer for Long-Tailed Recognition

slide-13
SLIDE 13

Classes Frequency

slide-14
SLIDE 14

Existing work

Class-wise weighting,

  • ver/under-sampling, etc.

[CVPR’18] Large Scale Fine-Grained Categorization and Domain-Specifjc Transfer Learning [CVPR’19] Class-Balanced Loss Based on Efgective Number of Samples [NeurIPS’19] Learning Imbalanced Datasets with Label-Distribution-Aware Margin Loss [ICLR’20] Decoupling Representation and Classifjer for Long-Tailed Recognition

Existing work assumes 𝝑=0

… as domain adaptation

Target Source

slide-15
SLIDE 15

Existing work assumes 𝝑=0

… as domain adaptation

Many training images in a head class: 𝝑=0 Few-shot training images in a tail class: 𝝑≠0

Head vs. tail

slide-16
SLIDE 16

CVPR 2020 (oral), long-tailed recognition ⩰ domain adaptation

slide-17
SLIDE 17

Our approach

Estimating both & by unifying [CVPR’19] & an improved meta-learning method

SOTA on six datasets

○ CIFAR-LT-10 ○ CIFAR-LT-100 ○ ImageNet-LT ○ Places-LT ○ iNaturalist 2017 ○ iNaturalist 2018

slide-18
SLIDE 18

Long-tailed visual recognition (LTVR)

Emerging challenge as the datasets grow in scale Timely topic Datasets: iNaturalist, LVIS, ImageNet, COCO, etc. Tasks: almost all

… as domain adaptation

New perspective to LTVR New powerhouse of methods

Domain-invariant representation learning Curriculum domain adaptation Adversarial learning Classifjer discrepancy Data augmentation & synthesis, etc.

Difg: no access to target data

slide-19
SLIDE 19

in the wild

slide-20
SLIDE 20

Open compound test cases (target)

slide-21
SLIDE 21

Open compound test cases (target)

slide-22
SLIDE 22

Open compound domain adaptation

Training:

Labeled source domain data Unlabeled data of the compound target

Testing:

in the compound target domain and in previously unseen domains

Liu, Ziwei, Zhongqi Miao, Xingang Pan, Xiaohang Zhan, Stella X. Yu, Dahua Lin, and Boqing Gong. "Compound domain adaptation in an open world." CVPR 2020. (oral)

slide-23
SLIDE 23

Experiments

slide-24
SLIDE 24

Our approach to break the compound target domain

into a series of bi-domain adaptation problems by “domain distances” between the source and latent domains in the target (curriculum training)

Source Latent domain 1 Latent domain 2 Latent domain 3

slide-25
SLIDE 25

into a series of bi-domain adaptation problems by “domain distances” between the source and latent domains in the target (curriculum training)

Source Latent domain 1 Latent domain 2 Latent domain 3

Our approach to break the compound target domain

slide-26
SLIDE 26

into a series of bi-domain adaptation problems by “domain distances” between the source and latent domains in the target (curriculum training)

Source Latent domain 1 Latent domain 2 Latent domain 3

Our approach to break the compound target domain

slide-27
SLIDE 27

into a series of bi-domain adaptation problems by “domain distances” between the source and latent domains in the target (curriculum training)

Source Latent domain 1 Latent domain 2 Latent domain 3

Our approach to break the compound target domain

slide-28
SLIDE 28

Pushing the boundary of visual recognition

Long-tailed source domains

The elephant in the room as we scale up classes / study the wild data Memory bank to enhance tail classes (CVPR’19, oral) Domain adaptation: a new powerhouse of techniques (CVPR’20, oral) Improved meta-learning for long-tailed recognition (undergoing)

Open compound target domains (CVPR’20, oral)

Learning from unlabeled, noisy data in the wild (undergoing)