Learn earn CN CNNs Ns fr from om Lar arge ge-scale cale We - - PowerPoint PPT Presentation

learn earn cn cnns ns fr from om lar arge ge scale cale
SMART_READER_LITE
LIVE PREVIEW

Learn earn CN CNNs Ns fr from om Lar arge ge-scale cale We - - PowerPoint PPT Presentation

Learn earn CN CNNs Ns fr from om Lar arge ge-scale cale We Web b Im Images ages wi without hout Hu Human an An Anno notations tations Weilin Huang Malong Technologies Ho How t w to Trai ain a H a High-Perf erfor orma


slide-1
SLIDE 1

Learn earn CN CNNs Ns fr from

  • m Lar

arge ge-scale cale We Web b Im Images ages wi without hout Hu Human an An Anno notations tations

Malong Technologies Weilin Huang

slide-2
SLIDE 2

June 26, 2017

Ho How t w to Trai ain a H a High-Perf erfor

  • rma

manc nce e CNN Data

ImageNet, Webvision

Network

AlexNet, GoogleNet, VggNet, ResNet, DenseNet

Loss

Learni ning ng resour urce ce Wh What to do Learni ning ng strategy gy How w to do How w to do Model el Capabi bility ity

slide-3
SLIDE 3

June 26, 2017

Motivations

  • Pe

Perfor

  • rmance

mance on

  • n Im

Imag ageNet eNet ha has s been n saturation turation. .

From

  • m ~30%

0% (2009) 009) --

  • ->~2

~2.2% .2% (2017) 17)

  • Develo

lop p ne new appro roaches aches work

  • rking

ing on

  • n large

ge-scale cale data a in n real-wo world rld scenari narios

  • s
  • Data

ta, mod

  • del

el archi hite tecture cture, los

  • ss,

s, train inin ing g strat ategy egy are e all important portant

  • Train

ain CNNs NNs from

  • m web

eb image ges s are e most

  • st com
  • mmon

mon tasks sks in n ind ndustri ustries es

  • Train

ain CNNs Ns withou thout t hu human man labelli elling ng --

  • ->

> weakly kly-superv upervis ised d learn rning ing

slide-4
SLIDE 4

June 26, 2017

Program Chairs General Chairs

  • J. Berent
  • A. Gupta
  • R. Sukthankar
  • L. Van Gool

Wen Li Limin Wang Wei Li

  • E. Agustsson

In Introdu ducti ction: WebVision Workshop Organizers

slide-5
SLIDE 5

June 26, 2017

WebVi Visi sion

  • n Dataset

et:

  • 2 Sources
  • 1,000 categories
  • 2.4M training images
  • 50K validation Images
  • 50K test Image

ges 1,000 0 semanti ntic c conce ncept pts from m ILSVRC VRC 2012

Wen Li, Limin in Wang ng, Wei Li, Eirikur kur Agus ustss tsson

  • n, Luc Van Gool
  • l, "WebV

bVisi ision

  • n Database

abase: Visual ual Learning rning and d Unde ders rstandin tanding g from

  • m Web Data".ar

arXi Xiv: : 1708.028 8.02862, , 2017. .

In Introdu

  • ductio

ction: n: Database Construction

slide-6
SLIDE 6

June 26, 2017

Mai ain Ch Chal alleng enge: e: Data Imbalance

slide-7
SLIDE 7

Tench Terrapin Caretta

June 26, 2017

Mai ain Challen enge: e: Label Noise

slide-8
SLIDE 8

June 26, 2017

Mai ain Challen enge: e: Label Noise

slide-9
SLIDE 9

June 26, 2017

Related Works

  • Di

Directly ectly le lear arn n from rom no nois isy y la label els s

  • 1. Noise-robu

robust st Algorith thms ms

  • 2. Label-cl

cleansi eansing ng methods hods

  • Se

Semi-Supervise Supervised d methods thods

  • ->N

>Need ed a s small ll set of manual ually ly-la labeled beled

  • Re

Recent ent deep ep le learn arning ing ap approaches proaches developed veloped for r both

  • th groups
  • ups of metho

thods ds

Im Improve e model l cap apab ability ity of s f stan andard ard neural al networks rks by introdu ducin cing g new trai aining ing strateg ategies. ies.

  • -> dif

ifficul ficult t to id identify ify mis isla labe beled led samples les from m hard train inin ing g samples les

slide-10
SLIDE 10

“Humans and animals learn much better when the examples are not randomly presented but organized in a meaningful order which illustrates gradually more concepts, and gradually more complex ones.”

  • Y. Bengio, J. Louradour, R. Collobert, and J. Weston, Curriculum Learning, ICML, 2009.

June 26, 2017

Methodo dolo logy gy: : Curriculum Learning

Curriculum iculum learning rning —Train CNNs on tasks with increasing difficulty —Train CNNs using samples with increasing complexity

slide-11
SLIDE 11

June 26, 2017

Methodo dolo logy gy: : Curriculum Learning Processing

Steps: ps:

  • Split a learning problem into a number of subtasks
  • Order subtasks by difficulty
  • Decide a task-transform threshold
  • Find an optimized path that leads to fast convergence

and better generalization

  • Simple principle: proceed harder tasks once easier ones

are handled

slide-12
SLIDE 12

June 26, 2017

Methodo dolo logy gy: : Idealistic Curriculum Learning Processing

  • T. Matiisen, A. Oliver, T. Cohen, and J. Schulman, Teacher-Student Curriculum Learning,

arXiv:170 :1707.0 7.001 0183, 2017.

slide-13
SLIDE 13

June 26, 2017

Methodo dolo logy gy: : Formulate our problem

slide-14
SLIDE 14

June 26, 2017

Methodo dolo logy gy: : Curriculum Design

  • Split the whole training set into multiple subsets
  • Density-Distance clustering in each category
  • Rank subsets with increasing complexity

Step One: Similarity Matrix: Step Two: Sample Density: Step Three: Sample Distance:

(Rodriguez and Laio, Science, 2014.)

slide-15
SLIDE 15

Methodo dolo logy gy: : Curriculum Design

Subset1 Subset2 Subset3

slide-16
SLIDE 16

Terrapin Tench Tench

Subse bset t 1 Subse bset t N

Terrapin

Methodo dolo logy gy: : Curriculum Design

slide-17
SLIDE 17

June 26, 2017

Subset t 1 Subset t 3 Subset t 2

Task One Task Two Task Three

Methodo dolo logy gy: : Train with Curriculum Learning

𝝐𝒈 𝛜𝒙𝒋𝒌

𝒏 = 𝒖 𝒍=𝟐 𝒐𝒖

𝝐𝒈 𝝐𝑷𝒍

𝒏 · 𝝐𝑷𝒍 𝒏

𝝐𝒙𝒋𝒌

𝒏

· 𝒔𝒖 𝑢 is number of subtasks, t= 3 𝑠𝑢 is sample weight, , 𝑠 = {1, 0.5,0.5} 𝑠

1=1

𝑠2=0.5 𝑠3=0.5

slide-18
SLIDE 18

Task sk Three ee Meta Data Curri rriculu culum m Learn earning ing Curriculum 2 Subsets Model- B Model- C Model- D

Subset 1

Model- A Curriculum 3 Subsets Curricu riculum lum Learn earning ing

June 26, 2017

Methodo dolo logy gy: : Models with Different Training Schemes

Baseli seline ne Model el 1 Task sk One Baseli seline ne Model el 2 Curricu riculum lum Model el 1 Curri rriculu culum m Model el 2

slide-19
SLIDE 19

Mini-batch = 256 Curriculum Design = 3 subsets

  • Classes balance only on Subset_1

—> Randomly select 128 classes —> Each class only has one sample

  • Samples balance among subsets (three subsets applied)

[Subset_1 = 128, Subset_2 = 64, Subset_3 = 64]

June 26, 2017

Methodo dolo logy gy: : Selective Data Balance

slide-20
SLIDE 20

Conv nv.1

9x9 5x5 7x7

Input Data

Feature ture Map-1 Feature ture Map-2 Featur ture e Map-3

Concat ncat

Feature ture Map

Enhan ance ce low-le level vel featu tures res whic ich imp mprov

  • ve

e the performa rmance nce (about ut 0.5% 5%).

June 26, 2017

Methodo

  • dolo

logy gy: : Multi-Scale Convolutional Kernel

slide-21
SLIDE 21
  • Figure1. Testing

ng loss of four ur diff fferent ent mode dels s with Incep eption tion_v2 v2 (also compar paring ing to K-mean an cluste terin ring g in curriculum culum design) gn)

June 26, 2017

Result ult: : Testing Loss

slide-22
SLIDE 22

June 26, 2017

Result ults: : Single Model, 10 Crops

Table le 1. Dif iffer ferent ent models els based d on Incep eption_v2 tion_v2 on validati ation

  • n set.

Table 3. Model el-D D with h various

  • us networks

works Table 2. Model el-D D with h various

  • us amoun

unts ts of hig ighl hly y nois isy y data.

slide-23
SLIDE 23

June 26, 2017

Co Compar ariso isons: ns: Model B & D – Top Positive Categories

Improve 668 categories, reduce 195 categories, and 137 unchanged

slide-24
SLIDE 24

June 26, 2017

Co Compar ariso isons: ns: Model B & D – Top Negative Categories

Improve 668 categories, reduce 195 categories, and 137 unchanged

slide-25
SLIDE 25
slide-26
SLIDE 26

June 26, 2017

Co Conclusi usions ns

—> > Train ain high-performa performance nce CNNs Ns from

  • m large

ge-scale cale web imag ages es —> > Be Bette ter r gene neraliza ralization tion capabil pability ity —> > Improv rove e our products ucts where re real-wo world rld data ta was as claw awed d from

  • m Internet

rnet with th less s human an labelli lling ng or

  • r labels

ls are incon

  • nsis

siste tence nce —> > Will ll develop

  • p semi-sup

supervised ervised and nd wea eakly kly-supervis upervised ed app ppro roache aches

Su Summary mary: :

—> > Hand ndle le label bel inc ncon

  • nsi

siste stence nce and nd data ta un unbalance alance

slide-27
SLIDE 27

Sheng Guo, Weilin Huang, Chenfan Zhuang, Dengke Dong, Haozhi Zhang, Matthew R. Scott, Dinglong Huang Malong Technologies Co., Ltd.

June 26, 2017

Team am Members rs

Our ur Te Team am:

slide-28
SLIDE 28

Team am histor

  • ry

y Team am members rs ac achievem vements ents on

  • n lar

arge-sca scale le chal allen enges: es:

  • ICCV - 15: ILSVRC2015 (ImageNet): scene classification - 2nd
  • CVPR - 15 :Large-scale Scene Understanding Challenge (LSUN): scene classification - 2nd
  • CVPR - 15 : ChaLearn Looking at People Challenge 2015: cultural event recognition - 3rd
  • CVPR - 16 : Large-scale Scene Understanding Challenge (LSUN): scene classification - 1st
  • ECCV - 16: ILSVRC2016 (ImageNet): scene classification - 4th
  • CVPR -17: Webvision Image classification – 1st
slide-29
SLIDE 29

AI for Product Recognition.

About Mal along

slide-30
SLIDE 30

We live in a world of products. In retail, manufacturing, and security scenarios, products need to be routinely recognized at a high-level, microscopic-level, and even the invisible (x-ray) level. If a machine can “see” products as well as people can, higher efficiency can be achieved in retail product checkouts, higher quality in manufacturing product testing, and higher safety via baggage scanning of products – just to name a few. Using breakthrough GPU-powered semi-supervised deep learning algorithms, scientists at Malong invented product recognition technology which operates at human-level performance across the full-stack of visual input levels – the big, the small, and the invisible, to help improve efficiency, quality, and safety, for our world.

FULL-STACK PRODUCT RECOGNITION

slide-31
SLIDE 31

We We ar are hiring ing - Say hello lo at: at: HR HR@MA MALONG LONG.COM .COM

Thank you !