The language of visual attributes Kristen Grauman Facebook AI - - PowerPoint PPT Presentation

the language of visual attributes
SMART_READER_LITE
LIVE PREVIEW

The language of visual attributes Kristen Grauman Facebook AI - - PowerPoint PPT Presentation

The language of visual attributes Kristen Grauman Facebook AI Research University of Texas at Austin Attributes vs. objects Red Round Visual Physical properties entity Ripe Fresh Value of attributes Find a more Zebras have stripes


slide-1
SLIDE 1

Kristen Grauman Facebook AI Research University of Texas at Austin

The language of visual attributes

slide-2
SLIDE 2

Attributes vs. objects

Red Round Ripe Fresh Physical entity Visual properties

slide-3
SLIDE 3

Value of attributes

A lone cow grazes in a green pasture.

Image/video description Zero-shot learning

Zebras have stripes and four legs…

Visual search

“Find a more formal shoe”

Interactive recognition

What color is the beak?

[Ferrari & Zisserman 2007, Kumar et al. 2008, Farhadi et al. 2009, Lampert et al. 2009, Wang & Mori 2010, Berg et al. 2010, Parikh & Grauman 2011, Branson et al. 2010, Kovashka et al. 2012, Kulkarni et al. 2011, Wang et al. 2016, Liu et al. 2015, Singh et al. 2016, …]

slide-4
SLIDE 4

The language of visual attributes

  • Attributes as operators

Attributes:adjectives that modify objects:nouns

  • Attributes for comparisons

Relative differences that people first describe

  • Attributes for visual styles

Semantic topic models for data-driven styles

slide-5
SLIDE 5

Attributes and objects

Red Round Ripe Fresh Physical entity Visual properties Attributes and objects are fundamentally different

slide-6
SLIDE 6

Attribute and Object Representations

sliced apple Yet status quo treats attributes and

  • bjects the same...

As latent vector encodings

e.g., Wang CVPR16, Liu CVPR15, Singh ECCV16, Lu CVPR17, Su ECCV16,…

slide-7
SLIDE 7

Attribute vs. Object Representations

Prototypical “car” instance

  • bject

Prototypical “sliced” instance

?

attribute

slide-8
SLIDE 8

Has to capture interactions with every object

Challenges for the status quo approach

Object-agnostic attribute representation

...

slide-9
SLIDE 9

Challenges for the status quo approach

Has to capture attributes’ distinct manifestations Object-agnostic attribute representation vs. Old car Old man

slide-10
SLIDE 10

= =

Attributes are operators that transform object encodings

Our idea – Attributes as operators

[Nagarajan & Grauman, ECCV 2018]

slide-11
SLIDE 11

Our idea – Attributes as operators

Objects are vectors

=

Attributes are operators Composition is: an attribute operator transforming an object vector

=

T

[Nagarajan & Grauman, ECCV 2018]

slide-12
SLIDE 12

Linguistically inspired regularizers

Antonym-consistency: “Unripe should undo the effect of ripe”

[Nagarajan & Grauman, ECCV 2018]

slide-13
SLIDE 13

Linguistically inspired regularizers

Attribute commutation: Attribute effects should stack.

[Nagarajan & Grauman, ECCV 2018]

slide-14
SLIDE 14

Learning attribute operators

[Nagarajan & Grauman, ECCV 2018]

slide-15
SLIDE 15

Learning attribute operators

Triplet loss to learn embedding space

[Nagarajan & Grauman, ECCV 2018]

slide-16
SLIDE 16

Learning attribute operators

Triplet loss [plus linguistic regularizers] to learn embedding space

Initialize with GloVe word embeddings [Pennington et al. EMNLP 2014]

slide-17
SLIDE 17

Learning attribute operators

Allows unseen compositions

[Nagarajan & Grauman, ECCV 2018]

slide-18
SLIDE 18

Evaluation

UT-Zappos 50k (Yu & Grauman, CVPR 14) MIT States (Isola et al., CVPR 15) 16 attributes x 12 objects 115 attributes x 245 objects

slide-19
SLIDE 19

Evaluating our composition model

Train time

Sliced carrot Unripe orange Diced onion Sliced apple

Diced carrot Sliced

  • range

Test time

slide-20
SLIDE 20

Train time

Diced onion

Diced carrot Test time

Evaluating our composition model

Combination never seen during training Sliced

  • range

Sliced carrot Unripe orange Sliced apple

slide-21
SLIDE 21

Results – Attribute+object composition recognition

MIT States: 6% increase in

  • pen world (3% h-mean)

UT-Zap: 14% increase in

  • pen world (12% h-mean)

*Misra et al. CVPR 2017 #Chen & Grauman CVPR 2014 * # [Nagarajan & Grauman, ECCV 2018]

slide-22
SLIDE 22

Results - Retrieving unseen (unseen) compositions

Rusty Lock query Nearest Images in ImageNet

slide-23
SLIDE 23

The language of visual attributes

  • Attributes as operators

Attributes:adjectives that modify objects:nouns

  • Attributes for comparisons

Relative differences that people first describe

  • Attributes for visual styles

Semantic topic models for data-driven styles

slide-24
SLIDE 24

Smiling Not Smiling

???

Parikh & Grauman, ICCV 2011 Singh & Lee, ECCV 2016

>?

Relative attributes

slide-25
SLIDE 25

Not Smiling

Parikh & Grauman, ICCV 2011 Singh & Lee, ECCV 2016

>?

  • Learn a ranking function per attribute

<

Relative attributes

slide-26
SLIDE 26

Relative attributes

Compare images by an attribute’s “strength”

bright smiling natural

[Parikh & Grauman, ICCV 2011]

slide-27
SLIDE 27

Coarse

v s.

Fine-Grained

v s.

Challenge #1: fine-grained comparisons

Sparsity of supervision problem:

  • 1. Label availability: lots of possible pairs.
  • 2. Image availability: subtleties hard to curate.

Which is more sporty?

slide-28
SLIDE 28

Overcome sparsity of available fine-grained image pairs with attribute-conditioned image generation

  • sporty
  • pen

comfort

+

Our idea: Semantic jitter

+ +

  • Status quo:

Low-level jitter

vs.

Idea: Semantic jitter

Yu & Grauman, ICCV 2017

slide-29
SLIDE 29

Train rankers with both real and synthetic image pairs, test on real fine-grained pairs.

Novel Pair

Real Pairs Synthetic Pairs

vs.

80 90 100

Attribute accuracy

Semantic jitter for attribute learning

Ranking functions trained with deep spatial transformer ranking networks [Singh & Lee 2016] or Local RankSVM [Yu & Grauman 2014]

Faces, Shoes

Yu & Grauman, ICCV 2017

slide-30
SLIDE 30

Challenge #2: Which attributes matter?

slide-31
SLIDE 31

Idea: Prominent relative attributes

Infer which comparisons are perceptually salient

Chen & Grauman, CVPR 2018

slide-32
SLIDE 32

Approach: What causes prominence?

  • Large difference in

attribute strength:

  • Unusual and uncommon

attribute occurrences:

  • Absence of other

noticeable differences:

Visible Forehead Colorful Dark Hair

In general: Interactions between all the relative attributes in an image pair cause prominent differences.

Prominent Difference:

Chen & Grauman, CVPR 2018

slide-33
SLIDE 33

Approach: Predicting prominent differences

  • Relative

Attribute Rankers Relative Attribute Rankers

  • Prominence

Multiclass Classifier

  • input:
  • Prominent

Difference: Visible Teeth

  • Symmetric

encoding

Chen & Grauman, CVPR 2018

slide-34
SLIDE 34

(Top 3 prominent differences for each pair)

Results: Prominent differences

slide-35
SLIDE 35

Accuracy Accuracy

# Top prominent as ground truth # Top prominent as ground truth

Rank-SVM Rank-CNN

Results: Prominent differences

slide-36
SLIDE 36

Prominent differences: impact on visual search

Feedback: “shinier than these” Feedback: “less formal than these”

Refined top search results Initial top search results

… …

Query: “white high-heeled shoes”

Leverage prominence to better focus search results

Chen & Grauman, CVPR 2018

slide-37
SLIDE 37

Prominent differences: impact on visual search

Leverage prominence to better focus search results

Faster retrieval of user’s target image without using any additional user feedback.

Chen & Grauman, CVPR 2018

slide-38
SLIDE 38

From items to styles

slide-39
SLIDE 39

The language of visual attributes

  • Attributes as operators

Attributes:adjectives that modify objects:nouns

  • Attributes for comparisons

Relative differences that people first describe

  • Attributes for visual styles

Semantic topic models for data-driven styles

slide-40
SLIDE 40

How to represent visual style?

Challenges:

  • Same “look” manifests in different garments
  • Emerges organically and evolves over time
  • Soft boundaries

CNN image similarity manually defined style labels stylistic similarity?

slide-41
SLIDE 41

Idea: Discovering visual styles

Unsupervised learning of a style-coherent embedding with a polylingual topic model

Mimno et al. "Polylingual topic models." EMNLP 2009.

Hsiao & Grauman, ICCV 2017

...

An outfit is a mixture of (latent) styles. A style is a distribution over attributes. An outfit is a mixture of (latent) styles. A style is a distribution over attributes.

slide-42
SLIDE 42

Example discovered styles (dresses)

Styles we automatically discover in the Amazon dataset [McAuley et al. 2015]

slide-43
SLIDE 43

Styles automatically discovered in the HipsterWars dataset [Kiapour et al]

Example discovered styles (full outfit)

slide-44
SLIDE 44

Bohemian Hipster

Our embedding naturally facilitates browsing for mixes of user-selected styles

Mixing styles

Hsiao & Grauman, ICCV 2017

slide-45
SLIDE 45

Capsule pieces

Outfit #1 Outfit #2 Outfit #3 Outfit #4 Outfit #5

Creating a “capsule” wardrobe

Goal: Select minimal set of pieces that mix and match well to create many viable outfits

Hsiao & Grauman, CVPR 2018 Inventory

set of garments = argmax compatibility + versatility

Pose as subset selection problem

slide-46
SLIDE 46

Personalized capsule Discover user’s style preferences from album

Creating a “capsule” wardrobe

Hsiao & Grauman, CVPR 2018

slide-47
SLIDE 47

Visual trend forecasting

We predict the future popularity of each style

Al-Halah et al., ICCV 2017

Amazon dataset [McAuley et al. SIGIR 2015]

slide-48
SLIDE 48

What kind of fabric, texture, color will be popular next year?

Visual trend forecasting

slide-49
SLIDE 49

VizWiz: Answer blind people’s visual questions

Hi there can you please tell me what flavor this is? Is my monitor

  • n?

What type of pills are these? What is this?

  • Goal-oriented visual

questions

  • Conversational

language

  • Assistive technology

[Gurari et al. CVPR 2018] Spotlight/Poster Wednesday

slide-50
SLIDE 50

Summary: the language of visual attributes

New ideas for attributes as operators, comparisons, style basis Applications for visual search and fashion image analysis

Aron Yu Kimberly Hsiao Steven Chen Ziad Al-Halah Tushar Nagarajan

Poster Tuesday Spotlight/Poster Thursday

slide-51
SLIDE 51

Papers/code

  • Attributes as Operators. T. Nagarajan and K. Grauman.

In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, Sept 2018. [pdf] [supp] [code]

  • Semantic Jitter: Dense Supervision for Visual Comparisons via Synthetic Images. A. Yu and
  • K. Grauman. In Proceedings of the International Conference on Computer Vision (ICCV), Venice,

Italy, Oct 2017. [pdf] [supp] [poster]

  • Compare and Contrast: Learning Prominent Visual Differences. S. Chen and K. Grauman. In

Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, June 2018. [pdf] [supp] [project page]

  • Fashion Forward: Forecasting Visual Style in Fashion. Z. Al-Halah, R. Stiefelhagen, and K.
  • Grauman. In Proceedings of the International Conference on Computer Vision (ICCV), Venice,

Italy, Oct 2017. [pdf] [supp] [project page]

  • Learning the Latent "Look": Unsupervised Discovery of a Style-Coherent Embedding from

Fashion Images. W-L. Hsiao and K. Grauman. In Proceedings of the International Conference on Computer Vision (ICCV), Venice, Italy, Oct 2017. [pdf] [supp] [project page/code]

  • Creating Capsule Wardrobes from Fashion Images. W-L. Hsiao and K. Grauman. In

Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, June 2018. (Spotlight) [pdf]

  • VizWiz Grand Challenge: Answering Visual Questions from Blind People. D. Gurari, Q. Li, A.

Stangl, A. Guo, C. Lin, K. Grauman, J. Luo, and J. Bigham. In Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, June 2018. (Spotlight) [pdf]