HexaGAN: Generative Adversarial Nets for Real World Classification - - PowerPoint PPT Presentation

β–Ά
hexagan generative adversarial nets for real world
SMART_READER_LITE
LIVE PREVIEW

HexaGAN: Generative Adversarial Nets for Real World Classification - - PowerPoint PPT Presentation

HexaGAN: Generative Adversarial Nets for Real World Classification Uiwon Hwang , Dahuin Jung, and Sungroh Yoon Seoul National University Electrical and Computer Engineering speaker Problem Definition Missing data problem


slide-1
SLIDE 1

Uiwon Hwang , Dahuin Jung, and Sungroh Yoon

HexaGAN: Generative Adversarial Nets for Real World Classification

Seoul National University Electrical and Computer Engineering

βˆ—

speaker

βˆ—

slide-2
SLIDE 2
  • Missing data problem
  • Missing data imputation
  • Imputing missing elements in a data level
  • Class imbalance problem
  • Class-conditional generation
  • Imputing the entire elements of a sample

conditioned on a label

  • Missing label problem
  • Semi-supervised learning
  • Imputing missing class labels using a classifier
  • Keyword: Imputation

Problem Definition

𝑦$ 𝑦% 𝑦& 𝑧 𝑦$ 𝑦% 𝑦& 𝑧 𝑦$ 𝑦% 𝑦& 𝑧 𝑦$ 𝑦% 𝑦& 𝑧 𝑦$ 𝑦% 𝑦& 𝑧 𝑦$ 𝑦% 𝑦& 𝑧

slide-3
SLIDE 3
  • Missing data problem
  • Missing data imputation
  • filling missing elements in a data level
  • Class imbalance problem
  • Class-conditional generation
  • Imputing the entire elements of a sample

conditioned on a label

  • Missing label problem
  • Semi-supervised learning
  • Imputing missing class labels using a classifier
  • Keyword: Imputation

Problem Definition

slide-4
SLIDE 4
  • Missing data problem
  • Missing data imputation
  • filling missing elements in a data level
  • Class imbalance problem
  • Class-conditional generation
  • Imputing the entire elements of a sample

conditioned on a label

  • Missing label problem
  • Semi-supervised learning
  • Imputing missing class labels using a classifier
  • Keyword: Imputation

Problem Definition

𝑦$ 𝑦% 𝑦& 𝑧 𝑦$ 𝑦% 𝑦& 𝑧 𝑦$ 𝑦% 𝑦& 𝑧 𝑦$ 𝑦% 𝑦& 𝑧

slide-5
SLIDE 5
  • Missing data problem
  • Missing data imputation
  • filling missing elements in a data level
  • Class imbalance problem
  • Class-conditional generation
  • Imputing the entire elements of a sample

conditioned on a label

  • Missing label problem
  • Semi-supervised learning
  • Imputing missing class labels using a classifier
  • Keyword: Imputation

Problem Definition

slide-6
SLIDE 6
  • Missing data problem
  • Missing data imputation
  • filling missing elements in a data level
  • Class imbalance problem
  • Class-conditional generation
  • Imputing the entire elements of a sample

conditioned on a label

  • Missing label problem
  • Semi-supervised learning
  • Imputing missing class labels using a classifier

β†’ Keyword: Imputation

Problem Definition

slide-7
SLIDE 7
  • We propose a generative adversarial network to solve the problems in real world classification

simultaneously

Overview of HexaGAN

slide-8
SLIDE 8
  • Missing data (element-wise) imputation (to solve the missing data problem)
  • Components
  • 𝐹: transfers both labeled and unlabeled instances into the hidden space
  • 𝐻*+: imputes missing data
  • 𝐸*+ - $:/: distinguishes b/w missing and non-missing elements

Addressing Three Problems

slide-9
SLIDE 9
  • Missing data (element-wise) imputation (to solve the missing data problem)
  • Components
  • 𝐹: transfers both labeled and unlabeled instances into the hidden space
  • 𝐻*+: imputes missing data
  • 𝐸*+ - $:/: distinguishes b/w missing and non-missing elements

Addressing Three Problems

slide-10
SLIDE 10
  • Missing data (element-wise) imputation (to solve the missing data problem)
  • Components
  • 𝐹: transfers both labeled and unlabeled instances into the hidden space
  • 𝐻*+: imputes missing data
  • 𝐸*+ - $:/: distinguishes b/w missing and non-missing elements

Addressing Three Problems

slide-11
SLIDE 11
  • Missing data (element-wise) imputation (to solve the missing data problem)
  • Components
  • 𝐹: transfers both labeled and unlabeled instances into the hidden space
  • 𝐻*+: imputes missing data
  • 𝐸*+ - $:/: distinguishes b/w missing and non-missing elements
  • A novel element-wise adversarial loss function and gradient penalty
  • max

345

min

845

  • Addressing Three Problems
slide-12
SLIDE 12
  • Class conditional generation (to solve the class imbalance problem)
  • Components
  • 𝐻93: creates conditional hidden vectors 𝐒;
  • 𝐸93: determines whether a hidden vector is from the dataset or has been created by 𝐻93
  • 𝐻*+ generates the entire elements conditioned on the minority class

Addressing Three Problems

slide-13
SLIDE 13
  • Class conditional generation (to solve the class imbalance problem)
  • Components
  • 𝐻93: creates conditional hidden vectors 𝐒;
  • 𝐸93: determines whether a hidden vector is from the dataset or has been created by 𝐻93
  • 𝐻*+ generates the entire elements conditioned on the minority class

Addressing Three Problems

slide-14
SLIDE 14
  • Class conditional generation (to solve the class imbalance problem)
  • Components
  • 𝐻93: creates conditional hidden vectors 𝐒;
  • 𝐸93: determines whether a hidden vector is from the dataset or has been created by 𝐻93
  • 𝐻*+ generates the entire elements conditioned on the minority class
  • Losses
  • WGAN loss + zero-centered gradient penalty
  • Add Loss of 𝐻*+ calculated from 𝐲

=; and the cross entropy of 𝐲 =, 𝐳; to 𝐻93

Addressing Three Problems

slide-15
SLIDE 15
  • Semi-supervised learning (to solve the missing label problem)
  • Components
  • 𝐷: estimates class labels. This also works as the label generator
  • 𝐸*+ - /A$: distinguishes b/w real and pseudo (fake) labels

Addressing Three Problems

slide-16
SLIDE 16
  • Semi-supervised learning (to solve the missing label problem)
  • Components
  • 𝐷: estimates class labels. This also works as the label generator
  • 𝐸*+ - /A$: distinguishes b/w real and pseudo (fake) labels
  • We adopt the pseudo-labeling technique of TripleGAN (Li et al., NIPS 2017)
  • The two components are related adversarially
  • Addressing Three Problems
slide-17
SLIDE 17
  • Semi-supervised learning (to solve the missing label problem)
  • Components
  • 𝐷: estimates class labels. This also works as the label generator
  • 𝐸*+ - /A$: distinguishes b/w real and pseudo (fake) labels
  • We adopt the pseudo-labeling technique of TripleGAN (Li et al., NIPS 2017)
  • The two components are related adversarially
  • Addressing Three Problems
slide-18
SLIDE 18
  • Not three separate models, this is ONEmodel dubbed HexaGAN
  • The six components of HexaGAN interplay to solve the problems effectively

Overview of HexaGAN(revisited)

slide-19
SLIDE 19
  • Theorem 1: Global optimality of π‘ž 𝑦 𝑛D = 1 = π‘ž(𝑦|𝑛D = 0) for HexaGAN
  • A generator distribution π‘ž(𝑦|𝑛D = 0) is a global optimum for the min-max game of 𝑯𝑡𝑱 and 𝑬𝑡𝑱,

if and only if 𝒒 π’š 𝒏𝒋 = 𝟐 = 𝒒(π’š|𝒏𝒋 = 𝟏) for all π’š ∈ ℝ/, except possibly on a set of zero Lebesgue measure.

Theorems

slide-20
SLIDE 20
  • Theorem 1: Global optimality of π‘ž 𝑦 𝑛D = 1 = π‘ž(𝑦|𝑛D = 0) for HexaGAN
  • A generator distribution π‘ž(𝑦|𝑛D = 0) is a global optimum for the min-max game of 𝑯𝑡𝑱 and 𝑬𝑡𝑱,

if and only if 𝒒 π’š 𝒏𝒋 = 𝟐 = 𝒒(π’š|𝒏𝒋 = 𝟏) for all π’š ∈ ℝ/, except possibly on a set of zero Lebesgue measure.

  • Theorem 2: The adversarial loss for semi-supervised learning is the ODM cost
  • Output distribution matching (ODM) cost function (Sutskeveret al., ICLR workshop 2016)
  • the global minimum of the supervised cost function is also a global minimum of the ODM cost function
  • Optimizing the adversarial losses for C and 𝐸*+ - /A$ imposes an unsupervised constraint on C.

Then, the adversarial losses for semi-supervised learning in HexaGAN satisfy the definition of the ODM cost.

Theorems

slide-21
SLIDE 21
  • Missing data imputation
  • HexaGAN shows good performances on various real world datasets
  • Medical, financial, vision, …

Experimental Results

slide-22
SLIDE 22

e define the three problems (missing data, class imbalance, and missing label) in real-world classification from the perspective of imputation Our framework is simple to use and works automatically when the absence of data and label (𝐧and 𝑛X) is indicated. We loss function and gradient penalty for element-wise imputation. Our imputation performance produces stable, state-of-the-art results. Our method significantly outperforms cascading combinations of the existing state-of-the-art methods.

  • For more details, please visit our poster session!
  • June 11th (Today), 6:30 – 9:00 pm, Pacific Ballroom #20

Conclusions

HexaGAN

𝑭 𝑫 𝑯𝑡𝑱 𝑬𝑡𝑱 𝑬𝑫𝑯 𝑯𝑫𝑯 Missing Data

Real World Classification

Missing Label Class Imbalance

slide-23
SLIDE 23

e define the three problems (missing data, class imbalance, and missing label) in real-world classification from the perspective of imputation Our framework is simple to use and works automatically when the absence of data and label (𝐧and 𝑛X) is indicated. We loss function and gradient penalty for element-wise imputation. Our imputation performance produces stable, state-of-the-art results. Our method significantly outperforms cascading combinations of the existing state-of-the-art methods.

  • For more details, please visit our poster session!
  • June 11th (Today), 6:30 – 9:00 pm, Pacific Ballroom #20

Conclusions

HexaGAN

𝑭 𝑫 𝑯𝑡𝑱 𝑬𝑡𝑱 𝑬𝑫𝑯 𝑯𝑫𝑯 Missing Data

Real World Classification

Missing Label Class Imbalance