Concrete Autoencoders Abubakar Abid Muhammed Fatih Balin James Zou - - PowerPoint PPT Presentation

concrete autoencoders
SMART_READER_LITE
LIVE PREVIEW

Concrete Autoencoders Abubakar Abid Muhammed Fatih Balin James Zou - - PowerPoint PPT Presentation

Differentiable Feature Selection with Concrete Autoencoders Abubakar Abid Muhammed Fatih Balin James Zou Poster: Thu Jun 13th 06:30 - 09:00 PM @ Pacific Ballroom #188 Unsupervised Feature Selection (UFS) is Widely Used in Machine


slide-1
SLIDE 1

Differentiable Feature Selection with

Concrete Autoencoders

Abubakar Abid★ Muhammed Fatih Balin★ James Zou

Poster: Thu Jun 13th 06:30 - 09:00 PM @ Pacific Ballroom #188

slide-2
SLIDE 2

Differentiable Feature Selection and Reconstruction with Concrete Autoencoders

Unsupervised Feature Selection (UFS) is Widely Used in Machine Learning

  • Identify the subset of most informative features in dataset
  • Simplifies the process of training models
  • Especially useful if the data is difficult or expensive to collect

2

slide-3
SLIDE 3

Differentiable Feature Selection and Reconstruction with Concrete Autoencoders

Unsupervised Feature Selection (UFS) is Used Widely in Applied ML

  • Example: the L1000 Landmark Genes [Lamb et al., 2006]

3

All Genes Samples Samples Samples L1000 Genes All Genes

Recons truction Feature Selection

slide-4
SLIDE 4

Differentiable Feature Selection and Reconstruction with Concrete Autoencoders

Unsupervised Discriminative Feature Selection (UDFS) [Yang et al., 2011] Multi-Cluster Feature Selection (MCFS) [Cai et al., 2010] Autoencoder Feature Selection (AEFS) [Han et al., 2017]

UFS Methods Typically Rely on Regularization

4

All based on L1 or L21 regularization

slide-5
SLIDE 5

Differentiable Feature Selection and Reconstruction with Concrete Autoencoders

What about directly backpropagating through discrete “feature selection” nodes?

5

slide-6
SLIDE 6

Differentiable Feature Selection and Reconstruction with Concrete Autoencoders

What about directly backpropagating through discrete “feature selection” nodes?

6

Replace the weights

  • f the encoder with

parameters of a Concrete Random Variable (Maddison, 2016)

slide-7
SLIDE 7

Differentiable Feature Selection and Reconstruction with Concrete Autoencoders

Results on the ISOLET dataset (reconstruction error)

7

Reconstruction error (lower is better) Number of features selected

slide-8
SLIDE 8

Differentiable Feature Selection and Reconstruction with Concrete Autoencoders

Results on the ISOLET dataset (classification accuracy)

8

Reconstruction error (lower is better) Classification accuracy (higher is better) Number of features selected Number of features selected

slide-9
SLIDE 9

Differentiable Feature Selection and Reconstruction with Concrete Autoencoders

Concrete Autoencoder (CAE) Genes Outperform the L1000 Landmark Genes!

9

Reconstruction error (lower is better) Number of genes selected by CAE

slide-10
SLIDE 10

Differentiable Feature Selection and Reconstruction with Concrete Autoencoders

Concrete Autoencoder Takeaways

10

  • More effective than other feature selection methods based on

regularization

  • Implementation is just a few lines of code from a standard

autoencoder

  • Training time is similar to standard autoencoder per epoch
  • Can be extended to supervised/semi-supervised settings
slide-11
SLIDE 11

Differentiable Feature Selection and Reconstruction with Concrete Autoencoders

Start using concrete autoencoders today!

Installation: pip install concrete-autoencoder Code: https://github.com/mfbalin/Concrete-Autoencoders For more details and results: Poster: Thu Jun 13th 06:30 - 09:00 PM @ Pacific Ballroom #188 Contact: a12d@stanford.edu, fatih.balin@boun.edu.tr

11