Empirical Study of the Benefits of Overparameterization in Learning - - PowerPoint PPT Presentation

empirical study of the benefits of overparameterization
SMART_READER_LITE
LIVE PREVIEW

Empirical Study of the Benefits of Overparameterization in Learning - - PowerPoint PPT Presentation

Empirical Study of the Benefits of Overparameterization in Learning Latent Variable Models Rares-Darius Buhai 1 , Yoni Halpern 2 , Yoon Kim 3 , Andrej Risteski 4 , David Sontag 1 1 MIT, 2 Google, 3 Harvard, 4 CMU Overparameterization = training a


slide-1
SLIDE 1

Empirical Study of the Benefits of Overparameterization in Learning Latent Variable Models

Rares-Darius Buhai1, Yoni Halpern2, Yoon Kim3, Andrej Risteski4, David Sontag1

1MIT, 2Google, 3Harvard, 4CMU

slide-2
SLIDE 2

Overparameterization

= training a larger model than necessary

→ practice: [Zhang et al., 2016] commonly used neural networks are so large that they can learn randomized labels. → theory: [Allen-Zhu et al., 2018; Allen-Zhu et al., 2019] overparameterized neural networks provably learn and generalize for certain classes of functions. Supervised learning: easier optimization, often without sacrificing generalization.

slide-3
SLIDE 3

Overparameterization in unsupervised learning

Task: learning latent variable models. Contribution: Empirical study of the benefits of

  • verparameterization in learning latent variable

models.

slide-4
SLIDE 4

Know . Maximum likelihood: . Typically intractable.

Latent variable models

(typically gradient step)

Iterative algorithms (e.g., EM, variational learning).

unobserved

  • bserved

unobserved

  • bserved
  • bserved

inferred inference

Task: learn .

slide-5
SLIDE 5

Our setting

Ground truth model.

latent variables

  • bserved variables

Task: learn model from samples.

(synthetic setting) non-overparameterized

  • verparameterized
slide-6
SLIDE 6

Our question

How does overparameterization affect the recovery of ground truth latent variables?

A ground truth latent variable is recovered if there exists a learned latent variable with the same parameters.

slide-7
SLIDE 7

Our finding

Demonstration through extensive experiments with:

  • noisy-OR network models
  • sparse coding models
  • neural PCFG models

The unmatched learned latent variables are typically redundant.

With overparameterization, the learned model recovers the ground truth latent variables more

  • ften than without overparameterization.

non-overparameterized

  • verparameterized
slide-8
SLIDE 8

Noisy-OR networks

Example: image model.

latent variables

  • bserved variables

latent variables

  • bserved variables

1 1 1

slide-9
SLIDE 9

noisy-OR network

Noisy-OR networks

Train using variational learning. recognition network Maximize the evidence lower bound (ELBO), alternating between gradient steps w.r.t and .

(in our experiments: logistic regression and independent Bernoulli)

slide-10
SLIDE 10

Noisy-OR networks: recovery

# recovered true latent variables % runs full recovery # latent variables of learned model # latent variables of learned model

Image model.

slide-11
SLIDE 11

Noisy-OR networks: recovery

Harm of extreme overparameterization is minor. Similar trends for held-out log-likelihood.

slide-12
SLIDE 12

Noisy-OR networks: unmatched latent variables

discarded or duplicates

Simple filtering step to recover ground truth:

  • eliminate latent variables with low prior or high failure
  • eliminate latent variables that are duplicates

high failure discarded low prior discarded

slide-13
SLIDE 13

Noisy-OR networks: algorithm variations

Overparameterization remains beneficial:

  • batch size: 20 → 1000
  • recognition network: logistic regression → independent Bernoulli

Suggests benefits are general when learning latent variable models with iterative algorithms.

slide-14
SLIDE 14

Noisy-OR networks: explanation

Hypothesis Actual finding

With overparameterization, more latent variables initialized close to ground truth latent variables. Then, the benefit is due to a “warm start”. Latent variables do not converge quickly to ground truth latent variables. In the beginning, undecided. Throughout, contentions.

slide-15
SLIDE 15

Noisy-OR networks: optimization stability

State of latent variables after 1/9, 2/9, and 3/9 of the first epoch. In the beginning, many latent variables are undecided.

both contend for the same ground truth latent variable

slide-16
SLIDE 16

Noisy-OR networks: optimization stability

State of latent variables after 10, 20, and 30 epochs. Throughout, latent variables often contend.

both contend for the same ground truth latent variable

slide-17
SLIDE 17

Sparse Coding

Linear model. Nonlinear model.

Neural PCFG

→ overparameterization gives better recovery → overparameterization gives better recovery Training with linear alternating minimization algorithm. Training with EM and neural network parameterization. Synthetic experiments. Semi-synthetic experiments with neural network parameterization. → simple filtering step

(similarity between parse trees)

slide-18
SLIDE 18

Discussion

Why is any of this surprising?

Typically, smaller models are more likely to be identifiable. However, our experiments show that larger models often make optimization easier and have an inductive bias toward ground truth recovery.

slide-19
SLIDE 19

Application

For practice: it is helpful to overparameterize. For theory: interesting phenomenon, may provide insights into learning and optimization.

slide-20
SLIDE 20

Future work

Study larger and more complex models, e.g., commonly used deep generative models.

  • Understand model identifiability.
  • Define overparameterization.
  • Define ground truth recovery

and design filtering steps.

slide-21
SLIDE 21

Thank you!

Our code is available at https://github.com/clinicalml/overparam.