Event Generation and Statistical Sampling with Deep Generative - - PowerPoint PPT Presentation

event generation and statistical sampling with deep
SMART_READER_LITE
LIVE PREVIEW

Event Generation and Statistical Sampling with Deep Generative - - PowerPoint PPT Presentation

Event Generation and Statistical Sampling with Deep Generative Models Rob Verheyen Introduction Event generation is really hard! 2 Introduction Can we use deep neural networks to do event generation? Possible applications: Faster


slide-1
SLIDE 1

Event Generation and Statistical Sampling with Deep Generative Models

Rob Verheyen

slide-2
SLIDE 2

2

Introduction

Event generation is really hard!

slide-3
SLIDE 3

3

Introduction

Can we use deep neural networks to do event generation? Possible applications:

  • Faster
  • Data driven generators
  • Targeted event generation
slide-4
SLIDE 4

4

Introduction

Study of different types of unsupervised generative models

  • Generative Adversarial Networks
  • Variational Autoencoders
  • Buffer Variational Autoencoder

Can these networks be used for event generation?

slide-5
SLIDE 5

Generative Adversarial Networks (GANs)

slide-6
SLIDE 6

6

Generative Adversarial Networks

Two networks (Generator & Discriminator) that play a game against each other

slide-7
SLIDE 7

7

Generative Adversarial Networks

Loss function: Nash equilibrium:

pdata(x) = pgen(x)

D(x) = 1 2

slide-8
SLIDE 8

8

Generative Adversarial Networks

1812.04948

slide-9
SLIDE 9

Variational Autoencoders (VAEs)

slide-10
SLIDE 10

10

Autoencoders

  • Data is encoded into latent space
  • Dim of latent space is often lower than dim of data
slide-11
SLIDE 11

11

Variational Autoencoders

Add degree of randomness to training procedure

slide-12
SLIDE 12

12

Variational Autoencoders

Points in latent space are ordered

slide-13
SLIDE 13

13

Variational Autoencoders

Loss function

Mean squared error Kullback–Leibler divergence

LVAE = (1 − ) 1 N (~ xi − ~ yi)2 + DKL(N(µi, i), N(0, 1))

MSE : Gaussians prefer being very narrow KL Div: Gaussians prefer being close to

N(0, 1)

is a hyperparameter: tune by hand

β

slide-14
SLIDE 14

14

Information Buffer

The latent space representation of our datapoints are now ordered But we can do better: Create information buffer

p(z) = 1 n

n

X

i

N(µi, σi)

Representation of distribution of training data in latent space Normally, one would now sample from in latent space

N(0, 1)

slide-15
SLIDE 15

Results

slide-16
SLIDE 16

16

Toy Model

1 → 2

decay with uniform angles and no exact momentum conservation Trained on four-vectors

slide-17
SLIDE 17

17

Top pair production

  • One top required to decay leptonically

5 × 105

  • Number of training points
  • Jets with pT > 20 GeV

Event generation with the B-VAE is faster!

O(108)

MG5 aMC@NLO 6.3.2 + Pythia 8.2 + Delphes 3

slide-18
SLIDE 18

18

Top pair production

slide-19
SLIDE 19

19

Latent space distributions

Distributions are still Gaussian-like Some have sharp cutoffs: Unphysical events outside Information buffer very important!

slide-20
SLIDE 20

20

Latent Space Principal Component Analysis

slide-21
SLIDE 21

21

Latent Space Principal Component Analysis

slide-22
SLIDE 22

22

Possible Applications

Most direct application: Importance sampling for ME generation

σ ∝ Z dΦ|M(Φ)|2 = Z dΦ p(Φ)|M(Φ)|2 p(Φ)

Current methods: VEGAS Recent ML techniques: Latent variable models

1810.11509

e+e− → qg¯ q efficiency:

  • VEGAS: ~4%
  • LVM: ~ 65%
  • B-VAE: ???
slide-23
SLIDE 23

23

Applications?

  • Data-driven event generators
  • Targeted event generation
  • Applications outside High Energy Physics?
  • ???