adversarial event generator tuning with bayesian
play

Adversarial event generator tuning with Bayesian Optimization Maxim - PowerPoint PPT Presentation

Adversarial event generator tuning with Bayesian Optimization Maxim Borisyak, Andrey Ustyuzhanin National Research University Higher School of Economics (HSE) July 7, 2018 Event Generator Tuning Intro We consider problem of tuning parameters


  1. Adversarial event generator tuning with Bayesian Optimization Maxim Borisyak, Andrey Ustyuzhanin National Research University Higher School of Economics (HSE) July 7, 2018

  2. Event Generator Tuning

  3. Intro We consider problem of tuning parameters of event generators to ’real’ data: • generating samples is expensive; • generator is non-differentiable. Working example: Pythia 8 generator. 2

  4. Approach I • Bayesian Optimization on the objective: n bins • additional assumptions on distributions are required to guarantee convergence; 3 • two histogram for each parameter: data i and MC i ; ( data i − MC i ) 2 χ 2 = ∑ σ 2 data , i + σ 2 MC , i i = 1

  5. Approach II • an adversarial objective: • Variational Optimization to search for distribution over generator parameters . 4 Wasserstein ( F real , F θ ) = sup x ∼ F real d ( x ) − E x ∼ F θ d ( x ) E d ∈ L 1

  6. Assumptions and goals We consider Adversarial Bayesian Optimization: • no additional restrictions on distribution shapes; Our primary concern is time complexity : • sampling from the target event generator is expensive; • number of generator calls dominates overall complexity; • minimizing number of event generator calls ; • there is a configuration of generator that perfectly matches ’real’ data. 5

  7. Adversarial Bayesian Optimization

  8. Adversarial Objective Jensen-Shannon distance: • Jensen-Shannon distance can be approximated by a classifier. f 6 2 [ P ( x ) Q ( x ) ] JS ( P , Q ) = log 2 + 1 x ∼ P log E P ( x ) + Q ( x ) + E x ∼ Q log = P ( x ) + Q ( x ) log 2 − min cross - entropy ( f , P , Q )

  9. Multi-Stage Adversarial Bayesian Optimization • sequence of classifier models with increasing power: 7 F 1 ⊆ F 2 ⊆ · · · ⊆ F m = F • classifier F i associated with ’pseudo’ JS distance: pJS i ( P , Q ) = log 2 − min f ∈F i cross - entropy ( f , P , Q ) pJS 1 ( P , Q ) ≤ pJS 2 ( P , Q ) ≤ · · · ≤ pJS m ( P , Q ) = JS ( P , Q ); pJS i ( P , Q ) ≥ 0 = ⇒ pJS i + 1 ( P , Q ) ≥ 0

  10. Multi-Stage Adversarial Bayesian Optimization • ’weak’ classifiers tend to require less samples; • ’weak’ classifiers can be used to rapidly explore search space; • these results are constraints for a more powerful classifier. 8 pJS i ( P , Q ) ≥ 0 = ⇒ pJS i + 1 ( P , Q ) ≥ 0

  11. Multi-Stage Adversarial Bayesian Optimization 3: 4: 5: end for 9 1: model 1 = unconstrained BO on pJS 1 ( data , generator θ ) 2: for k = 2 , . . . , m do ( ) constraint k ( θ ) = P pJS k − 1 ≤ 0 | θ, model k − 1 model k = BO on pJS k ( data , · ) s.t. constraint j ( theta ) > τ , j = 0 , . . . , k − 1

  12. Experiments

  13. Experiment We follow problem statement from Ilten P, Williams M, Yang Y. Event generator tuning using Bayesian optimization. Journal of Instrumentation. 2017 Apr 27;12(04):P04028. • values of Monash tune as parameters of the ’real’ distribution; • 2-stage Adversarial Bayesian Optimization; • number of samples required to avoid overfitting of the classifier is measured. 10 • e + e − modeled by Pythia 8 ;

  14. Experiment 1 Target generator options: • alphaSvalue . 11

  15. Experiment 1: stage 1 12

  16. Experiment 1: stage 1 13

  17. Experiment 1: stage 2 14

  18. Experiment 1: stage 2 15

  19. Experiment 1: single stage 16

  20. Experiment 1: results 17

  21. Experiment 2 Target generator options: • bLund ; • sigma ; • aExtraSQuark ; • aExtraDiQuark ; • rFactC ; • rFactB . Second group of varables from Ilten P, Williams M, Yang Y. Event generator tuning using Bayesian optimization. Journal of Instrumentation. 2017 Apr 27;12(04):P04028. 18

  22. Experiment 2: results 19

  23. Summary

  24. Summary • Adversarial Bayesian Optimization is a promising tool for tuning event generators; • Multi-stage Adversarial Bayesian Optimization utilizes ’weak’ classifiers to incrementally constrain search space: • rapid exploration of search space on first stages; • late stages search for solution only among promising candidates; • reduction in overall cost of optimization. 20

  25. Backup 20

  26. Bayesian Adversarial Optimization 5: 8: end while 7: 1 1: initialize Bayesian Optimization train 6: 21 2: while not bored do 3: 4: θ ← askBO () X θ train , X θ test ← sample ( θ ) f ← train discriminator on X θ train and X real [∑ m ] i = 1 log f ( X θ, i i = 1 log( 1 − f ( X real , i L ← test ) + ∑ m test )) 2 · m tellBO ( θ, log 2 − L )

  27. Possible Caveats • constraints are observed by authors to mess with GP; • it is likely that the method would still work (modifying constraints) if classifiers are from the same family of algorithms; • it is possible, that BO with weak classifier carries no information about BO with a strong classifier. 22 • without assumption ∃ θ : JS ( generator ( θ ) , real ) = 0 :

  28. Expected Improvement with Constraints Problem: s.t. • improvement is impossible if constraints are violated: Gelbart, M.A., Snoek, J. and Adams, R.P., 2014. Bayesian optimization with unknown constraints. arXiv preprint arXiv:1403.5607. 23 EI ( x ) → min; g ( x ) ≥ 0 . CEI ( x ) = P ( g ( x ) ≥ 0 ) · EI ( x ) + P ( g ( x ) < 0 ) · 0 • constraints in our case: model i ( x ) ≤ 0 .

  29. Technical details • training set is incrementally extended until over-fitting becomes insignificant. • 2 stage ABO: • 1 stage: XGboost with 1 tree and max depth = 3; • 2 stage: XGboost with 20 tree and max depth = 6. 24

  30. Experiment 1 25

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend