Generative Networks for Precision Enthusiasts

Generative networks are opening new avenues in fast event generation for the LHC. We show how generative flow networks can reach percent-level precision for kinematic distributions, how they can be trained jointly with a discriminator, and how this discriminator improves the generation. Our joint training relies on a novel coupling of the two networks which does not require a Nash equilibrium. We then estimate the generation uncertainties through a Bayesian network setup and through conditional data augmentation, while the discriminator ensures that there are no systematic inconsistencies compared to the training data

[1]  F. Siegert,et al.  Accelerating Monte Carlo event generation -- rejection sampling using neural network event-weight estimates , 2021, SciPost Physics.

[2]  David Shih,et al.  CaloFlow: Fast and Accurate Generation of Calorimeter Showers with Normalizing Flows , 2021, ArXiv.

[3]  Thong Q. Nguyen,et al.  Analysis-Specific Fast Simulation at the LHC with Deep Learning , 2021, Computing and Software for Big Science.

[4]  Rob Verheyen,et al.  Phase space sampling and inference from weighted events with autoregressive flows , 2020, SciPost Physics.

[5]  G. Kasieczka,et al.  Getting High: High Fidelity Simulation of High Granularity Calorimeters with High Speed , 2020, Computing and Software for Big Science.

[6]  Ivan Kobyzev,et al.  Normalizing Flows: An Introduction and Review of Current Methods , 2020, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[7]  L. Santi,et al.  Towards a computer vision particle flow , 2020, The European Physical Journal C.

[8]  Michelle P. Kuchera,et al.  Simulation of electron-proton scattering events by a Feature-Augmented and Transformed Generative Adversarial Network (FAT-GAN) , 2020, IJCAI.

[9]  Damian Podareanu,et al.  Event generation and statistical sampling for physics with deep generative models and a density information buffer , 2019, Nature Communications.

[10]  Stefan T. Radev,et al.  Measuring QCD Splittings with Invertible Networks , 2020, SciPost Physics.

[11]  A. Butter,et al.  How to GAN Event Unweighting , 2020, 2012.07873.

[12]  Matthew D. Klimek,et al.  Improved neural network Monte Carlo simulation , 2020, 2009.07819.

[13]  G. Kasieczka,et al.  DCTRGAN: improving the precision of generative models with reweighting , 2020, Journal of Instrumentation.

[14]  G. Kasieczka,et al.  GANplifying event samples , 2020, SciPost Physics.

[15]  Ullrich Kothe,et al.  Invertible networks or partons to detector and back again , 2020, 2006.06685.

[16]  G. Kasieczka,et al.  Per-object systematics using deep-learned calibration , 2020, SciPost Physics.

[17]  S. Badger,et al.  Using neural networks for efficient evaluation of high multiplicity scattering amplitudes , 2020, Journal of High Energy Physics.

[18]  H. Schulz,et al.  Event generation with normalizing flows , 2020, Physical Review D.

[19]  S. Schumann,et al.  Exploring phase space with Neural Importance Sampling , 2020, SciPost Physics.

[20]  Christina Gao,et al.  i- flow: High-dimensional integration and sampling with normalizing flows , 2020, Mach. Learn. Sci. Technol..

[21]  Wei Wei,et al.  Calorimetry with deep learning: particle simulation and reconstruction for collider physics , 2019, The European Physical Journal C.

[22]  G. Kasieczka,et al.  How to GAN away Detector Effects , 2019, SciPost Physics.

[23]  B. Nachman A guide for deploying Deep Learning in LHC searches: How to achieve optimality and account for uncertainty , 2019, SciPost Physics.

[24]  Jennifer Thompson,et al.  Deep-learning jets with uncertainties and more , 2019, SciPost Physics.

[25]  A. Butter,et al.  How to GAN event subtraction , 2019, SciPost Physics Core.

[26]  Natalia Gimelshein,et al.  PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.

[27]  Tilman Plehn,et al.  How to GAN LHC events , 2019, SciPost Physics.

[28]  Diederik P. Kingma,et al.  An Introduction to Variational Autoencoders , 2019, Found. Trends Mach. Learn..

[29]  Iain Murray,et al.  Cubic-Spline Flows , 2019, ICML 2019.

[30]  Steffen Schumann,et al.  Event generation with Sherpa 2.2 , 2019, SciPost Physics.

[31]  Sana Ketabchi Haghighat,et al.  DijetGAN: a Generative-Adversarial Network approach for the simulation of QCD dijet events at the LHC , 2019, Journal of High Energy Physics.

[32]  Enrico Bothmann,et al.  Reweighting a parton shower using a neural network: the final-state case , 2018, Journal of High Energy Physics.

[33]  Martin Erdmann,et al.  Precise Simulation of Electromagnetic Calorimeter Showers Using a Wasserstein Generative Adversarial Network , 2018, Computing and Software for Big Science.

[34]  C. Frye,et al.  JUNIPR: a framework for unsupervised machine learning in particle physics , 2018, The European Physical Journal C.

[35]  Nicholay Topin,et al.  Super-convergence: very fast training of neural networks using large learning rates , 2018, Defense + Commercial Sensing.

[36]  J. Monk,et al.  Deep learning as a parton shower , 2018, Journal of High Energy Physics.

[37]  Francesco Pandolfi,et al.  Fast and Accurate Simulation of Particle Detectors Using Generative Adversarial Networks , 2018, Computing and Software for Big Science.

[38]  Martin Erdmann,et al.  Generating and Refining Particle Detector Simulations Using the Wasserstein Distance in Adversarial Networks , 2018, Computing and Software for Big Science.

[39]  Tom White,et al.  Generative Adversarial Networks: An Overview , 2017, IEEE Signal Processing Magazine.

[40]  Stefano Ermon,et al.  Flow-GAN: Combining Maximum Likelihood and Adversarial Learning in Generative Models , 2017, AAAI.

[41]  Benjamin Nachman,et al.  Accelerating Science with Generative Adversarial Networks: An Application to 3D Particle Showers in Multilayer Calorimeters. , 2017, Physical review letters.

[42]  Michela Paganini,et al.  CaloGAN: Simulating 3D High Energy Particle Showers in Multi-Layer Electromagnetic Calorimeters with Generative Adversarial Networks , 2017, ArXiv.

[43]  Alex Kendall,et al.  What Uncertainties Do We Need in Bayesian Deep Learning for Computer Vision? , 2017, NIPS.

[44]  Luke de Oliveira,et al.  Learning Particle Physics by Example: Location-Aware Generative Adversarial Networks for Physics Synthesis , 2017, Computing and Software for Big Science.

[45]  David M. Blei,et al.  Variational Inference: A Review for Statisticians , 2016, ArXiv.

[46]  Yarin Gal,et al.  Uncertainty in Deep Learning , 2016 .

[47]  Shakir Mohamed,et al.  Variational Inference with Normalizing Flows , 2015, ICML.

[48]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[49]  M. Cacciari,et al.  FastJet user manual , 2011, 1111.6097.

[50]  M. Cacciari,et al.  The anti-$k_t$ jet clustering algorithm , 2008, 0802.1189.

[51]  F. Krauss,et al.  QCD Matrix Elements + Parton Showers , 2001, hep-ph/0109231.

[52]  Radford M. Neal Bayesian learning for neural networks , 1995 .

[53]  David Mackay,et al.  Probable networks and plausible predictions - a review of practical Bayesian methods for supervised neural networks , 1995 .