Distributed Training of Generative Adversarial Networks for Fast Detector Simulation

The simulation of the interaction of particles in High Energy Physics detectors is a computing intensive task. Since some level of approximation is acceptable, it is possible to implement fast simulation simplified models that have the advantage of being less computationally intensive. Here we present a fast simulation based on Generative Adversarial Networks (GANs). The model is constructed from a generative network describing the detector response and a discriminative network, trained in adversarial manner. The adversarial training process is compute-intensive and the application of a distributed approach becomes particularly important. We present scaling results of a data-parallel approach to distribute GANs training across multiple nodes on TACC’s Stampede2. The efficiency achieved was above 94% when going from 1 to 128 Xeon Scalable Processor nodes. We report on the accuracy of the generated samples and on the scaling of time-to-solution. We demonstrate how HPC installations could be utilized to globally optimize this kind of models leading to quicker research cycles and experimentation, thanks to their large computation power and excellent connectivity.

[1]  Yuan Yu,et al.  TensorFlow: A system for large-scale machine learning , 2016, OSDI.

[2]  Alex Krizhevsky,et al.  One weird trick for parallelizing convolutional neural networks , 2014, ArXiv.

[3]  Kilian Q. Weinberger,et al.  Snapshot Ensembles: Train 1, get M for free , 2017, ICLR.

[4]  Razvan Pascanu,et al.  Sharp Minima Can Generalize For Deep Nets , 2017, ICML.

[5]  Luke de Oliveira,et al.  Learning Particle Physics by Example: Location-Aware Generative Adversarial Networks for Physics Synthesis , 2017, Computing and Software for Big Science.

[6]  Soumith Chintala,et al.  Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks , 2015, ICLR.

[7]  Martín Abadi,et al.  TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems , 2016, ArXiv.

[8]  Maria Spiropulu,et al.  An MPI-Based Python Framework for Distributed Training with Keras , 2017, ArXiv.

[9]  Mike Williams,et al.  Efficient, reliable and fast high-level triggering using a bonsai boosted decision tree , 2012, 1210.6861.

[10]  P. Vannerem,et al.  Classifying LEP data with support vector algorithms. , 1999 .

[11]  S. Vallecorsa,et al.  Generative models for fast simulation , 2018, Journal of Physics: Conference Series.

[12]  P. Mendez Lorenzo,et al.  Three dimensional Generative Adversarial Networks for fast simulation , 2018, Journal of Physics: Conference Series.

[13]  Václav Snásel,et al.  Towards the Modeling of Atomic and Molecular Clusters Energy by Support Vector Regression , 2013, 2013 5th International Conference on Intelligent Networking and Collaborative Systems.

[14]  Philippe Canal,et al.  GeantV: from CPU to accelerators , 2016 .

[15]  Stefano Soatto,et al.  Entropy-SGD: biasing gradient descent into wide valleys , 2016, ICLR.

[16]  Shimon Whiteson,et al.  Machine learning for event selection in high energy physics , 2009, Eng. Appl. Artif. Intell..

[17]  Petr Savický,et al.  Methods for multidimensional event classification: A case study using images from a Cherenkov gamma-ray telescope , 2004 .

[18]  Jorge Nocedal,et al.  On Large-Batch Training for Deep Learning: Generalization Gap and Sharp Minima , 2016, ICLR.

[19]  Michela Paganini,et al.  CaloGAN: Simulating 3D High Energy Particle Showers in Multi-Layer Electromagnetic Calorimeters with Generative Adversarial Networks , 2017, ArXiv.