Transfer-learning-based Autotuning using Gaussian Copula

As diverse high-performance computing (HPC) systems are built, many opportunities arise for applications to solve larger problems than ever before. Given the significantly increased complexity of these HPC systems and application tuning, empirical performance tuning, such as autotuning, has emerged as a promising approach in recent years. Despite its effectiveness, autotuning is often a computationally expensive approach. Transfer learning (TL)-based autotuning seeks to address this issue by leveraging the data from prior tuning. Current TL methods for autotuning spend significant time modeling the relationship between parameter configurations and performance, which is ineffective for few-shot (that is, few empirical evaluations) tuning on new tasks. We introduce the first generative TL-based autotuning approach based on the Gaussian copula (GC) to model the high-performing regions of the search space from prior data and then generate high-performing configurations for new tasks. This allows a sampling-based approach that maximizes few-shot performance and provides the first probabilistic estimation of the few-shot budget for effective TL-based autotuning. We compare our generative TL approach with state-of-the-art autotuning techniques on several benchmarks. We find that the GC is capable of achieving 64.37% of peak few-shot performance in its first evaluation. Furthermore, the GC model can determine a few-shot transfer budget that yields up to 33.39× speedup, a dramatic improvement over the 20.58× speedup using prior techniques.

[1]  Mary W. Hall,et al.  ytopt: Autotuning Scientific Applications for Energy Efficiency at Large Scales , 2023, ArXiv.

[2]  T. J. Anande,et al.  Generative adversarial networks for network traffic feature generation , 2023, International Journal of Computers and Applications.

[3]  Zheng Zhang,et al.  A fast parameter tuning framework via transfer learning and multi-objective bayesian optimization , 2022, DAC.

[4]  Vijay Gadepally,et al.  Bliss: auto-tuning complex applications using a pool of diverse lightweight learning models , 2021, PLDI.

[5]  Hongyu Zhang,et al.  Efficient Compiler Autotuning via Bayesian Optimization , 2021, 2021 IEEE/ACM 43rd International Conference on Software Engineering (ICSE).

[6]  James Demmel,et al.  GPTune: multitask learning for autotuning exascale applications , 2021, PPoPP.

[7]  Carlos H. A. Costa,et al.  You Only Run Once: Spark Auto-Tuning From a Single Run , 2020, IEEE Transactions on Network and Service Management.

[8]  Prasanna Balaprakash,et al.  Autotuning PolyBench Benchmarks with LLVM Clang/Polly Loop Optimization Pragmas Using Bayesian Optimization , 2020, 2020 IEEE/ACM Performance Modeling, Benchmarking and Simulation of High Performance Computer Systems (PMBS).

[9]  David Salinas,et al.  A Quantile-based Approach for Hyperparameter Transfer Learning , 2020, ICML.

[10]  Andrew Rice,et al.  Tuneful: An Online Significance-Aware Configuration Tuner for Big Data Analytics , 2020, ArXiv.

[11]  James Demmel,et al.  Multitask and Transfer Learning for Autotuning Exascale Applications , 2019, ArXiv.

[12]  Lei Xu,et al.  Modeling Tabular data using Conditional GAN , 2019, NeurIPS.

[13]  Prasanna Balaprakash,et al.  Autotuning in High-Performance Computing Applications , 2018, Proceedings of the IEEE.

[14]  Tao Wang,et al.  Bootstrapping Parameter Space Exploration for Fast Tuning , 2018, ICS.

[15]  Rushil Anirudh,et al.  Performance Modeling under Resource Constraints Using Deep Transfer Learning , 2017, SC17: International Conference for High Performance Computing, Networking, Storage and Analysis.

[16]  Christian Kästner,et al.  Transfer Learning for Improving Model Predictions in Highly Configurable Software , 2017, 2017 IEEE/ACM 12th International Symposium on Software Engineering for Adaptive and Self-Managing Systems (SEAMS).

[17]  Kalyan Veeramachaneni,et al.  The Synthetic Data Vault , 2016, 2016 IEEE International Conference on Data Science and Advanced Analytics (DSAA).

[18]  Jeffrey K. Hollingsworth,et al.  ANGEL: A Hierarchical Approach to Multi-Objective Online Auto-Tuning , 2015, ROSS@HPDC.

[19]  Christian Lengauer,et al.  Polly - Performing Polyhedral Optimizations on a Low-Level Intermediate Representation , 2012, Parallel Process. Lett..

[20]  Una-May O'Reilly,et al.  Siblingrivalry: online autotuning through local competitions , 2012, CASES '12.

[21]  Ananta Tiwari,et al.  Online Adaptive Code Generation and Tuning , 2011, 2011 IEEE International Parallel & Distributed Processing Symposium.

[22]  Hedibert Freitas Lopes,et al.  Copula, marginal distributions and model selection: a Bayesian note , 2008, Stat. Comput..

[23]  Joshua D. Knowles,et al.  ParEGO: a hybrid algorithm with on-line landscape approximation for expensive multiobjective optimization problems , 2006, IEEE Transactions on Evolutionary Computation.

[24]  R. A. Leibler,et al.  On Information and Sufficiency , 1951 .

[25]  C. Varin,et al.  Gaussian Copula Marginal Regression , 2012 .