暂无分享,去创建一个
Karima Benatchba | Taha Arbaoui | Riyadh Baghdadi | Kamel Abdous | Saman Amarasinghe | Massinissa Merouani | Mohamed-Hicham Leghettas | S. Amarasinghe | K. Benatchba | Riyadh Baghdadi | Massinissa Merouani | T. Arbaoui | K. Abdous | Mohamed-Hicham Leghettas
[1] Albert Cohen,et al. Improved loop tiling based on the removal of spurious false dependences , 2013, TACO.
[2] Thierry Moreau,et al. Learning to Optimize Tensor Programs , 2018, NeurIPS.
[3] Albert Cohen,et al. Minimal Unroll Factor for Code Generation of Software Pipelining , 2012, International Journal of Parallel Programming.
[4] Paul Feautrier,et al. Polyhedron Model , 2011, Encyclopedia of Parallel Computing.
[5] Nicholay Topin,et al. Super-convergence: very fast training of neural networks using large learning rates , 2018, Defense + Commercial Sensing.
[6] Monica S. Lam,et al. A Loop Transformation Theory and an Algorithm to Maximize Parallelism , 1991, IEEE Trans. Parallel Distributed Syst..
[7] Albert Cohen,et al. PENCIL Language Specification , 2015 .
[8] Frank Hutter,et al. Fixing Weight Decay Regularization in Adam , 2017, ArXiv.
[9] Michael Carbin,et al. Ithemal: Accurate, Portable and Fast Basic Block Throughput Estimation using Deep Neural Networks , 2018, ICML.
[10] Natalia Gimelshein,et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.
[11] Albert Cohen,et al. Hybrid Hexagonal/Classical Tiling for GPUs , 2014, CGO '14.
[12] Albert Cohen,et al. GRAPHITE Two Years After First Lessons Learned From Real-World Polyhedral Compilation , 2010 .
[13] Christian Lengauer,et al. Polly - Performing Polyhedral Optimizations on a Low-Level Intermediate Representation , 2012, Parallel Process. Lett..
[14] Sanjay V. Rajopadhye,et al. Optimizing memory usage in the polyhedral model , 2000, TOPL.
[15] Uday Bondhugula,et al. A practical automatic polyhedral parallelizer and locality optimizer , 2008, PLDI '08.
[16] Michael F. P. O'Boyle,et al. MILEPOST GCC: machine learning based research compiler , 2008 .
[17] Michael Carbin,et al. TIRAMISU: A Polyhedral Compiler for Dense and Sparse Deep Learning , 2020, ArXiv.
[18] Chris Cummins,et al. End-to-End Deep Learning of Optimization Heuristics , 2017, 2017 26th International Conference on Parallel Architectures and Compilation Techniques (PACT).
[19] P. Cochat,et al. Et al , 2008, Archives de pediatrie : organe officiel de la Societe francaise de pediatrie.
[20] Paul Feautrier,et al. Array expansion , 1988, ICS '88.
[21] Neural Network Assisted Tile Size Selection , 2010 .
[22] Frédo Durand,et al. Decoupling algorithms from schedules for easy optimization of image processing pipelines , 2012, ACM Trans. Graph..
[23] Shoaib Kamil,et al. Tiramisu: A Polyhedral Compiler for Expressing Fast and Portable Code , 2018, 2019 IEEE/ACM International Symposium on Code Generation and Optimization (CGO).
[24] Michael F. P. O'Boyle,et al. Automatic optimization of thread-coarsening for graphics processors , 2014, 2014 23rd International Conference on Parallel Architecture and Compilation (PACT).
[25] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[26] Albert Cohen,et al. Polyhedral-Model Guided Loop-Nest Auto-Vectorization , 2009, 2009 18th International Conference on Parallel Architectures and Compilation Techniques.
[27] Albert Cohen,et al. Tensor Comprehensions: Framework-Agnostic High-Performance Machine Learning Abstractions , 2018, ArXiv.
[28] Albert Cohen,et al. PENCIL: Towards a Platform-Neutral Compute Intermediate Language for DSLs , 2013, HiPC 2013.
[29] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[30] Paul Feautrier,et al. Automatic Storage Management for Parallel Programs , 1998, Parallel Comput..
[31] Elnar Hajiyev,et al. PENCIL: A Platform-Neutral Compute Intermediate Language for Accelerator Programming , 2015, 2015 International Conference on Parallel Architecture and Compilation (PACT).
[32] Frédo Durand,et al. Learning to optimize halide with tree search and random programs , 2019, ACM Trans. Graph..
[33] Yoshua Bengio,et al. Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.