A History-Based Auto-Tuning Framework for Fast and High-Performance DNN Design on GPU
暂无分享,去创建一个
Jun Yang | Jiandong Mu | Wei Zhang | Mengdi Wang | Wei Lin | Lanbo Li | Wei Lin | Jun Yang | Mengdi Wang | Jiandong Mu | Lanbo Li | Wei Zhang
[1] B. Roe,et al. Boosted decision trees as an alternative to artificial neural networks for particle identification , 2004, physics/0408124.
[2] J. Viers,et al. Hydrologic Variability of the Cosumnes River Floodplain , 2006 .
[3] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[4] Tianqi Chen,et al. XGBoost: A Scalable Tree Boosting System , 2016, KDD.
[5] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[6] Yi Yang,et al. Optimizing Memory Efficiency for Deep Convolutional Neural Networks on GPUs , 2016, SC16: International Conference for High Performance Computing, Networking, Storage and Analysis.
[7] Jaejin Lee,et al. Performance analysis of CNN frameworks for GPUs , 2017, 2017 IEEE International Symposium on Performance Analysis of Systems and Software (ISPASS).
[8] Xiaoming Chen,et al. Optimizing memory efficiency for convolution kernels on kepler GPUs , 2017, 2017 54th ACM/EDAC/IEEE Design Automation Conference (DAC).
[9] Song Han,et al. AMC: AutoML for Model Compression and Acceleration on Mobile Devices , 2018, ECCV.
[10] Wei Liu,et al. PocketFlow: An Automated Framework for Compressing and Accelerating Deep Neural Networks , 2018 .
[11] Haichen Shen,et al. TVM: An Automated End-to-End Optimizing Compiler for Deep Learning , 2018, OSDI.
[12] Hadi Esmaeilzadeh,et al. Reinforcement Learning and Adaptive Sampling for Optimized DNN Compilation , 2019, ArXiv.
[13] Yida Wang,et al. A Unified Optimization Approach for CNN Model Inference on Integrated GPUs , 2019, ICPP.
[14] Yida Wang,et al. Optimizing CNN Model Inference on CPUs , 2018, USENIX Annual Technical Conference.