Layer Skip Learning using LARS variables for 39% Faster Conversion Time and Lower Bandwidth
暂无分享,去创建一个
Masahiko Yoshimoto | Hiroshi Kawaguchi | Haruki Mori | Atsuki Inoue | Kazuki Yamada | Tetsuya Youkawa | Yuki Miyauchi | Shintato Izumi
[1] James Philbin,et al. FaceNet: A unified embedding for face recognition and clustering , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[2] Shintaro Izumi,et al. A layer-block-wise pipeline for memory and bandwidth reduction in distributed deep learning , 2017, 2017 IEEE 27th International Workshop on Machine Learning for Signal Processing (MLSP).
[3] Yang You,et al. Large Batch Training of Convolutional Networks , 2017, 1708.03888.
[4] Kilian Q. Weinberger,et al. Deep Networks with Stochastic Depth , 2016, ECCV.
[5] Qiang Xu,et al. ApproxANN: An approximate computing framework for artificial neural network , 2015, 2015 Design, Automation & Test in Europe Conference & Exhibition (DATE).
[6] Andrew Zisserman,et al. Return of the Devil in the Details: Delving Deep into Convolutional Nets , 2014, BMVC.
[7] Guigang Zhang,et al. Deep Learning , 2016, Int. J. Semantic Comput..
[8] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[9] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[10] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.