Scaling Up a Multispectral Resnet-50 to 128 GPUs
暂无分享,去创建一个
Matthias Book | Morris Riedel | Gabriele Cavallaro | Rocco Sedona | Jenia Jitsev | Alexandre Strube | M. Riedel | J. Jitsev | Rocco Sedona | Gabriele Cavallaro | A. Strube | Matthias Book
[1] Alex Krizhevsky,et al. One weird trick for parallelizing convolutional neural networks , 2014, ArXiv.
[2] Geoff Holmes,et al. Classifier chains for multi-label classification , 2009, Machine Learning.
[3] Yang You,et al. Large Batch Training of Convolutional Networks , 2017, 1708.03888.
[4] Begüm Demir,et al. Bigearthnet: A Large-Scale Benchmark Archive for Remote Sensing Image Understanding , 2019, IGARSS 2019 - 2019 IEEE International Geoscience and Remote Sensing Symposium.
[5] Jon Atli Benediktsson,et al. Remote Sensing Big Data Classification with High Performance Distributed Deep Learning , 2019, Remote Sensing.
[6] James Demmel,et al. Large Batch Optimization for Deep Learning: Training BERT in 76 minutes , 2019, ICLR.
[7] Alexander Sergeev,et al. Horovod: fast and easy distributed deep learning in TensorFlow , 2018, ArXiv.
[8] Torsten Hoefler,et al. Demystifying Parallel and Distributed Deep Learning: An In-Depth Concurrency Analysis. , 2018 .
[9] Dorian Krause,et al. JURECA: Modular supercomputer at Jülich Supercomputing Centre , 2018, Journal of large-scale research facilities JLSRF.
[10] Josef Aschbacher,et al. ESA’s Earth Observation Strategy and Copernicus , 2017 .
[11] Elad Hoffer,et al. Train longer, generalize better: closing the generalization gap in large batch training of neural networks , 2017, NIPS.
[12] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[13] Kaiming He,et al. Accurate, Large Minibatch SGD: Training ImageNet in 1 Hour , 2017, ArXiv.