You Look Twice: GaterNet for Dynamic Filter Selection in CNNs
暂无分享,去创建一个
Yang Li | Si Si | Samy Bengio | Zhourong Chen | Samy Bengio | Zhourong Chen | Si Si | Yang Li
[1] Geoffrey E. Hinton,et al. Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer , 2017, ICLR.
[2] Lukasz Kaiser,et al. Neural GPUs Learn Algorithms , 2015, ICLR.
[3] Serge J. Belongie,et al. Convolutional Networks with Adaptive Inference Graphs , 2017, International Journal of Computer Vision.
[4] Brendan J. Frey,et al. Adaptive dropout for training deep neural networks , 2013, NIPS.
[5] Jürgen Schmidhuber,et al. Deep Networks with Internal Selective Attention through Feedback Connections , 2014, NIPS.
[6] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[7] Joelle Pineau,et al. Conditional Computation in Neural Networks for faster models , 2015, ArXiv.
[8] Samy Bengio,et al. Discrete Autoencoders for Sequence Models , 2018, ArXiv.
[9] Yoshua Bengio,et al. Estimating or Propagating Gradients Through Stochastic Neurons for Conditional Computation , 2013, ArXiv.
[10] Yoshua Bengio,et al. Deep Learning of Representations: Looking Forward , 2013, SLSP.
[11] Samy Bengio,et al. Can Active Memory Replace Attention? , 2016, NIPS.
[12] Noam Shazeer,et al. HydraNets: Specialized Dynamic Architectures for Efficient Inference , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[13] Vijay Vasudevan,et al. Learning Transferable Architectures for Scalable Image Recognition , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[14] Xiaopeng Li,et al. Learning Sparse Deep Feedforward Networks via Tree Skeleton Expansion , 2018, ArXiv.
[15] Aurko Roy,et al. Fast Decoding in Sequence Models using Discrete Latent Variables , 2018, ICML.
[16] Itamar Arel,et al. Low-Rank Approximations for Conditional Feedforward Computation in Deep Neural Networks , 2013, ICLR.
[17] Jian Sun,et al. Identity Mappings in Deep Residual Networks , 2016, ECCV.
[18] Xavier Gastaldi,et al. Shake-Shake regularization , 2017, ArXiv.
[19] Gang Sun,et al. Squeeze-and-Excitation Networks , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[20] Geoffrey E. Hinton,et al. Adaptive Mixtures of Local Experts , 1991, Neural Computation.
[21] Yoshua Bengio,et al. Show, Attend and Tell: Neural Image Caption Generation with Visual Attention , 2015, ICML.
[22] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[23] Ludovic Denoyer,et al. Deep Sequential Neural Network , 2014, NIPS 2014.
[24] Geoffrey E. Hinton,et al. Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.
[25] Graham W. Taylor,et al. Improved Regularization of Convolutional Neural Networks with Cutout , 2017, ArXiv.
[26] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[27] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[28] Kilian Q. Weinberger,et al. Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[29] Michael S. Bernstein,et al. ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.
[30] Sergey Ioffe,et al. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning , 2016, AAAI.
[31] Geoffrey E. Hinton,et al. Visualizing Data using t-SNE , 2008 .