暂无分享,去创建一个
Xianglong Liu | Bo Li | Bo Lang | Adams Wei Yu | Lei Huang | Lei Huang | B. Lang | Xianglong Liu | A. Yu | Bo Li
[1] Yonina C. Eldar,et al. MMSE whitening and subspace whitening , 2003, IEEE Transactions on Information Theory.
[2] Tim Salimans,et al. Weight Normalization: A Simple Reparameterization to Accelerate Training of Deep Neural Networks , 2016, NIPS.
[3] Jaime G. Carbonell,et al. Normalized Gradient with Adaptive Stepsize Method for Deep Neural Network Training , 2017, ArXiv.
[4] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[5] Cristian Sminchisescu,et al. Training Deep Networks with Structured Layers by Matrix Backpropagation , 2015, ArXiv.
[6] Jack J. Dongarra,et al. Towards dense linear algebra for hybrid GPU accelerated manycore systems , 2009, Parallel Comput..
[7] Jian Sun,et al. Identity Mappings in Deep Residual Networks , 2016, ECCV.
[8] Surya Ganguli,et al. Exact solutions to the nonlinear dynamics of learning in deep linear neural networks , 2013, ICLR.
[9] Minh N. Do,et al. Special paraunitary matrices, Cayley transform, and multidimensional orthogonal filter banks , 2006, IEEE Transactions on Image Processing.
[10] Jérôme Malick,et al. Projection-like Retractions on Matrix Manifolds , 2012, SIAM J. Optim..
[11] Les E. Atlas,et al. Full-Capacity Unitary Recurrent Neural Networks , 2016, NIPS.
[12] Anders Krogh,et al. A Simple Weight Decay Can Improve Generalization , 1991, NIPS.
[13] Moustapha Cissé,et al. Parseval Networks: Improving Robustness to Adversarial Examples , 2017, ICML.
[14] F. Xavier Roca,et al. Regularizing CNNs with Locally Constrained Decorrelations , 2016, ICLR.
[15] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[16] Victor D. Dorobantu,et al. DizzyRNN: Reparameterizing Recurrent Neural Networks for Norm-Preserving Backpropagation , 2016, ArXiv.
[17] Yoshua Bengio,et al. Extracting and composing robust features with denoising autoencoders , 2008, ICML '08.
[18] Toshihisa Tanaka,et al. Empirical Arithmetic Averaging Over the Compact Stiefel Manifold , 2013, IEEE Transactions on Signal Processing.
[19] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[20] Levent Tunçel,et al. Optimization algorithms on matrix manifolds , 2009, Math. Comput..
[21] Wotao Yin,et al. A feasible method for optimization with orthogonality constraints , 2013, Math. Program..
[22] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[23] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[24] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[25] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[26] Hemant D. Tagare,et al. Notes on Optimization on Stiefel Manifolds , 2011 .
[27] Shiliang Pu,et al. All You Need is Beyond a Good Init: Exploring Better Solution for Training Extremely Deep Convolutional Neural Networks with Orthonormality and Modulation , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[28] Christopher Joseph Pal,et al. On orthogonality and learning recurrent networks with long term dependencies , 2017, ICML.
[29] Tapani Raiko,et al. Semi-supervised Learning with Ladder Networks , 2015, NIPS.
[30] Silvere Bonnabel,et al. Stochastic Gradient Descent on Riemannian Manifolds , 2011, IEEE Transactions on Automatic Control.
[31] Yann LeCun,et al. Effiicient BackProp , 1996, Neural Networks: Tricks of the Trade.
[32] Frank Hutter,et al. SGDR: Stochastic Gradient Descent with Restarts , 2016, ArXiv.
[33] Michael S. Bernstein,et al. ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.
[34] James Bailey,et al. Efficient Orthogonal Parametrisation of Recurrent Neural Networks Using Householder Reflections , 2016, ICML.
[35] Takayuki Okatani,et al. Optimization on Submanifolds of Convolution Kernels in CNNs , 2016, ArXiv.
[36] Basura Fernando,et al. Generalized BackPropagation, Étude De Cas: Orthogonality , 2016, ArXiv.
[37] Yoshua Bengio,et al. Unitary Evolution Recurrent Neural Networks , 2015, ICML.
[38] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[39] Zoubin Ghahramani,et al. Unifying linear dimensionality reduction , 2014, 1406.0873.
[40] Shuicheng Yan,et al. Training Group Orthogonal Neural Networks with Privileged Information , 2017, IJCAI.
[41] Dumitru Erhan,et al. Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[42] Feiping Nie,et al. Convolutional 2D LDA for Nonlinear Dimensionality Reduction , 2017, IJCAI.
[43] Razvan Pascanu,et al. Natural Neural Networks , 2015, NIPS.
[44] Adams Wei Yu,et al. BLOCK-NORMALIZED GRADIENT METHOD: AN EMPIRICAL STUDY FOR TRAINING DEEP NEURAL NETWORK , 2018 .