Bayesian Nested Neural Networks for Uncertainty Calibration and Adaptive Compression
暂无分享,去创建一个
Antoni B. Chan | Yufei Cui | Ziquan Liu | Qiao Li | Yu Mao | Antoni B. Chan | Chun Jason Xue | Qiao Li | C. Xue | Yushun Mao | Yufei Cui | Ziquan Liu
[1] Chuang Gan,et al. Once for All: Train One Network and Specialize it for Efficient Deployment , 2019, ICLR.
[2] Tom Minka,et al. A* Sampling , 2014, NIPS.
[3] Peder A. Olsen,et al. Crowd Counting with Decomposed Uncertainty , 2019, AAAI.
[4] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[5] Max Welling,et al. Learning Sparse Neural Networks through L0 Regularization , 2017, ICLR.
[6] Kilian Q. Weinberger,et al. On Calibration of Modern Neural Networks , 2017, ICML.
[7] Han Zhang,et al. Improving GANs Using Optimal Transport , 2018, ICLR.
[8] Christopher Burgess,et al. beta-VAE: Learning Basic Visual Concepts with a Constrained Variational Framework , 2016, ICLR 2016.
[9] Didrik Nielsen,et al. Fast yet Simple Natural-Gradient Descent for Variational Inference in Complex Models , 2018, 2018 International Symposium on Information Theory and Its Applications (ISITA).
[10] Thomas S. Huang,et al. Universally Slimmable Networks and Improved Training Techniques , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[11] Max Welling,et al. Bayesian Compression for Deep Learning , 2017, NIPS.
[12] Antoni B. Chan,et al. Modeling Noisy Annotations for Crowd Counting , 2020, NeurIPS.
[13] Zoubin Ghahramani,et al. Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning , 2015, ICML.
[14] Alex Kendall,et al. Concrete Dropout , 2017, NIPS.
[15] Yee Whye Teh,et al. The Concrete Distribution: A Continuous Relaxation of Discrete Random Variables , 2016, ICLR.
[16] Sanjeev Khudanpur,et al. Parallel training of DNNs with Natural Gradient and Parameter Averaging , 2014 .
[17] Ryan P. Adams,et al. Learning Ordered Representations with Nested Dropout , 2014, ICML.
[18] Max Welling,et al. Multiplicative Normalizing Flows for Variational Bayesian Neural Networks , 2017, ICML.
[19] Trevor Darrell,et al. Learning Compact Convolutional Neural Networks with Nested Dropout , 2015, ICLR.
[20] Jiayu Li,et al. ADMM-NN: An Algorithm-Hardware Co-Design Framework of DNNs Using Alternating Direction Methods of Multipliers , 2018, ASPLOS.
[21] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[22] Kevin P. Murphy,et al. Machine learning - a probabilistic perspective , 2012, Adaptive computation and machine learning series.
[23] David Page,et al. Area under the Precision-Recall Curve: Point Estimates and Confidence Intervals , 2013, ECML/PKDD.
[24] Zheng O'Neill,et al. One for Many: Transfer Learning for Building HVAC Control , 2020, BuildSys@SenSys.
[25] Ben Poole,et al. Categorical Reparameterization with Gumbel-Softmax , 2016, ICLR.
[26] Lawrence Carin,et al. Large-Scale Bayesian Multi-Label Learning via Topic-Based Label Embeddings , 2015, NIPS.
[27] Song Han,et al. Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.
[28] Dmitry P. Vetrov,et al. Variational Dropout Sparsifies Deep Neural Networks , 2017, ICML.
[29] Marc G. Bellemare,et al. The Cramer Distance as a Solution to Biased Wasserstein Gradients , 2017, ArXiv.
[30] Roberto Cipolla,et al. Bayesian SegNet: Model Uncertainty in Deep Convolutional Encoder-Decoder Architectures for Scene Understanding , 2015, BMVC.
[31] Klaus H. Maier-Hein,et al. A Probabilistic U-Net for Segmentation of Ambiguous Images , 2018, NeurIPS.
[32] Charles Blundell,et al. Simple and Scalable Predictive Uncertainty Estimation using Deep Ensembles , 2016, NIPS.
[33] Zhuowen Tu,et al. Aggregated Residual Transformations for Deep Neural Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[34] Mark Sandler,et al. MobileNetV2: Inverted Residuals and Linear Bottlenecks , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[35] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[36] Maria L. Rizzo,et al. Energy statistics: A class of statistics based on distances , 2013 .
[37] E. Gumbel. Statistical Theory of Extreme Values and Some Practical Applications : A Series of Lectures , 1954 .
[38] Xiangyu Zhang,et al. Channel Pruning for Accelerating Very Deep Neural Networks , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[39] Wei Wen,et al. DeepHoyer: Learning Sparser Neural Network with Differentiable Scale-Invariant Sparsity Measures , 2019, ICLR.
[40] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[41] Songhwai Oh,et al. NestedNet: Learning Nested Sparse Structures in Deep Neural Networks , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[42] Kevin Gimpel,et al. A Baseline for Detecting Misclassified and Out-of-Distribution Examples in Neural Networks , 2016, ICLR.
[43] Ariel D. Procaccia,et al. Variational Dropout and the Local Reparameterization Trick , 2015, NIPS.
[44] Iain Murray,et al. Ordering Dimensions with Nested Dropout Normalizing Flows , 2020, ArXiv.
[45] Stephen M. Moore,et al. The Cancer Imaging Archive (TCIA): Maintaining and Operating a Public Information Repository , 2013, Journal of Digital Imaging.
[46] Quoc V. Le,et al. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks , 2019, ICML.
[47] Yarin Gal,et al. Uncertainty in Deep Learning , 2016 .
[48] Tei-Wei Kuo,et al. Fully Nested Neural Network for Adaptive Compression and Quantization , 2020, IJCAI.