暂无分享,去创建一个
[1] Roland Vollgraf,et al. Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.
[2] Neil Thompson,et al. The Decline of Computers As a General Purpose Technology: Why Deep Learning and the End of Moore’s Law are Fragmenting Computing , 2018 .
[3] Thomas S. Huang,et al. Image Classification Using Super-Vector Coding of Local Image Descriptors , 2010, ECCV.
[4] Stefano Soatto,et al. Mono3D++: Monocular 3D Vehicle Detection with Two-Scale 3D Hypotheses and Task Priors , 2019, AAAI.
[5] Erik F. Tjong Kim Sang,et al. Introduction to the CoNLL-2003 Shared Task: Language-Independent Named Entity Recognition , 2003, CoNLL.
[6] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[7] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[8] Rogério Schmidt Feris,et al. Big-Little Net: An Efficient Multi-Scale Feature Representation for Visual and Speech Recognition , 2018, ICLR.
[9] Quoc V. Le,et al. Self-Training With Noisy Student Improves ImageNet Classification , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[10] Adel Javanmard,et al. Theoretical Insights Into the Optimization Landscape of Over-Parameterized Shallow Neural Networks , 2017, IEEE Transactions on Information Theory.
[11] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[12] Quoc V. Le,et al. The Evolved Transformer , 2019, ICML.
[13] Andrew McCallum,et al. Energy and Policy Considerations for Deep Learning in NLP , 2019, ACL.
[14] Marcello Federico,et al. Report on the 11th IWSLT evaluation campaign , 2014, IWSLT.
[15] A. Krizhevsky. Convolutional Deep Belief Networks on CIFAR-10 , 2010 .
[16] Yi Luo,et al. All-optical machine learning using diffractive deep neural networks , 2018, Science.
[17] Jaehoon Lee,et al. Bayesian Deep Convolutional Networks with Many Channels are Gaussian Processes , 2018, ICLR.
[18] Mikhail Belkin,et al. Overfitting or perfect fitting? Risk bounds for classification and regression rules that interpolate , 2018, NeurIPS.
[19] Treebank Penn,et al. Linguistic Data Consortium , 1999 .
[20] R. Tibshirani. Regression Shrinkage and Selection via the Lasso , 1996 .
[21] Geoffrey E. Hinton,et al. Attend, Infer, Repeat: Fast Scene Understanding with Generative Models , 2016, NIPS.
[22] Mike Davies. Progress in Neuromorphic Computing : Drawing Inspiration from Nature for Gains in AI and Computing , 2019, VLSI-DAT.
[23] Quoc V. Le,et al. GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism , 2018, ArXiv.
[24] Andrew W. Cross,et al. Validating quantum computers using randomized model circuits , 2018, Physical Review A.
[25] Michael S. Bernstein,et al. ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.
[26] Rajat Raina,et al. Large-scale deep unsupervised learning using graphics processors , 2009, ICML '09.
[27] Florent Perronnin,et al. Fisher Kernels on Visual Vocabularies for Image Categorization , 2007, 2007 IEEE Conference on Computer Vision and Pattern Recognition.
[28] Philipp Koehn,et al. Findings of the 2014 Workshop on Statistical Machine Translation , 2014, WMT@ACL.
[29] Benjamin Recht,et al. Do ImageNet Classifiers Generalize to ImageNet? , 2019, ICML.
[30] Ran El-Yaniv,et al. Binarized Neural Networks , 2016, ArXiv.
[31] Alessandro Rudi,et al. Statistical Optimality of Stochastic Gradient Descent on Hard Learning Problems through Multiple Passes , 2018, NeurIPS.
[32] Michael I. Jordan,et al. Deep Transfer Learning with Joint Adaptation Networks , 2016, ICML.
[33] G LoweDavid,et al. Distinctive Image Features from Scale-Invariant Keypoints , 2004 .
[34] Yihong Gong,et al. Locality-constrained Linear Coding for image classification , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.
[35] Frank Rosenblatt,et al. Perceptron Simulation Experiments , 1960, Proceedings of the IRE.
[36] J. Welser,et al. Future Computing Hardware for AI , 2018, 2018 IEEE International Electron Devices Meeting (IEDM).
[37] Yi Yang,et al. More is Less: A More Complicated Network with Less Inference Complexity , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[38] N. Meinshausen,et al. LASSO-TYPE RECOVERY OF SPARSE REPRESENTATIONS FOR HIGH-DIMENSIONAL DATA , 2008, 0806.0145.
[39] Quoc V. Le,et al. Listen, attend and spell: A neural network for large vocabulary conversational speech recognition , 2015, 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[40] P. Cochat,et al. Et al , 2008, Archives de pediatrie : organe officiel de la Societe francaise de pediatrie.
[41] S. Ambrogio,et al. Confined PCM-based Analog Synaptic Devices offering Low Resistance-drift and 1000 Programmable States for Deep Learning , 2019, 2019 Symposium on VLSI Technology.
[42] Luca Carlone,et al. Outlier-Robust Spatial Perception: Hardness, General-Purpose Algorithms, and Guarantees , 2019, 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[43] Zhuowen Tu,et al. Aggregated Residual Transformations for Deep Neural Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[44] Jürgen Schmidhuber,et al. Connectionist temporal classification: labelling unsegmented sequence data with recurrent neural networks , 2006, ICML.
[45] Kurt Hornik,et al. Multilayer feedforward networks are universal approximators , 1989, Neural Networks.
[46] Amos J. Storkey,et al. School of Informatics, University of Edinburgh , 2022 .
[47] Cong Xu,et al. Coordinating Filters for Faster Deep Neural Networks , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[48] Chuang Gan,et al. Neural-Symbolic VQA: Disentangling Reasoning from Vision and Language Understanding , 2018, NeurIPS.
[49] Max Tegmark,et al. AI Feynman: A physics-inspired method for symbolic regression , 2019, Science Advances.
[50] G. Griffin,et al. Caltech-256 Object Category Dataset , 2007 .
[51] Yang Song,et al. The iNaturalist Species Classification and Detection Dataset , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[52] Song Han,et al. Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.
[53] Honglak Lee,et al. An Analysis of Single-Layer Networks in Unsupervised Feature Learning , 2011, AISTATS.
[54] Mikhail Belkin,et al. Reconciling modern machine-learning practice and the classical bias–variance trade-off , 2018, Proceedings of the National Academy of Sciences.
[55] Yann LeCun,et al. The mnist database of handwritten digits , 2005 .
[56] Chong Wang,et al. Deep Speech 2 : End-to-End Speech Recognition in English and Mandarin , 2015, ICML.
[57] Salim Roukos,et al. Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.
[58] Jonathan Krause,et al. Collecting a Large-scale Dataset of Fine-grained Cars , 2013 .
[59] Pietro Perona,et al. Microsoft COCO: Common Objects in Context , 2014, ECCV.
[60] F. Jelinek,et al. Continuous speech recognition by statistical methods , 1976, Proceedings of the IEEE.
[61] Tao Zhang,et al. A Survey of Model Compression and Acceleration for Deep Neural Networks , 2017, ArXiv.
[62] Gregory Cohen,et al. EMNIST: Extending MNIST to handwritten letters , 2017, 2017 International Joint Conference on Neural Networks (IJCNN).
[63] Christian Muise,et al. Learning Neural-Symbolic Descriptive Planning Models via Cube-Space Priors: The Voyage Home (to STRIPS) , 2020, ArXiv.
[64] Navdeep Jaitly,et al. Towards End-To-End Speech Recognition with Recurrent Neural Networks , 2014, ICML.
[65] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[66] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[67] Alex Lamb,et al. Deep Learning for Classical Japanese Literature , 2018, ArXiv.
[68] Boris Katz,et al. ObjectNet: A large-scale bias-controlled dataset for pushing the limits of object recognition models , 2019, NeurIPS.
[69] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[70] Jianguo Zhang,et al. The PASCAL Visual Object Classes Challenge , 2006 .
[71] David A. Patterson,et al. Computer Architecture: A Quantitative Approach , 1969 .
[72] Busra Celikkaya,et al. Comprehend Medical: A Named Entity Recognition and Relationship Extraction Web Service , 2019, 2019 18th IEEE International Conference On Machine Learning And Applications (ICMLA).
[73] Marvin Minsky,et al. Perceptrons: An Introduction to Computational Geometry , 1969 .
[74] Thomas Mensink,et al. Improving the Fisher Kernel for Large-Scale Image Classification , 2010, ECCV.
[75] Song Han,et al. ProxylessNAS: Direct Neural Architecture Search on Target Task and Hardware , 2018, ICLR.
[76] Mikhail Belkin,et al. Does data interpolation contradict statistical optimality? , 2018, AISTATS.
[77] Matti Pietikäinen,et al. Face Description with Local Binary Patterns: Application to Face Recognition , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[78] Ming Yang,et al. Large-scale image classification: Fast feature extraction and SVM training , 2011, CVPR 2011.
[79] Chuang Gan,et al. The Neuro-Symbolic Concept Learner: Interpreting Scenes Words and Sentences from Natural Supervision , 2019, ICLR.
[80] Catherine D. Schuman,et al. A Study of Complex Deep Learning Networks on High Performance, Neuromorphic, and Quantum Computers , 2016, 2016 2nd Workshop on Machine Learning in HPC Environments (MLHPC).
[81] Quoc V. Le,et al. Efficient Neural Architecture Search via Parameter Sharing , 2018, ICML.
[82] Chuang Gan,et al. Once for All: Train One Network and Specialize it for Efficient Deployment , 2019, ICLR.
[83] Pritish Narayanan,et al. Equivalent-accuracy accelerated neural-network training using analogue memory , 2018, Nature.
[84] Jian Zhang,et al. SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.