暂无分享,去创建一个
[1] Karl Pearson F.R.S.. LIII. On lines and planes of closest fit to systems of points in space , 1901 .
[2] H. Hotelling. Relations Between Two Sets of Variates , 1936 .
[3] Kurt Hornik,et al. Multilayer feedforward networks are universal approximators , 1989, Neural Networks.
[4] Geoffrey E. Hinton,et al. Bayesian Learning for Neural Networks , 1995 .
[5] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[6] Yoshua Bengio,et al. Extracting and composing robust features with denoising autoencoders , 2008, ICML '08.
[7] Geoffrey E. Hinton,et al. A Scalable Hierarchical Distributed Language Model , 2008, NIPS.
[8] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[9] Yoshua Bengio,et al. Why Does Unsupervised Pre-training Help Deep Learning? , 2010, AISTATS.
[10] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[11] Yee Whye Teh,et al. A fast and simple algorithm for training neural probabilistic language models , 2012, ICML.
[12] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[13] Thorsten Brants,et al. One billion word benchmark for measuring progress in statistical language modeling , 2013, INTERSPEECH.
[14] Stefan Carlsson,et al. CNN Features Off-the-Shelf: An Astounding Baseline for Recognition , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition Workshops.
[15] Nir Ailon,et al. Deep Metric Learning Using Triplet Network , 2014, SIMBAD.
[16] Quoc V. Le,et al. Semi-supervised Sequence Learning , 2015, NIPS.
[17] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[18] L. Floridi,et al. Data ethics , 2021, Effective Directors.
[19] Aapo Hyvärinen,et al. Unsupervised Feature Extraction by Time-Contrastive Learning and Nonlinear ICA , 2016, NIPS.
[20] Uri Shalit,et al. Learning Representations for Counterfactual Inference , 2016, ICML.
[21] Kihyuk Sohn,et al. Improved Deep Metric Learning with Multi-class N-pair Loss Objective , 2016, NIPS.
[22] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[23] Max Welling,et al. Causal Effect Inference with Deep Latent-Variable Models , 2017, NIPS 2017.
[24] Jascha Sohl-Dickstein,et al. SVCCA: Singular Vector Canonical Correlation Analysis for Deep Learning Dynamics and Interpretability , 2017, NIPS.
[25] Lukasz Kaiser,et al. Generating Wikipedia by Summarizing Long Sequences , 2018, ICLR.
[26] Oriol Vinyals,et al. Representation Learning with Contrastive Predictive Coding , 2018, ArXiv.
[27] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[28] Samy Bengio,et al. Insights on representational similarity in neural networks with canonical correlation , 2018, NeurIPS.
[29] R'emi Louf,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[30] Aapo Hyvärinen,et al. Nonlinear ICA Using Auxiliary Variables and Generalized Contrastive Learning , 2018, AISTATS.
[31] Inioluwa Deborah Raji,et al. Model Cards for Model Reporting , 2018, FAT.
[32] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[33] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[34] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[35] Ali Razavi,et al. Data-Efficient Image Recognition with Contrastive Predictive Coding , 2019, ICML.
[36] Aapo Hyvärinen,et al. Variational Autoencoders and Nonlinear ICA: A Unifying Framework , 2019, AISTATS.
[37] Mark Chen,et al. Language Models are Few-Shot Learners , 2020, NeurIPS.
[38] 俊一 甘利. 5分で分かる!? 有名論文ナナメ読み:Jacot, Arthor, Gabriel, Franck and Hongler, Clement : Neural Tangent Kernel : Convergence and Generalization in Neural Networks , 2020 .
[39] Diederik P. Kingma,et al. ICE-BeeM: Identifiable Conditional Energy-Based Deep Models , 2020, NeurIPS.
[40] Ullrich Köthe,et al. Disentanglement by Nonlinear ICA with General Incompressible-flow Networks (GIN) , 2020, ICLR.