暂无分享,去创建一个
[1] H Neven,et al. A blueprint for demonstrating quantum supremacy with superconducting qubits , 2017, Science.
[2] Song Han,et al. Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.
[3] Chin-Yew Lin,et al. ROUGE: A Package for Automatic Evaluation of Summaries , 2004, ACL 2004.
[4] Christopher D. Manning,et al. Effective Approaches to Attention-based Neural Machine Translation , 2015, EMNLP.
[5] Zhe Gan,et al. Improving Sequence-to-Sequence Learning via Optimal Transport , 2019, ICLR.
[6] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[7] Christopher Ré,et al. Low-Memory Neural Network Training: A Technical Report , 2019, ArXiv.
[8] Tri Dao,et al. On the Downstream Performance of Compressed Word Embeddings , 2019, NeurIPS.
[9] Pritish Narayanan,et al. Deep Learning with Limited Numerical Precision , 2015, ICML.
[10] Xin Wang,et al. Parameter Efficient Training of Deep Convolutional Neural Networks by Dynamic Sparse Reparameterization , 2019, ICML.
[11] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[12] Hideki Nakayama,et al. Compressing Word Embeddings via Deep Compositional Code Learning , 2017, ICLR.
[13] Eric P. Xing,et al. Texar: A Modularized, Versatile, and Extensible Toolkit for Text Generation , 2018, ACL.
[14] Hao Wu,et al. Mixed Precision Training , 2017, ICLR.
[15] Mikhail Khodak,et al. A Compressed Sensing View of Unsupervised Text Embeddings, Bag-of-n-Grams, and LSTMs , 2018, ICLR.
[16] Ameya Velingker,et al. Random Fourier Features for Kernel Ridge Regression: Approximation Bounds and Statistical Guarantees , 2018, ICML.
[17] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[18] Masaaki Nagata,et al. Learning Compact Neural Word Embeddings by Parameter Space Sharing , 2016, IJCAI.
[19] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[20] Marc'Aurelio Ranzato,et al. Sequence Level Training with Recurrent Neural Networks , 2015, ICLR.
[21] Tri Dao,et al. Low-Precision Random Fourier Features for Memory-Constrained Kernel Approximation , 2018, AISTATS.
[22] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[23] Ryan LaRose,et al. Overview and Comparison of Gate Level Quantum Software Platforms , 2018, Quantum.
[24] Erich Elsen,et al. The State of Sparsity in Deep Neural Networks , 2019, ArXiv.
[25] Ilya Sutskever,et al. Generating Long Sequences with Sparse Transformers , 2019, ArXiv.
[26] Martin Andrews. Compressing Word Embeddings , 2016, ICONIP.
[27] Kunle Olukotun,et al. High-Accuracy Low-Precision Training , 2018, ArXiv.
[28] Andrew Gordon Wilson,et al. GPyTorch: Blackbox Matrix-Matrix Gaussian Process Inference with GPU Acceleration , 2018, NeurIPS.
[29] Jason Weston,et al. Reading Wikipedia to Answer Open-Domain Questions , 2017, ACL.
[30] Tianqi Chen,et al. Training Deep Nets with Sublinear Memory Cost , 2016, ArXiv.
[31] J. Landsberg,et al. Ranks of tensors and a generalization of secant varieties , 2009, 0909.4262.