Language Modeling in the Era of Abundant Data
暂无分享,去创建一个
[1] Slava M. Katz,et al. Estimation of probabilities from sparse data for the language model component of a speech recognizer , 1987, IEEE Trans. Acoust. Speech Signal Process..
[2] Hermann Ney,et al. Improved backing-off for M-gram language modeling , 1995, 1995 International Conference on Acoustics, Speech, and Signal Processing.
[3] Thorsten Brants,et al. One billion word benchmark for measuring progress in statistical language modeling , 2013, INTERSPEECH.
[4] Geoffrey E. Hinton,et al. Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer , 2017, ICLR.
[5] Frederick Jelinek,et al. Statistical methods for speech recognition , 1997 .
[6] Robert L. Mercer,et al. An Estimate of an Upper Bound for the Entropy of English , 1992, CL.
[7] Joris Pelemans,et al. Sparse Non-negative Matrix Language Modeling , 2016, Transactions of the Association for Computational Linguistics.
[8] Thorsten Brants,et al. Study on interaction between entropy pruning and kneser-ney smoothing , 2010, INTERSPEECH.
[9] Andreas Stolcke,et al. Entropy-based Pruning of Backoff Language Models , 2000, ArXiv.
[10] Joshua Goodman,et al. A bit of progress in language modeling , 2001, Comput. Speech Lang..
[11] Thomas M. Cover,et al. A convergent gambling estimate of the entropy of English , 1978, IEEE Trans. Inf. Theory.