Stacked Sequential Learning
暂无分享,去创建一个
[1] R. H. Myers,et al. Probability and Statistics for Engineers and Scientists , 1978 .
[2] Raymond H. Myers,et al. Probability and Statistics for Engineers and Scientists. , 1973 .
[3] David H. Wolpert,et al. Stacked generalization , 1992, Neural Networks.
[4] D. Signorini,et al. Neural networks , 1995, The Lancet.
[5] Thomas G. Dietterich. What is machine learning? , 2020, Archives of Disease in Childhood.
[6] Adam L. Berger,et al. A Maximum Entropy Approach to Natural Language Processing , 1996, CL.
[7] Yoav Freund,et al. Large Margin Classification Using the Perceptron Algorithm , 1998, COLT.
[8] George Tzanetakis,et al. MARSYAS: a framework for audio analysis , 1999, Organised Sound.
[9] Andrew McCallum,et al. Maximum Entropy Markov Models for Information Extraction and Segmentation , 2000, ICML.
[10] Andrew McCallum,et al. Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data , 2001, ICML.
[11] Dan Klein,et al. Conditional Structure versus Conditional Estimation in NLP Models , 2002, EMNLP.
[12] Thomas G. Dietterich. Machine Learning for Sequential Data: A Review , 2002, SSPR/SPR.
[13] Michael Collins,et al. Discriminative Training Methods for Hidden Markov Models: Theory and Experiments with Perceptron Algorithms , 2002, EMNLP.
[14] Dan Klein,et al. Feature-Rich Part-of-Speech Tagging with a Cyclic Dependency Network , 2003, NAACL.
[15] Thomas G. Dietterich,et al. Training conditional random fields via gradient tree boosting , 2004, ICML.
[16] William W. Cohen,et al. Learning to Extract Signature and Reply Lines from Email , 2004, CEAS.
[17] Adwait Ratnaparkhi,et al. Learning to Parse Natural Language with Maximum Entropy Models , 1999, Machine Learning.