Noise sensitivity signatures for model selection
暂无分享,去创建一个
Presents a method for calculating the "noise sensitivity signature" of a learning algorithm which is based on scrambling the output classes of various fractions of the training data. This signature can be used to indicate a good (or bad) match between the complexity of the classifier and the complexity of the data and hence to improve the predictive accuracy of a classification algorithm. Use of noise sensitivity signatures is distinctly different from other schemes to avoid overtraining, such as cross-validation, which uses only part of the training data, or various penalty functions, which are not data-adaptive. Noise sensitivity signature methods use all of the training data and are manifestly data-adaptive and nonparametric. They are well suited for situations with limited training data.
[1] James Theiler,et al. Testing for nonlinearity in time series: the method of surrogate data , 1992 .
[2] B. LeBaron,et al. Nonlinear Dynamics and Stock Returns , 2021, Cycles and Chaos in Economic Equilibrium.
[3] A. Lapedes,et al. Covariation of mutations in the V3 loop of human immunodeficiency virus type 1 envelope protein: an information theoretic analysis. , 1993, Proceedings of the National Academy of Sciences of the United States of America.
[4] J. Rissanen. Stochastic Complexity in Statistical Inquiry Theory , 1989 .