Some notes on neural learning algorithm benchmarking
暂无分享,去创建一个
Abstract New neural learning algorithms are often benchmarked only poorly. This article gathers some important DOs and DON'Ts for researchers in order to improve on that situation. The essential requirements are (1) Volume: benchmarking has to be broad enough, i.e. must use several problems; (2) Validity: common errors that invalidate the results have to be avoided; (3) Reproducibility: benchmarking has to be documented well enough to be completely reproducible; and (4) Comparability: benchmark results should, if possible, be directly comparable with the results achieved by others using different algorithms.
[1] David W. Aha,et al. Generalizing from Case studies: A Case Study , 1992, ML.
[2] Ray J. Hickey. Artificial Universes - Towards a Systematic Approach to Evaluation Algorithms which Learn form Examples , 1992, ML.
[3] Lutz Prechelt. A quantitative study of neural network learning algorithm evaluation practices , 1995 .
[4] Lutz Prechelt,et al. PROBEN 1 - a set of benchmarks and benchmarking rules for neural network training algorithms , 1994 .