Assessing prediction systems

For some years software engineers have been attempting to develop useful prediction systems to estimate such attributes as the effort to develop a piece of software and the likely number of defects. Typically, prediction systems are proposed and then subjected to empirical evaluation. Claims are then made with regard to the quality of the prediction systems. A wide variety of prediction quality indicators have been suggested in the literature. Unfortunately, we believe that a somewhat confusing state of affairs prevails and that this impedes research progress. This paper aims to provide the research community with a better understanding of the meaning of, and relationship between, these indicators. We critically review twelve different approaches by considering them as descriptors of the residual variable. We demonstrate that the two most popular indicators MMRE and pred(25) are in fact indicators of the spread and shape respectively of prediction accuracy where prediction accuracy is the ratio of estimate to 1 Version 11 28/5/99

[1]  Peter J. Rousseeuw,et al.  Robust regression and outlier detection , 1987 .

[2]  Barbara Kitchenham,et al.  The MERMAID Approach to software cost estimation , 1990 .

[3]  Meir M. Lehman,et al.  Characteristics of Large Systems , 1978 .

[4]  Stephen G. MacDonell Metrics for database systems: an empirical study , 1997, Proceedings Fourth International Software Metrics Symposium.

[5]  H. E. Dunsmore,et al.  Software engineering metrics and models , 1986 .

[6]  N. Nakagawa,et al.  Method to estimate parameter values in software prediction models , 1991 .

[7]  Magne Jørgensen,et al.  Experience With the Accuracy of Software Maintenance Task Effort Prediction Models , 1995, IEEE Trans. Software Eng..

[8]  Ingunn Myrtveit,et al.  Human performance estimating with analogy and regression models: an empirical validation , 1998, Proceedings Fifth International Software Metrics Symposium. Metrics (Cat. No.98TB100262).

[9]  Y. Miyazaki,et al.  Robust regression for developing software estimation models , 1994, J. Syst. Softw..

[10]  Barry W. Boehm,et al.  Calibrating the COCOMO II Post-Architecture model , 1998, Proceedings of the 20th International Conference on Software Engineering.

[11]  Chris F. Kemerer,et al.  An empirical validation of software cost estimation models , 1987, CACM.

[12]  Barbara A. Kitchenham,et al.  An investigation of analysis techniques for software datasets , 1999, Proceedings Sixth International Software Metrics Symposium (Cat. No.PR00403).

[13]  Martin J. Shepperd,et al.  Estimating Software Project Effort Using Analogies , 1997, IEEE Trans. Software Eng..

[14]  Barbara A. Kitchenham,et al.  Empirical studies of assumptions that underlie software cost-estimation models , 1992, Inf. Softw. Technol..

[15]  Xiangzhu Gao,et al.  Assessing Software Cost Estimation Models: criteria for accuracy, consistency and regression , 1997, Australas. J. Inf. Syst..

[16]  Barbara A. Kitchenham,et al.  Estimates, Uncertainty, and Risk , 1997, IEEE Softw..