Abstract
This paper describes models whose purpose is to explain the accuracy and bias variation of an organization’s estimates of software development effort through regression analysis. We collected information about variables that we believed would affect the accuracy or bias of estimates of the performance of tasks completed by the organization. In total, information about 49 software development tasks was collected. We found that the following conditions led to inaccuracies in estimates: (1) Estimates were provided by a person in the role of “software developer” instead of “project leader”, (2) The project had as its highest priority time-to-delivery instead of quality or cost, and (3) The estimator did not participate in the completion of the task. The following conditions led to an increased bias towards under-estimation: (1) Estimates were provided by a person with the role of “project leader” instead of “software developer”. (2) The estimator assessed the accuracy of own estimates of similar, previously completed tasks to be low (more than 20% error). Although all variables included in the models were significant p < 0.1), the explanatory and predictive power of both models was poor, that is, most of the variance in the accuracy and bias of estimates was not explained or predicted by our models. In addition, there were several important threats to the validity of the coefficients suggested by the models. An analysis of the estimators’ own descriptions of the reasons for achieved estimation accuracy on each task suggests that it will be difficult to include all important estimation accuracy and bias factors in regression-based models. It is, for this reason, not realistic to expect such models to replace human judgment in estimation uncertainty assessments and as input to plans for the improvement of estimates. It is, nevertheless, possible that the type of formal analysis and regression-based models presented in this paper may, in some cases, be useful as support for human judgment.
Similar content being viewed by others
References
Angelis, L., and Stamelos, I. 2000. A simulation tool for efficient analogy based cost estimation. Empirical Software Engineering 5: 35–68.
Briand, L. C., El Emam, K., Surmann, D., Wieczorek, I., and Maxwell, K. D. 1999. An assessment and comparison of common software cost estimation modeling techniques. International Conference on Software Engineering, Los Angeles, USA, ACM, New York, pp. 313–323.
Connolly, T., and Dean, D. 1997. Decomposed versus holistic estimates of effort required for software writing tasks. Management Science 43(7): 1029–1045.
Gray, A., MacDonnell, S., and Shepperd, M. 1999. Factors systematically associated with errors in subjective estimates of software development effort: The stability of expert judgment. Sixth International Software Metrics Symposium, IEEE Comput. Soc. Los Alamitos, CA, USA, pp. 216–227.
Jørgensen, M. 1995. Experience with the accuracy of software maintenance task effort prediction models. IEEE Transactions on Software Engineering 21: 674–681.
Jørgensen, M. 2004. A review of studies on expert estimation of software development effort. Journal of Systems and Software 70(1-2): 37–60.
Jørgensen, M., Løvstad, N., and Moen, L. 2002. Combining quantitative software development cost estimation precision data with qualitative data from project experience reports at Ericsson Design Center in Norway. Empirical Assessments of Software Engineering (EASE), Keele, UK.
Jørgensen, M., and Sjøberg, D. I. K. 2003. An effort prediction interval approach based on the empirical distribution of previous estimation accuracy. Journal of Information Software and Technology 45: 123–136.
Jørgensen, M., and Teigen, K. H. 2002. Uncertainty intervals versus interval uncertainty: An alternative method for eliciting effort prediction intervals in software development projects. International Conference on Project Management (ProMAC), Singapore, pp. 343–352.
Jørgensen, M., Teigen, K. H., and Moløkken-østvold, K. J, 2004. Better sure than safe? Overconfidence in judgment based software development effort prediction intervals. Journal of Systems and Software 70(1-2): 79–93.
Miyazaki, Y., Terakado, K., Ozaki, K., and Nozaki, H. 1994. Robust regression for developing software estimation models. Journal of Systems and Software 27: 3–16.
Rights and permissions
About this article
Cite this article
Jørgensen, M. Regression Models of Software Development Effort Estimation Accuracy and Bias. Empirical Software Engineering 9, 297–314 (2004). https://doi.org/10.1023/B:EMSE.0000039881.57613.cb
Issue Date:
DOI: https://doi.org/10.1023/B:EMSE.0000039881.57613.cb