Advertisement

Eliminating Over-Confidence in Software Development Effort Estimates

  • Magne Jørgensen
  • Kjetil Moløkken
Part of the Lecture Notes in Computer Science book series (LNCS, volume 3009)

Abstract

Previous studies show that software development projects strongly underestimate the uncertainty of their effort estimates. This overconfidence in estimation accuracy may lead to poor project planning and execution. In this paper, we investigate whether the use of estimation error information from previous projects improves the realism of uncertainty assessments. As far as we know, there have been no empirical software studies on this topic before. Nineteen realistically composed estimation teams provided minimum-maximum effort intervals for the same software project. Ten of the teams (Group A) received no instructions about how to complete the uncertainty assessment process. The remaining nine teams (Group B) were instructed to apply a history-based uncertainty assessment process. The main results is that software professionals seem to willing to consider the error of previous effort estimates as relevant information when assessing the minimum effort of a new project, but not so much when assessing the maximum effort!

Keywords

Actual Effort Maximum Effort Minimum Effort Effort Estimate Similar Project 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. 1.
    Connolly, T., Dean, D.: Decomposed versus holistic estimates of effort required for software writing tasks. Management Science 43(7), 1029–1045 (1997)zbMATHCrossRefGoogle Scholar
  2. 2.
    Jørgensen, M., Teigen, K.H.: Uncertainty Intervals versus Interval Uncertainty: An Alternative Method for Eliciting Effort Prediction Intervals in Software Development Projects. In: International conference on Project Management (ProMAC), Singapore, pp. 343–352 (2002)Google Scholar
  3. 3.
    Jørgensen, M., Teigen, K.H., Moløkken, K.: Better Sure than Safe? Overconfidence in Judgment Based Software Development Effort Prediction Intervals. Journal of System and Software (2004) (to appear)Google Scholar
  4. 4.
    Jørgensen, M.: Top-Down and Bottom-Up Expert Estimation of Software Development Effort 46(1), 3–16 (2004)Google Scholar
  5. 5.
    Alpert, M., Raiffa, H.: A progress report on the training of probability assessors. In: Tversky, A. (ed.) Judgment under uncertainty: Heuristics and biases, pp. 294–305. Cambridge University Press, Cambridge (1982)Google Scholar
  6. 6.
    Kahnemann, D., Slovic, P., Tversky, A.: Judgement under uncertainty: Heuristics and biases. Cambridge University Press, Cambridge (1982)Google Scholar
  7. 7.
    Tversky, A., Kahneman, D.: Judgment under uncertainty: Heuristics and biases. Science 185, 1124–1130 (1974)CrossRefGoogle Scholar
  8. 8.
    Yaniv, I., Foster, D.P.: Precision and accuracy of judgmental estimation. Journal of behavioral decision making 10, 21–32 (1997)CrossRefGoogle Scholar
  9. 9.
    Lichtenstein, S., Fischhoff, B.: Do those who know more also know more about how much they know? Organizational Behaviour and Human Decision Processes 20(2), 159–183 (1977)Google Scholar
  10. 10.
    Arkes, H.R.: Overconfidence in judgmental forecasting. In: Armstrong, J.S. (ed.) Principles of forecasting: A handbook for researchers and practitioners, pp. 495–515. Kluwer Academic Publishers, Boston (2001)Google Scholar
  11. 11.
    Kahneman, D., Tversky, A.: Variants of uncertainty. In: Kahneman, D., Slovic, P., Tversky, A. (eds.) Judgment under uncertainty: Heuristics and biases, pp. 509–520. Cambridge University Press, Cambridge (1982)Google Scholar
  12. 12.
    Griffin, D., Buehler, R.: Frequency, probability, and prediction: Easy solutions to cognitive illusions? Cognitive Psychology 38(1), 48–78 (1999)CrossRefGoogle Scholar
  13. 13.
    Kahneman, D., Lovallo, D.: Timid choices and bold forecasts: A cognitive perspective on risk taking. Management Science 39(1), 17–31 (1993)CrossRefGoogle Scholar
  14. 14.
    Jørgensen, M., Sjøberg, D.I.K.: An effort prediction interval approach based on the empirical distribution of previous estimation accuracy. Journal of Information and Software Technology 45(3), 123–136 (2003)CrossRefGoogle Scholar
  15. 15.
    Jørgensen, M., Sjøberg, D.I.K.: Impact of effort estimates on software project work. Information and Software Technology 43(15), 939–948 (2001), 2001Google Scholar
  16. 16.
    Klein, W.M., Kunda, Z.: Exaggerated self-assessments and the preference for controllable risks. Organizational behavior and human decision processes 59(3), 410–427 (1994)CrossRefGoogle Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2004

Authors and Affiliations

  • Magne Jørgensen
    • 1
  • Kjetil Moløkken
    • 1
    • 2
  1. 1.Simula Research LaboratoryLysakerNorway
  2. 2.University of OsloNorway

Personalised recommendations