1. Is it possible to disregard obsolete requirements? a family of experiments in software effort estimation
- Author
-
Richard Berntsson Svensson and Lucas Gren
- Subjects
Programvaruteknik ,Computer science ,Errors ,Bayesian probability ,Small Sample Size ,02 engineering and technology ,01 natural sciences ,95% credible intervals ,Bayesian data analysis ,Field (computer science) ,Research subjects ,010104 statistics & probability ,Software ,Order (exchange) ,0202 electrical engineering, electronic engineering, information engineering ,Credible interval ,Family of experiments ,0101 mathematics ,Students ,Bayesian approaches ,Estimation ,Software effort estimation ,business.industry ,Software Engineering ,Estimator ,020207 software engineering ,Systematic error ,Industrial engineering ,Cognitive bias ,Expert judgement ,Bayesian networks ,Software estimation ,business ,Information Systems - Abstract
Expert judgement is a common method for software effort estimations in practice today. Estimators are often shown extra obsolete requirements together with the real ones to be implemented. Only one previous study has been conducted on if such practices bias the estimations. We conducted six experiments with both students and practitioners to study, and quantify, the effects of obsolete requirements on software estimation. By conducting a family of six experiments using both students and practitioners as research subjects ($$N=461$$ N = 461 ), and by using a Bayesian Data Analysis approach, we investigated different aspects of this effect. We also argue for, and show an example of, how we by using a Bayesian approach can be more confident in our results and enable further studies with small sample sizes. We found that the presence of obsolete requirements triggered an overestimation in effort across all experiments. The effect, however, was smaller in a field setting compared to using students as subjects. Still, the over-estimations triggered by the obsolete requirements were systematically around twice the percentage of the included obsolete ones, but with a large 95% credible interval. The results have implications for both research and practice in that the found systematic error should be accounted for in both studies on software estimation and, maybe more importantly, in estimation practices to avoid over-estimations due to this systematic error. We partly explain this error to be stemming from the cognitive bias of anchoring-and-adjustment, i.e. the obsolete requirements anchored a much larger software. However, further studies are needed in order to accurately predict this effect.
- Published
- 2021