A practical guide for using statistical tests to assess randomized algorithms in software engineering

Andrea Arcuri, Lionel Briand

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Randomized algorithms have been used to successfully address many different types of software engineering problems. This type of algorithms employ a degree of randomness as part of their logic. Randomized algorithms are useful for difficult problems where a precise solution cannot be derived in a deterministic way within reasonable time. However, randomized algorithms produce different results on every run when applied to the same problem instance. It is hence important to assess the effectiveness of randomized algorithms by collecting data from a large enough number of runs. The use of rigorous statistical tests is then essential to provide support to the conclusions derived by analyzing such data. In this paper, we provide a systematic review of the use of randomized algorithms in selected software engineering venues in 2009. Its goal is not to perform a complete survey but to get a representative snapshot of current practice in software engineering research. We show that randomized algorithms are used in a significant percentage of papers but that, in most cases, randomness is not properly accounted for. This casts doubts on the validity of most empirical results assessing randomized algorithms. There are numerous statistical tests, based on different assumptions, and it is not always clear when and how to use these tests. We hence provide practical guidelines to support empirical research on randomized algorithms in software engineering

Original languageEnglish
Title of host publicationICSE 2011 - 33rd International Conference on Software Engineering, Proceedings of the Conference
Pages1-10
Number of pages10
DOIs
Publication statusPublished - 2011
Externally publishedYes
Event33rd International Conference on Software Engineering, ICSE 2011 - Waikiki, Honolulu, HI, United States
Duration: 21 May 201128 May 2011

Publication series

NameProceedings - International Conference on Software Engineering
ISSN (Print)0270-5257

Conference

Conference33rd International Conference on Software Engineering, ICSE 2011
Country/TerritoryUnited States
CityWaikiki, Honolulu, HI
Period21/05/1128/05/11

Keywords

  • bonferroni adjustment
  • confidence interval
  • effect size
  • non-parametric test
  • parametric test
  • statistical difference
  • survey
  • systematic review

Fingerprint

Dive into the research topics of 'A practical guide for using statistical tests to assess randomized algorithms in software engineering'. Together they form a unique fingerprint.

Cite this