Please use this identifier to cite or link to this item: http://bura.brunel.ac.uk/handle/2438/13448
Full metadata record
DC FieldValueLanguage
dc.contributor.authorFucci, D-
dc.contributor.authorScanniello, G-
dc.contributor.authorRomano, S-
dc.contributor.authorShepperd, M-
dc.contributor.authorSigweni, B-
dc.contributor.authorUyaguari, F-
dc.contributor.authorTurhan, B-
dc.contributor.authorJuristo, N-
dc.contributor.authorOivo, M-
dc.date.accessioned2016-11-04T11:07:32Z-
dc.date.available2016-09-08-
dc.date.available2016-11-04T11:07:32Z-
dc.date.issued2016-
dc.identifier.citationInternational Symposium on Empirical Software Engineering and Measurement (ESEM), 2016, 08-09-September-2016en_US
dc.identifier.isbn9781450344272-
dc.identifier.issn1949-3770-
dc.identifier.urihttp://bura.brunel.ac.uk/handle/2438/13448-
dc.description.abstractContext: Test-driven development (TDD) is an agile practice claimed to improve the quality of a software product, as well as the productivity of its developers. A previous study (i.e., baseline experiment) at the University of Oulu (Finland) compared TDD to a test-last development (TLD) approach through a randomized controlled trial. The results failed to support the claims. Goal: We want to validate the original study results by replicating it at the University of Basilicata (Italy), using a different design. Method: We replicated the baseline experiment, using a crossover design, with 21 graduate students. We kept the settings and context as close as possible to the baseline experiment. In order to limit researchers bias, we involved two other sites (UPM, Spain, and Brunel, UK) to conduct blind analysis of the data. Results: The Kruskal-Wallis tests did not show any significant difference between TDD and TLD in terms of testing effort (p-value = .27), external code quality (p-value = .82), and developers' productivity (p-value = .83). Nevertheless, our data revealed a difference based on the order in which TDD and TLD were applied, though no carry over effect. Conclusions: We verify the baseline study results, yet our results raises concerns regarding the selection of experimental objects, particularly with respect to their interaction with the order in which of treatments are applied. We recommend future studies to survey the tasks used in experiments evaluating TDD. Finally, to lower the cost of replication studies and reduce researchers' bias, we encourage other research groups to adopt similar multi-site blind analysis approach described in this paper.en_US
dc.description.sponsorshipThis research is supported in part by the Academy of Finland Project 278354.en_US
dc.language.isoenen_US
dc.publisherACMen_US
dc.subjectTest-driven developmenten_US
dc.subjectExternal experiment replicationen_US
dc.subjectblind analysisen_US
dc.titleAn external replication on the effects of test-driven development using a multi-site blind analysis approachen_US
dc.typeConference Paperen_US
dc.identifier.doihttp://dx.doi.org/10.1145/2961111.2962592-
dc.relation.isPartOfInternational Symposium on Empirical Software Engineering and Measurement-
pubs.publication-statusPublished-
pubs.volume08-09-September-2016-
Appears in Collections:Dept of Computer Science Research Papers

Files in This Item:
File Description SizeFormat 
FullText.pdf432.5 kBAdobe PDFView/Open


Items in BURA are protected by copyright, with all rights reserved, unless otherwise indicated.