Please use this identifier to cite or link to this item: http://bura.brunel.ac.uk/handle/2438/16073
Full metadata record
DC FieldValueLanguage
dc.contributor.authorMahmood, Z-
dc.contributor.authorBowes, D-
dc.contributor.authorHall, T-
dc.contributor.authorLane, PCR-
dc.contributor.authorPetrić, J-
dc.date.accessioned2018-04-06T11:04:08Z-
dc.date.available2018-01-01-
dc.date.available2018-04-06T11:04:08Z-
dc.date.issued2018-
dc.identifier.citationInformation and Software Technology, 2018en_US
dc.identifier.issn0950-5849-
dc.identifier.urihttp://bura.brunel.ac.uk/handle/2438/16073-
dc.description.abstract© 2018 Elsevier B.V. Context: Replications are an important part of scientific disciplines. Replications test the credibility of original studies and can separate true results from those that are unreliable.Objective: In this paper we investigate the replication of defect prediction studies and identify the characteristics of replicated studies. We further assess how defect prediction replications are performed and the consistency of replication findings.Method: Our analysis is based on tracking the replication of 208 defect prediction studies identified by a highly cited Systematic Literature Review (SLR) [1]. We identify how often each of these 208 studies has been replicated and determine the type of replication carried out. We identify quality, citation counts, publication venue, impact factor, and data availability from all 208 SLR defect prediction papers to see if any of these factors are associated with the frequency with which they are replicated.Results: Only 13 (6%) of the 208 studies are replicated. Replication seems related to original papers appearing in the Transactions of Software Engineering (TSE) journal. The number of citations an original paper had was also an indicator of replications. In addition, studies conducted using closed source data seems to have more replications than those based on open source data. Where a paper has been replicated, 11 (38%) out of 29 studies revealed different results to the original study.Conclusion: Very few defect prediction studies are replicated. The lack of replication means that it remains unclear how reliable defect prediction is. We provide practical steps for improving the state of replication.en_US
dc.language.isoenen_US
dc.titleReproducibility and replicability of software defect prediction studiesen_US
dc.typeArticleen_US
dc.identifier.doihttp://dx.doi.org/10.1016/j.infsof.2018.02.003-
dc.relation.isPartOfInformation and Software Technology-
pubs.publication-statusAccepted-
Appears in Collections:Dept of Computer Science Research Papers

Files in This Item:
File Description SizeFormat 
Fulltext.pdf416.51 kBAdobe PDFView/Open


Items in BURA are protected by copyright, with all rights reserved, unless otherwise indicated.