Replication of Studies in Empirical Software Engineering: A Systematic Mapping Study, From 2013 to 2018

<italic>Context</italic>: In any discipline, replications of empirical studies are necessary to consolidate the acquired knowledge. In Software Engineering, replications have been reported since the 1990s, although their number is still small. The difficulty in publishing, the lack of guidelines, and the unavailability of replication packages are pointed out by the community as some of the main causes. <italic>Objective</italic>: Understanding the current state of replications in Software Engineering studies by evaluating current trends and evolution during the last 6 years. <italic>Method</italic>: A Systematic Mapping Study including articles published in the 2013–2018 period that report at least one replication of an empirical study in Software Engineering. <italic>Results</italic>: 137 studies were selected and analysed, identifying: <inline-formula> <tex-math notation="LaTeX">${i}$ </tex-math></inline-formula>) forums; ii) authors, co-authorships and institutions; iii) most cited studies; iv) research topics addressed; <inline-formula> <tex-math notation="LaTeX">${v}$ </tex-math></inline-formula>) empirical methods used; vi) temporal distribution of publications; and vii) distribution of studies according to research topics and empirical methods. <italic>Conclusions</italic>: According to our results, the most relevant forums are the <italic>Empirical Software Engineering</italic> and <italic>Information and Software Technology</italic> journals, and the <italic>Empirical Software Engineering and Measurement</italic> conference. We observed that, as in previous reviews by other researchers, most of the studies were carried out by European institutions, especially Italian, Spanish, and German researchers and institutions. The studies attracting more citations were published mainly in journals and in the <italic>International Conference on Software Engineering</italic>. Testing, requirements, and software construction were the most frequent topics of replication studies, whereas the usual empirical method was the controlled experiment. On the other hand, we identified research gaps in areas such as software engineering process, software configuration management, and software engineering economics. When analysed together with previous reviews, there is a clear increasing trend in the number of published replications in the 2013–2018 period.

[1]  Yuanfang Cai,et al.  A replication case study to measure the architectural quality of a commercial system , 2014, ESEM '14.

[2]  Birgit Penzenstadler,et al.  Two Replication Studies for Evaluating Artefact Models in RE: Results and Lessons Learnt , 2013, 2013 3rd International Workshop on Replication in Empirical Software Engineering Research.

[3]  Roel Wieringa,et al.  Understandability of Goal Concepts by Requirements Engineering Experts , 2014, ER Workshops.

[4]  Marco Torchiano,et al.  A family of experiments to assess the effectiveness and efficiency of source code obfuscation techniques , 2013, Empirical Software Engineering.

[5]  Richard Torkar,et al.  Outliers and Replication in Software Engineering , 2014, 2014 21st Asia-Pacific Software Engineering Conference.

[6]  Yutaka Matsuo,et al.  Replication issues in syntax-based aspect extraction for opinion mining , 2017, EACL.

[7]  Tayana Conte,et al.  Towards Validating Complexity-Based Metrics for Software Product Line Architectures , 2013, 2013 VII Brazilian Symposium on Software Components, Architectures and Reuse.

[8]  Ville Leppänen,et al.  Investigating the Effects of Agile Practices and Processes on Technical Debt - The Viewpoint of the Brazilian Software Industry , 2018, SEKE.

[9]  Ali Idri,et al.  Software cost estimation by classical and Fuzzy Analogy for Web Hypermedia Applications: A replicated study , 2013, 2013 IEEE Symposium on Computational Intelligence and Data Mining (CIDM).

[10]  Jens Grabowski,et al.  Global vs. local models for cross-project defect prediction , 2017, Empirical Software Engineering.

[11]  Christian Quesada-López,et al.  An Empirical Validation of Function Point Structure and Applicability: A Replication Study , 2015, CIbSE.

[12]  Ronnie E. S. Santos,et al.  Replication of Empirical Studies in Software Engineering: An Update of a Systematic Mapping Study , 2015, 2015 ACM/IEEE International Symposium on Empirical Software Engineering and Measurement (ESEM).

[13]  Michael Felderer,et al.  Manual test case derivation from UML activity diagrams and state machines: A controlled experiment , 2015, Inf. Softw. Technol..

[14]  Shane McIntosh,et al.  Revisiting the Impact of Classification Techniques on the Performance of Defect Prediction Models , 2015, 2015 IEEE/ACM 37th IEEE International Conference on Software Engineering.

[15]  Miroslaw Staron,et al.  On the effect of using SysML requirement diagrams to comprehend requirements: results from two controlled experiments , 2014, EASE '14.

[16]  Mariano Ceccato,et al.  Do Automatically Generated Test Cases Make Debugging Easier? An Experimental Assessment of Debugging Effectiveness and Efficiency , 2015, ACM Trans. Softw. Eng. Methodol..

[17]  Bonita Sharif,et al.  Investigating Eye Movements in Natural Language and C++ Source Code - A Replication Experiment , 2017, HCI.

[18]  Giuseppe Scanniello,et al.  Distributed modeling of use case diagrams with a method based on think-pair-square: Results from two controlled experiments , 2014, J. Vis. Lang. Comput..

[19]  Marco Torchiano,et al.  Do UML object diagrams affect design comprehensibility? Results from a family of four controlled experiments , 2017, J. Vis. Lang. Comput..

[20]  Tsvi Kuflik,et al.  Comparing the comprehensibility of requirements models expressed in Use Case and Tropos: Results from a family of experiments , 2013, Inf. Softw. Technol..

[21]  Peter C. Rigby,et al.  Organizational Volatility and Post-release Defects: A Replication Case Study Using Data from Google Chrome , 2015, 2015 IEEE/ACM 12th Working Conference on Mining Software Repositories.

[22]  Paolo Tonella,et al.  A large scale empirical comparison of state-of-the-art search-based test case generators , 2018, Inf. Softw. Technol..

[23]  Marco Torchiano,et al.  On the impact of state-based model-driven development on maintainability: a family of experiments using UniMod , 2018, Empirical Software Engineering.

[24]  Silvia Teresita Acuña,et al.  A replicated quasi-experimental study on the influence of personality and team climate in software development , 2013, Empirical Software Engineering.

[25]  Jeffrey C. Carver,et al.  Effects of cloned code on software maintainability: A replicated developer study , 2013, 2013 20th Working Conference on Reverse Engineering (WCRE).

[26]  José Antonio Pow-Sang,et al.  Evaluating and Comparing Perceptions Between Undergraduate Students and Practitioners in Controlled Experiments for Requirements Prioritization , 2016 .

[27]  James Walden,et al.  Using software birthmarks and clustering to identify similar classes and major functionalities , 2018, ACM Southeast Regional Conference.

[28]  Kai Petersen,et al.  Systematic Mapping Studies in Software Engineering , 2008, EASE.

[29]  Huai Liu,et al.  Evaluating and Comparing Fault-Based Testing Strategies for General Boolean Specifications: A Series of Experiments , 2015, Comput. J..

[30]  Wilhelm Hasselbring,et al.  Comparing Trace Visualizations for Program Comprehension through Controlled Experiments , 2015, 2015 IEEE 23rd International Conference on Program Comprehension.

[31]  Danilo Caivano,et al.  Are Forward Designed or Reverse-Engineered UML diagrams more helpful for code maintenance?: A family of experiments , 2015, Inf. Softw. Technol..

[32]  Omar S. Gómez,et al.  Efficiency of Software Testing Techniques: A Controlled Experiment Replication and Network Meta-analysis , 2017, e Informatica Softw. Eng. J..

[33]  Dietmar Pfahl,et al.  Using text clustering to predict defect resolution time: a conceptual replication and an evaluation of prediction accuracy , 2015, Empirical Software Engineering.

[34]  Magne Jørgensen,et al.  An experimental evaluation of a de-biasing intervention for professional software developers , 2018, SAC.

[35]  Linda Newman,et al.  Advancing viewpoint merging in requirements engineering: a theoretical replication and explanatory study , 2017, Requirements Engineering.

[36]  Andrew Brooks,et al.  A Partial Replication with a Sample Size of One: A Smoke Test for Empirical Software Engineering , 2013, 2013 3rd International Workshop on Replication in Empirical Software Engineering Research.

[37]  Yann-Gaël Guéhéneuc,et al.  An experimental investigation on the effects of context on source code identifiers splitting and expansion , 2013, Empirical Software Engineering.

[38]  Natalia Juristo Juzgado,et al.  Effect of Domain Knowledge on Elicitation Effectiveness: An Internally Replicated Controlled Experiment , 2016, IEEE Transactions on Software Engineering.

[39]  Gabriel Costa Silva,et al.  An empirical study for evaluating the performance of multi-cloud APIs , 2018, Future Gener. Comput. Syst..

[40]  Natalia Juristo Juzgado,et al.  The role of non-exact replications in software engineering experiments , 2011, Empirical Software Engineering.

[41]  Giuseppe Scanniello,et al.  Link analysis algorithms for static concept location: an empirical assessment , 2014, Empirical Software Engineering.

[42]  Óscar Dieste Tubío,et al.  Effectiveness for detecting faults within and outside the scope of testing techniques: an independent replication , 2013, Empirical Software Engineering.

[43]  Forrest Shull,et al.  Building Knowledge through Families of Experiments , 1999, IEEE Trans. Software Eng..

[44]  Ana I. Molina,et al.  Evaluating a graphical notation for modeling collaborative learning activities: A family of experiments , 2014, Sci. Comput. Program..

[45]  Kathleen M. Carley,et al.  Assessing team performance from a socio-technical congruence perspective , 2012, 2012 International Conference on Software and System Process (ICSSP).

[46]  He Zhang,et al.  A replicated experiment for evaluating the effectiveness of pairing practice in PSP education , 2018, J. Syst. Softw..

[47]  Danilo Caivano,et al.  Does the level of detail of UML diagrams affect the maintainability of source code?: a family of experiments , 2016, Empirical Software Engineering.

[48]  Natalia Juristo Juzgado,et al.  Analyzing Families of Experiments in SE: A Systematic Mapping Study , 2018, IEEE Transactions on Software Engineering.

[49]  Clauirton A. Siebra,et al.  The importance of replications in software engineering: a case study in defect prediction , 2015, RACS.

[50]  Sheikh Umar Farooq,et al.  A replicated empirical study to evaluate software testing methods , 2017, J. Softw. Evol. Process..

[51]  Nils Göde,et al.  Cloned code: stable code , 2013, J. Softw. Evol. Process..

[52]  Natalia Juristo Juzgado,et al.  An External Replication on the Effects of Test-driven Development Using a Multi-site Blind Analysis Approach , 2016, ESEM.

[53]  Massila Kamalrudin,et al.  Eliciting Security Requirments For Mobile Apps: A Replication Study , 2017 .

[54]  Burak Turhan,et al.  On the role of tests in test-driven development: a differentiated and partial replication , 2013, Empirical Software Engineering.

[55]  Erik Kamsties,et al.  Do Words Make a Difference? An Empirical Study on the Impact of Taxonomies on the Classification of Requirements , 2017, 2017 IEEE 25th International Requirements Engineering Conference (RE).

[56]  Simone R. S. Souza,et al.  The Use of Visual Text Mining to Support the Study Selection Activity in Systematic Literature Reviews: A Replication Study , 2013, 2013 3rd International Workshop on Replication in Empirical Software Engineering Research.

[57]  Natalia Juristo Juzgado,et al.  A process for managing interaction between experimenters to get useful similar replications , 2013, Inf. Softw. Technol..

[58]  Fabio Q. B. da Silva,et al.  Motivation in software engineering industrial practice: A cross-case analysis of two software organisations , 2014, Inf. Softw. Technol..

[59]  Peng Liang,et al.  Do architectural design decisions improve the understanding of software architecture? two controlled experiments , 2014, ICPC 2014.

[60]  Fabio Q. B. da Silva,et al.  Team building criteria in software projects: A mix-method replicated study , 2013, Inf. Softw. Technol..

[61]  Roel Wieringa,et al.  Naming the pain in requirements engineering , 2016, Empirical Software Engineering.

[62]  Antonio Ruiz Cortés,et al.  A controlled experiment to evaluate the effects of mindfulness in software engineering , 2014, ESEM '14.

[63]  Barbara A. Kitchenham,et al.  The role of replications in empirical software engineering—a word of warning , 2008, Empirical Software Engineering.

[64]  Antonio Ruiz-Cortés,et al.  Effects of Mindfulness on Conceptual Modeling Performance: A Series of Experiments , 2020, IEEE Transactions on Software Engineering.

[65]  Natalia Juristo Juzgado,et al.  Difficulties in Running Experiments in the Software Industry: Experiences from the Trenches , 2015, 2015 IEEE/ACM 3rd International Workshop on Conducting Empirical Studies in Industry.

[66]  Keqing He,et al.  A replicated experiment on architecture pattern recommendation based on quality requirements , 2014, 2014 IEEE 5th International Conference on Software Engineering and Service Science.

[67]  Sheikh Umar Farooq,et al.  An externally replicated experiment to evaluate software testing methods , 2013, EASE '13.

[68]  Stefan Wagner,et al.  Naming the pain in requirements engineering: design of a global family of surveys and first results from Germany , 2013, EASE '13.

[69]  Martin Pinzger,et al.  Co-evolution analysis of production and test code by learning association rules of changes , 2018, 2018 IEEE Workshop on Machine Learning Techniques for Software Quality Evaluation (MaLTeSQuE).

[70]  Linda Newman,et al.  Advancing Repeated Research in Requirements Engineering: A Theoretical Replication of Viewpoint Merging , 2016, 2016 IEEE 24th International Requirements Engineering Conference (RE).

[71]  Patrícia Duarte de Lima Machado,et al.  Test case prioritization techniques for model-based testing: a replicated study , 2017, Software Quality Journal.

[72]  Ana Regina Cavalcanti da Rocha,et al.  Towards Successful Software Process Improvement Initiatives: Experiences from the Battlefield , 2015, AMCIS.

[73]  Mika Mäntylä,et al.  Are test cases needed? Replicated comparison between exploratory and test-case-based software testing , 2014, Empirical Software Engineering.

[74]  Antonio Ruiz Cortés,et al.  An experimental replication on the effect of the practice of mindfulness in conceptual modeling performance , 2018, J. Syst. Softw..

[75]  Sérgio Soares,et al.  Mechanisms to Characterize Context of Empirical Studies in Software Engineering , 2015, CIbSE.

[76]  Jens Grabowski,et al.  A Comparative Study to Benchmark Cross-Project Defect Prediction Approaches , 2018, IEEE Transactions on Software Engineering.

[77]  Martin P. Robillard,et al.  Revisiting Turnover-Induced Knowledge Loss in Software Projects , 2017, 2017 IEEE International Conference on Software Maintenance and Evolution (ICSME).

[78]  M Wood,et al.  Replication of Experimental Results in Software Engineering , 2022 .

[79]  M. Roper,et al.  Replication of Software Engineering Experiments , 2000 .

[80]  Jeffrey C. Carver Towards Reporting Guidelines for Experimental Replications: A Proposal , 2010 .

[81]  Birgit Penzenstadler,et al.  Sustainability analysis and ease of learning in artifact-based requirements engineering: The newest member of the family of studies (It's a girl!) , 2017, Inf. Softw. Technol..

[82]  Wilkerson de L. Andrade,et al.  A family of test selection criteria for Timed Input-Output Symbolic Transition System models , 2016, Sci. Comput. Program..

[83]  Jeffrey C. Carver,et al.  A Multi-Site Joint Replication of a Design Patterns Experiment Using Moderator Variables to Generalize across Contexts , 2016, IEEE Transactions on Software Engineering.

[84]  Alice E. Smith,et al.  Toward a model-driven engineering framework for reproducible simulation experiment lifecycle management , 2014, Proceedings of the Winter Simulation Conference 2014.

[85]  Giuseppe Scanniello,et al.  Fixing Faults in C and Java Source Code , 2017, ACM Trans. Softw. Eng. Methodol..

[86]  Giuseppe Scanniello,et al.  On the impact of UML analysis models on source-code comprehensibility and modifiability , 2014, ACM Trans. Softw. Eng. Methodol..

[87]  Bora Caglayan,et al.  Predicting bug-fixing time: A replication study using an open source software project , 2018, J. Syst. Softw..

[88]  Christian Quesada-López,et al.  An Empirical Evaluation of Automated Function Points , 2016, CIbSE.

[89]  Kathryn T. Stolee,et al.  Exploring Crowd Consistency in a Mechanical Turk Survey , 2016, 2016 IEEE/ACM 3rd International Workshop on CrowdSourcing in Software Engineering (CSI-SE).

[90]  Rini van Solingen,et al.  A replicated study on correlating agile team velocity measured in function and story points , 2014, WETSoM 2014.

[91]  Ralph E. Johnson,et al.  Effective Software Merging in the Presence of Object-Oriented Refactorings , 2008, IEEE Transactions on Software Engineering.

[92]  Santiago Matalonga,et al.  A Controlled Experiment to Explore Potentially Undetectable Defects for Testing Techniques , 2014, SEKE.

[93]  Sven Apel,et al.  Comparing the influence of using feature-oriented programming and conditional compilation on comprehending feature-oriented software , 2018, Empirical Software Engineering.

[94]  Ellen Francine Barbosa,et al.  A Visual Approach to Validate the Selection Review of Primary Studies in Systematic Reviews: A Replication Study , 2013, SEKE.

[95]  Fabio Q. B. da Silva,et al.  Replication of empirical studies in software engineering research: a systematic mapping study , 2012, Empirical Software Engineering.

[96]  Silvia Mara Abrahão,et al.  Assessing the Effectiveness of Sequence Diagrams in the Comprehension of Functional Requirements: Results from a Family of Five Experiments , 2013, IEEE Transactions on Software Engineering.

[97]  Leticia Montalvillo-Mendizabal,et al.  Requirement-driven evolution in software product lines: A systematic mapping study , 2016, J. Syst. Softw..

[98]  Davide Taibi,et al.  How developers perceive smells in source code: A replicated study , 2017, Inf. Softw. Technol..

[99]  Tihana Galinac Grbac,et al.  Co-evolutionary multi-population genetic programming for classification in software defect prediction: An empirical case study , 2017, Appl. Soft Comput..

[100]  Dietmar Pfahl,et al.  Reporting guidelines for controlled experiments in software engineering , 2005, 2005 International Symposium on Empirical Software Engineering, 2005..

[101]  Amela Karahasanovic,et al.  A survey of controlled experiments in software engineering , 2005, IEEE Transactions on Software Engineering.

[102]  José Miguel Morales,et al.  A family of experiments to evaluate the understandability of TRiStar and i* for modeling teleo-reactive systems , 2016, J. Syst. Softw..

[103]  Stefan Sobernig,et al.  Attributed variability models: outside the comfort zone , 2017, ESEC/SIGSOFT FSE.

[104]  Silvia Mara Abrahão,et al.  Empirical validation of a usability inspection method for model-driven Web development , 2013, J. Syst. Softw..

[105]  Paulo Borba,et al.  Assessing Semistructured Merge in Version Control Systems: A Replicated Experiment , 2015, 2015 ACM/IEEE International Symposium on Empirical Software Engineering and Measurement (ESEM).

[106]  Manishankar Mondal,et al.  Is cloned code really stable? , 2018, Empirical Software Engineering.

[107]  Davide Taibi,et al.  Functional Size Measures and Effort Estimation in Agile Development: A Replicated Study , 2015, XP.

[108]  Tayana Conte,et al.  Towards Building Knowledge on Causes of Critical Requirements Engineering Problems , 2015, SEKE.

[109]  Marco Torchiano,et al.  Assessing the Effect of Screen Mockups on the Comprehension of Functional Requirements , 2014, TSEM.

[110]  José Amancio M. Santos,et al.  Exploring decision drivers on god class detection in three controlled experiments , 2015, SAC.

[111]  Giuseppe Scanniello,et al.  Studying the Effect of UML-Based Models on Source-Code Comprehensibility: Results from a Long-Term Investigation , 2015, PROFES.

[112]  Michael Felderer,et al.  Preventing Incomplete/Hidden Requirements: Reflections on Survey Data from Austria and Brazil , 2016, SWQD.

[113]  Jeffrey C. Carver,et al.  How do Practitioners Perceive the Relevance of Requirements Engineering Research? An Ongoing Study , 2017, 2017 IEEE 25th International Requirements Engineering Conference (RE).

[114]  Uwe Zdun,et al.  The Supportive Effect of Traceability Links in Architecture-Level Software Understanding: Two Controlled Experiments , 2014, 2014 IEEE/IFIP Conference on Software Architecture.

[115]  Antonio Ruiz Cortés,et al.  EXEMPLAR: An Experimental Information Repository for Software Engineering Research , 2014 .

[116]  Silvia Mara Abrahão,et al.  Validating a model-driven software architecture evaluation and improvement method: A family of experiments , 2015, Inf. Softw. Technol..

[117]  Pearl Brereton,et al.  Systematic literature reviews in software engineering - A systematic literature review , 2009, Inf. Softw. Technol..

[118]  Emilia Mendes,et al.  Investigating the effects of personality traits on pair programming in a higher education setting through a family of experiments , 2012, Empirical Software Engineering.

[119]  Burak Turhan,et al.  Lessons learned from a partial replication of an experiment in the context of a software engineering course , 2017 .

[120]  Ronnie E. S. Santos,et al.  Investigations about replication of empirical studies in software engineering: A systematic mapping study , 2015, Inf. Softw. Technol..

[121]  A. Brooks,et al.  Verification of results in software maintenance through external replication , 1994, Proceedings 1994 International Conference on Software Maintenance.

[122]  Lionel C. Briand,et al.  Does aspect-oriented modeling help improve the readability of UML state machines? , 2012, Software & Systems Modeling.

[123]  Jeffrey C. Carver,et al.  The role of replications in Empirical Software Engineering , 2008, Empirical Software Engineering.

[124]  Per Runeson,et al.  A quantitative analysis of the unit verification perspective on fault distributions in complex software systems: an operational replication , 2016, Software Quality Journal.

[125]  Natalia Juristo Juzgado,et al.  Understanding replication of experiments in software engineering: A classification , 2014, Inf. Softw. Technol..

[126]  Michele Lanza,et al.  Answering software evolution questions: An empirical evaluation , 2013, Inf. Softw. Technol..

[127]  Natalia Juristo Juzgado,et al.  Evidence of the presence of bias in subjective metrics: analysis within a family of experiments , 2014, EASE '14.

[128]  Michael D. Ernst,et al.  Evaluating and Improving Fault Localization , 2017, 2017 IEEE/ACM 39th International Conference on Software Engineering (ICSE).

[129]  Jeffrey C. Carver,et al.  Program comprehension of domain-specific and general-purpose languages: replication of a family of experiments using integrated development environments , 2017, Empirical Software Engineering.

[130]  Keith Quille,et al.  Programming: predicting student success early in CS1. a re-validation and replication study , 2018, ITiCSE.

[131]  Natalia Juristo Juzgado,et al.  Are team personality and climate related to satisfaction and software quality? Aggregating results from a twice replicated experiment , 2015, Inf. Softw. Technol..

[132]  Ellen Francine Barbosa,et al.  Visual Text Mining: Ensuring the Presence of Relevant Studies in Systematic Literature Reviews , 2015, Int. J. Softw. Eng. Knowl. Eng..

[133]  Jeffrey C. Carver,et al.  Investigation of individual factors impacting the effectiveness of requirements inspections: a replicated experiment , 2014, Empirical Software Engineering.

[134]  Ömer Faruk Arar,et al.  Deriving thresholds of software metrics to predict faults on open source software: Replicated case studies , 2016, Expert Syst. Appl..

[135]  Alessandro F. Garcia,et al.  On the Influence of Human Factors for Identifying Code Smells: A Multi-Trial Empirical Study , 2017, 2017 ACM/IEEE International Symposium on Empirical Software Engineering and Measurement (ESEM).

[136]  Fang Jun Wu Empirical Tests of Scale-Free Characteristic in Open Source Software: A Replicated Case Study , 2012 .

[137]  Giuseppe Scanniello,et al.  Source-code comprehension tasks supported by UML design models: Results from a controlled experiment and a differentiated replication , 2015, J. Vis. Lang. Comput..

[138]  Gabriele Bavota,et al.  A fine-grained analysis of the support provided by UML class diagrams and ER diagrams during data model maintenance , 2012, Software & Systems Modeling.

[139]  Manoel G. Mendonça,et al.  The most common causes and effects of technical debt: first results from a global family of industrial surveys , 2018, ESEM.

[140]  Giuseppe Scanniello,et al.  On the comprehension of workflows modeled with a precise style: results from a family of controlled experiments , 2013, Software & Systems Modeling.

[141]  Silvia Mara Abrahão,et al.  Evaluating Software Architecture Evaluation Methods: An Internal Replication , 2017, EASE.

[142]  Marco Torchiano,et al.  On the Benefits and Barriers When Adopting Software Modelling and Model Driven Techniques - An External, Differentiated Replication , 2015, 2015 ACM/IEEE International Symposium on Empirical Software Engineering and Measurement (ESEM).

[143]  Martin Hiller,et al.  Experimental Comparison of Two Safety Analysis Methods and Its Replication , 2013, 2013 ACM / IEEE International Symposium on Empirical Software Engineering and Measurement.

[144]  Giuseppe Scanniello,et al.  Documenting Design-Pattern Instances: A Family of Experiments on Source-Code Comprehensibility , 2015, TSEM.

[145]  Jacob Krüger,et al.  Commenting source code: is it worth it for small programming tasks? , 2018, Empirical Software Engineering.

[146]  Gordon Fraser,et al.  How Do Automatically Generated Unit Tests Influence Software Maintenance? , 2018, 2018 IEEE 11th International Conference on Software Testing, Verification and Validation (ICST).

[147]  Maria Teresa Baldassarre,et al.  An Analysis of Software Engineering Experiments Using Human Subjects , 2015, 2015 ACM/IEEE International Symposium on Empirical Software Engineering and Measurement (ESEM).

[148]  Fábio Levy Siqueira,et al.  Comparing the comprehensibility of requirements models: An experiment replication , 2017, Inf. Softw. Technol..

[149]  Johan Per Fredrik Almqvist,et al.  Replication of Controlled Experiments in Empirical Software Engineering - A Survey , 2006 .

[150]  Ayse Basar Bener,et al.  A Replication Study: Just-in-Time Defect Prediction with Ensemble Learning , 2018, 2018 IEEE/ACM 6th International Workshop on Realizing Artificial Intelligence Synergies in Software Engineering (RAISE).

[151]  Márcio Ribeiro,et al.  Feature maintenance with emergent interfaces , 2014, ICSE.

[152]  Per Runeson,et al.  Variation factors in the design and analysis of replicated controlled experiments , 2013, Empirical Software Engineering.

[153]  Annie I. Antón,et al.  Evaluating Legal Implementation Readiness Decision-Making , 2015, IEEE Transactions on Software Engineering.

[154]  Steve Counsell,et al.  The role and value of replication in empirical software engineering results , 2018, Inf. Softw. Technol..

[155]  Burak Turhan,et al.  A Replicated Experiment on the Effectiveness of Test-First Development , 2013, 2013 ACM / IEEE International Symposium on Empirical Software Engineering and Measurement.

[156]  Beatriz Marín,et al.  Evaluating Model-Driven Development Claims with Respect to Quality: A Family of Experiments , 2018, IEEE Transactions on Software Engineering.

[157]  Sven Apel,et al.  Views on Internal and External Validity in Empirical Software Engineering , 2015, 2015 IEEE/ACM 37th IEEE International Conference on Software Engineering.

[158]  Claudia Pons,et al.  Gender Gap in Computing: A Preliminary Empirical Study , 2018, CIbSE.

[159]  Claes Wohlin,et al.  Guidelines for snowballing in systematic literature studies and a replication in software engineering , 2014, EASE '14.

[160]  Dawn J. Lawrie,et al.  The impact of identifier style on effort and comprehension , 2012, Empirical Software Engineering.

[161]  Petri Ihantola,et al.  In Search of the Emotional Design Effect in Programming , 2015, 2015 IEEE/ACM 37th IEEE International Conference on Software Engineering.

[162]  Laurie A. Williams,et al.  Identifying the implied: Findings from three differentiated replications on the use of security requirements templates , 2016, Empirical Software Engineering.

[163]  Santiago Matalonga,et al.  Toward the characterization of software testing practices in South America: looking at Brazil and Uruguay , 2017, Software Quality Journal.

[164]  George M. Marakas,et al.  Effectiveness of Pair and Solo Programming Methods: a Survey and an Analytical Approach , 2015, AMCIS.

[165]  Giuseppe Scanniello,et al.  Studying abbreviated vs. full-word identifier names when dealing with faults: an external replication , 2014, ESEM '14.

[166]  João Araújo,et al.  Comparing business value modeling methods: A family of experiments , 2018, Inf. Softw. Technol..

[167]  Nicole Novielli,et al.  Sensing Developers’ Emotions: The Design of a Replicated Experiment , 2018, 2018 IEEE/ACM 3rd International Workshop on Emotion Awareness in Software Engineering (SEmotion).

[168]  Nadir Shah,et al.  Fine-Granular Model Merge Solution For Model-Based Version Control System , 2016 .