Empirical Software Engineering

, Volume 13, Issue 1, pp 97–121

Evaluating guidelines for reporting empirical software engineering studies

  • Barbara Kitchenham
  • Hiyam Al-Khilidar
  • Muhammed Ali Babar
  • Mike Berry
  • Karl Cox
  • Jacky Keung
  • Felicia Kurniawati
  • Mark Staples
  • He Zhang
  • Liming Zhu
Article

DOI: 10.1007/s10664-007-9053-5

Cite this article as:
Kitchenham, B., Al-Khilidar, H., Babar, M.A. et al. Empir Software Eng (2008) 13: 97. doi:10.1007/s10664-007-9053-5

Abstract

Background

Several researchers have criticized the standards of performing and reporting empirical studies in software engineering. In order to address this problem, Jedlitschka and Pfahl have produced reporting guidelines for controlled experiments in software engineering. They pointed out that their guidelines needed evaluation. We agree that guidelines need to be evaluated before they can be widely adopted.

Aim

The aim of this paper is to present the method we used to evaluate the guidelines and report the results of our evaluation exercise. We suggest our evaluation process may be of more general use if reporting guidelines for other types of empirical study are developed.

Method

We used a reading method inspired by perspective-based and checklist-based reviews to perform a theoretical evaluation of the guidelines. The perspectives used were: Researcher, Practitioner/Consultant, Meta-analyst, Replicator, Reviewer and Author. Apart from the Author perspective, the reviews were based on a set of questions derived by brainstorming. A separate review was performed for each perspective. The review using the Author perspective considered each section of the guidelines sequentially.

Results

The reviews detected 44 issues where the guidelines would benefit from amendment or clarification and 8 defects.

Conclusions

Reporting guidelines need to specify what information goes into what section and avoid excessive duplication. The current guidelines need to be revised and then subjected to further theoretical and empirical validation. Perspective-based checklists are a useful validation method but the practitioner/consultant perspective presents difficulties.

Categories and Subject Descriptors

K.6.3 [Software Engineering]: Software Management—Software process.

General Terms

Management, Experimentation.

Keywords

Controlled experiments Software engineering Guidelines Perspective-based reading Checklist-based reviews 

Copyright information

© Springer Science+Business Media, LLC 2007

Authors and Affiliations

  • Barbara Kitchenham
    • 4
  • Hiyam Al-Khilidar
    • 1
    • 3
  • Muhammed Ali Babar
    • 2
  • Mike Berry
    • 1
    • 3
  • Karl Cox
    • 1
  • Jacky Keung
    • 1
    • 3
  • Felicia Kurniawati
    • 1
  • Mark Staples
    • 1
  • He Zhang
    • 1
    • 3
  • Liming Zhu
    • 1
  1. 1.National ICT Australia LtdSydneyAustralia
  2. 2.Lero, The Irish Software Engineering Research CentreUniversity of LimerickLimerickIreland
  3. 3.School of Computer Science & EngineeringUniversity of New South WalesSydneyAustralia
  4. 4.Keele UniversitySchool of Computing and MathematicsStaffordshireUK