Abstract
Conducting a critical appraisal is considered as one of the essential steps in any type of qualitative synthesis. While various appraisal tools are available in the literature, the reliability and validity of these tools tend to be limited due to inconsistencies in their conceptualizations and operationalizations. In addition, limited practical guidelines are available for researchers to develop and use these appraisal form for their own synthesis studies. Our goal for this study is to use and modify components of existing tools and work to address the concerns that have been laid out in existing research. Specifically, we set out to (a) keep detailed notes of our development of the appraisal tool over time, (b) attend to the inter-rater reliability and content validity of the appraisal tool, (c) provide a guidance on how to use the tool, and (d) produce an appraisal tool that could be used by all levels of researchers. In this paper, we explain the process we undertook to establish a reliable and valid critical appraisal tool for qualitative syntheses, share the developed 22-question critical appraisal form, and provide an example of how to use the results of a critical appraisal in a sample qualitative synthesis. The developed tool will help focus qualitative synthesis researchers’ attention on the transparency and completeness of the information reported in primary studies and thereby increase the credibility of a qualitative synthesis.
Similar content being viewed by others
Data availability and material (data transparency)
The developed appraisal form was included in the article as "Appendix".
Code availability (software application or custom code)
No specialized software was used for the data management.
References
Caskurlu, S., Richardson, J. C., Maeda, Y., Kozan, K.: The qualitative evidence behind the factors impacting online learning experiences as informed by the community of inquiry framework: A thematic synthesis. Computers & Educ. 165, 104111 (2021). https://doi.org/10.1016/j.compedu.2020.104111
Maeda, Y., Caskurlu, S., Kenney, R.H., Kozan, K., Richardson, J.C.: Moving qualitative synthesis research forward in education: A methodological systematic review. Educ. Res. Rev. 35, 100424 (2022). https://doi.org/10.1016/j.edurev.2021.100424
Kenney, R.H., Maeda, Y., & Caskurlu, S. Let’s get meta: Using meta-aggregation to synthesize qualitative research in mathematics education. Proceedings of the 41st Annual Meeting of the North American Chapter of the International Group for the Psychology of Mathematics Education (PME-NA), St. Louis, MO. (2019)
Atkins, S., Lewin, S., Smith, H., Engel, M., Fretheim, A., Volmink, J.: Conducting a meta-ethnography of qualitative literature: lessons learnt. BMC Med. Res. Method. (2008). https://doi.org/10.1186/1471-2288-8-21
Bearman, M., Dawson, P.: Qualitative synthesis and systematic review in health professions education. Med. Educ. 47, 252–260 (2013). https://doi.org/10.1111/medu.12092
Burls, A.: What is critical appraisal?. Hayward Medical Communications, Kent https://www.whatisseries.co.uk/what-is-critical-appraisal/ (2009)
Blaxter, M.: Criteria for the evaluation of qualitative research papers. Med. Sociol. News 22, 68–71 (1996)
Booth, A., Papaioannou, D., Sutton, A.: Systematic Approaches to a Successful Literature Review. Sage, Thousand Oaks (2012)
Carroll, C., Booth, A., Lloyd-Jones, M.: Should we exclude inadequately reported studies from qualitative systematic reviews? An evaluation of sensitivity analyses in two case study reviews. Qual. Health Res. 22, 1425–1434 (2012). https://doi.org/10.1177/1049732312452937
Carroll, C., Booth, A.: Quality assessment of qualitative evidence for systematic review and synthesis: Is it meaningful, if so, how should it be performed? Res. Synth. Meth. 6, 149–154 (2015). https://doi.org/10.1002/jrsm.1128
Cho, M.K., Bero, L.A.: Instruments for assessing the quality of drug studies published in the medical literature. JAMA 272, 101–104 (1994)
Critical Appraisal Skills Programme.: CASP qualitative checklist https://casp-uk.net/wp-content/uploads/2018/01/CASP-Qualitative-Checklist-2018.pdf (2018)
Crowe, M., Sheppard, L.: A review of critical appraisal tools show they lack rigor: alternative tool structure is proposed. J. Clin. Epidemiol. 64, 79–89 (2011). https://doi.org/10.1016/j.jclinepi.2010.02.008
Dixon-Woods, M., Shaw, R.L., Agarwal, S., Smith, J.A.: The problem of appraising qualitative research. BMJ Qual. Saf. 13, 223–225 (2004). https://doi.org/10.1136/qshc.2003.008714
Dixon-Woods, M., Sutton, A., Shaw, R., Miller, T., Smith, J., Young, B., Bonas, S., Booth, A., Jones, D.: Appraising qualitative research for inclusion in systematic reviews: a quantitative and qualitative comparison of three methods. J. Health Serv. Res. Pol. 12, 42–47 (2007). https://doi.org/10.1258/135581907779497486
EPPI-Centre.: Review guidelines for extracting data and quality assessing primary studies in educational research. Version 0.9.7. EPPI-Centre, Social Science Research Unit (2003) https://eppi.ioe.ac.uk/CMS/Portals/0/PDF%20reviews%20and%20summaries/Guidelines%20for%20extracting%20data%20and%20quality%20assessing%20primary%20studies%20in%20educational%20research%20Version%200.97_2.doc
Finlayson, K.W., Dixon, A.: Qualitative meta-synthesis: a guide for the novice. Nurs. Researcher 15, 59–71 (2008). https://doi.org/10.7748/nr2008.01.15.2.59.c6330
Gisev, N., Bell, J.S., Chen, T.F.: Interrater agreement and interrater reliability: key concepts, approaches, and applications. Res. in Soc. and Adm. Pharm. 9(3), 330–338 (2013). https://doi.org/10.1016/j.sapharm.2012.04.004
Gough, D.: Weight of evidence: a framework for the appraisal of the quality and relevance of evidence. Res. Papers Educ. 22, 213–228 (2007). https://doi.org/10.1080/02671520701296189
Gough, D., Thomas, J., Oliver, S.: Clarifying differences between review designs and methods. Syst. Rev. 1, 1–9 (2012). https://doi.org/10.1186/2046-4053-1-28
Hannes, K., Lockwood, C., Pearson, A.: A comparative analysis of three online appraisal instruments’ ability to assess validity in qualitative research. Qual. Health Res. 20, 1736–1743 (2010). https://doi.org/10.1177/1049732310378656
Hannes K.: Chapter 4: Critical appraisal of qualitative research. In: Noyes J, Booth A, Hannes K, Harden A, Harris J, Lewin S, Lockwood C (editors), Supplementary Guidance for Inclusion of Qualitative Research in Cochrane Systematic Reviews of Interventions. (Version 1). Cochrane Collaboration Qualitative Methods Group (2011) http://cqrmg.cochrane.org/supplemental-handbook-guidance
Higgins, J.P., Green, S. (eds.): Cochrane Handbook for Systematic Reviews of Interventions. Wiley & Sons, Hoboken (2008)
Hong, Q.N., Pluye, P.: A conceptual framework for critical appraisal in systematic mixed studies reviews. J. Mixed Methods Res. 13, 446–460 (2019). https://doi.org/10.1177/1558689818770058
Katikireddi, S.V., Egan, M., Petticrew, M.: How do systematic reviews incorporate risk of bias assessments into the synthesis of evidence? A methodological study. J. Epidemiol. Commun. Health 69, 189–195 (2015). https://doi.org/10.1136/jech-2014-204711
Katrak, P., Bialocerkowski, A.E., Massy-Westropp, N., Kumar, S., Grimmer, K.A.: A systematic review of the content of critical appraisal tools. BMC Med. Res. Methodol. 4, 1–11 (2004). https://doi.org/10.1186/1471-2288-4-22
Lockwood, C., Munn, Z., Porritt, K.: Qualitative research synthesis: methodological guidance for systematic reviewers utilizing meta-aggregation. Int. J. Evid. Based Health Care 13, 179–187 (2015)
Long, A. F, Godfrey, M., Randall, T., Brettle, A. J., Grant, M. J.: Developing evidence based social care policy and practice (Part 3): feasibility of undertaking systematic reviews in social care. Nuffield Institute for Health (2002)
Long, A.F., Godfrey, M.: An evaluation tool to assess the quality of qualitative research studies. Int. J. Soc. Res. Method 7, 181–196 (2004). https://doi.org/10.1080/1364557032000045302
MacEachen, E., Clarke, J., Franche, R.L., Irvin, E.: Systematic review of the qualitative literature on return to work after injury. Scand. J. Work Environ. Health 32, 257–269 (2006)
Merriam, S.B.: Qualitative Research: A Guide to Design and Implementation. Jossey-Bass, Hoboken (2009)
Majid, U., Vanstone, M.: Appraising qualitative research for evidence syntheses: a compendium of quality appraisal tools. Qual. Health Res. 28, 2115–2131 (2018). https://doi.org/10.1177/1049732318785358
Mays, N., Pope, C.: Qualitative research: rigour and qualitative research. BMJ 311, 109–112 (1995). https://doi.org/10.1136/bmj.311.6997.109
Munthe-Kaas, H.M., Glenton, C., Booth, A., Noyes, J., Lewin, S.: Systematic mapping of existing tools to appraise methodological strengths and limitations of qualitative research: First stage in the development of the CAMELOT tool. BMC Med. Res. Method. 19, 113 (2019). https://doi.org/10.1186/s12874-019-0728-6
Noblit, G.W., Hare, R.D.: Meta Ethnography: Synthesizing Qualitative Studies. Sage, Thousand Oaks, California (1988)
Noyes, J., Popay, J., Pearson, A., Hannes, K., Booth, A.: Qualitative research and cochrane reviews. In: Higgins, J., Green, S. (eds.) Cochrane Handbook for Systematic Reviews of Interventions, pp. 571–591. John Wiley & Sons, Hoboken (2008)
Nye, E., Melendez-Torres, G.J., Bonell, C.: Origins, methods and advances in qualitative meta-synthesis. Rev. Educ. 4, 57–79 (2016). https://doi.org/10.1002/rev3.3065
Pawson, R., Greenhalgh, T., Harvey, G., Walshe, K.: Realist review-a new method of systematic review designed for complex policy interventions. J. Health Services Res. Policy 10, 21–34 (2005). https://doi.org/10.1258/1355819054308530
Pinsk, R., Curran, M.J., Poirier, R., Coulson, G.: Student perceptions of the use of student-generated video in online discussions as a mechanism to establish social presence for non-traditional students: a case study. Issues in Inf. Syst. 15, 267–276 (2014)
Sandelowski, M.: A matter of taste: Evaluating the quality of qualitative research. Nurs. Inq. 22, 86–94 (2015). https://doi.org/10.1111/nin.12080
Sandelowski, M., Barroso, J.: Handbook for Synthesizing Qualitative Research. Springer, New York (2007)
Soilemezi, D., Linceviciute, S.: Synthesizing qualitative research: reflections and lessons learnt by two new reviewers. Intl. J. Qual. Methods 17, 1–14 (2018). https://doi.org/10.1177/1609406918768014
Suri, H., Clarke, D.: Advancements in research synthesis methods: from a methodologically inclusive perspective. Rev. Educ. Res. 79, 395–430 (2009). https://doi.org/10.3102/0034654308326349
Syed, M., Nelson, S.C.: Guidelines for establishing reliability when coding narrative data. Emerg. Adulthood 3, 375–387 (2015)
Funding
The author(s) received no financial support for the research, authorship, and/or publication of this article.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this article.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix
Appendix
Critical appraisal tool and a coded example of the report by Pinsk et al. (2014).
Criteria | Question(s) | Code* | Evidence from the paper (Direct quote) and Reviewer’s notes |
---|---|---|---|
Purpose of study Description of the main purpose(s)/ aim(s)/objective(s) of the study and /or research question(s) guided research | Did author(s) provide research questions? If not, is there a purpose statement to guide the investigation? | Yes | “To what degree did the use of student-generated video discussion posts create social presence for those students?” (p. 268) |
Research design Description of what and how qualitative research deign was used and justified | Is qualitative research used for addressing the research purpose? (This may include considering the qualitative component of a mixed methods study) | Yes | “A case study methodology was adopted for this project. “(p. 268) |
Did author(s) specify their selected qualitative method? (e.g., case study, phenomenology, ethnography, grounded theory, action research, narrative inquiry, and other) | Yes | “A case study methodology was adopted for this project. “(p. 268) References supported for the methodological selection was provided | |
Did author(s) provide an explanation of why the selected qualitative method was chosen? | Yes | “when the researcher has clearly identifiable cases within established boundaries and seeks to provide a greater in-depth understanding of a phenomenon, and allows for the emergence of sometimes subtle themes that are consistent throughout the experience” (p. 268) | |
Does the provided research question(s) include qualitative language (e.g., explore, describe, understand…) and avoid quantitative language (e.g., relate, influence, effect, cause…) ? | No | Reviewers agreed the questions were not written for the purpose of the study, i.e., “The study attempted to discover the influence that this type of action [lf-created video discussion posts for a class] had on establishing social presence in online courses” (p. 267). Reviewers agreed that “to what degree” does not suit for the case study | |
Sampling Description of setting and what, how and why participants/ documents/events are selected | Did author(s) describe the setting/context for data collection? | Yes | Online course (pp. 267–268) Described about online course learning environment, Blackboard and Kaltura-brand video tools and course settings. But the authors did not describe about the contexts beyond online course |
Did author(s) provide description about the participants? (e.g., Administrators; Secondary science teachers; 4-6th grade math students; History textbooks before 2010; etc.) | Yes | Section “Participants” (pp. 269- 270) Students in class. The authors gave detailed description (age, background, level of computer literacy) for each participant | |
Did author(s) explain why the selected participants/documents/events were chosen? | Yes | "After these sessions, seven agreed to participate. Of the seven, five actually produced at least 3 video discussion responses and were interviewed for this study" (p. 269) | |
Did author(s) describe the recruitment process? (If so, was there any discussion on criteria for exclusion and inclusion?) | Yes | “The five students that participated in this study were selected based on a convenience sample.” (p. 269) | |
Did the author specify the sample size? | Yes | ||
5 (p. 269) | |||
Data collection Description of how data collection carried out and justification of the procedure | Did author(s) specify their data collection method(s)? (e.g., interviews, observation) | Yes | 60–90 min. semi-structured telephone interviews with cases (Data collection section, p. 270) “data collected by the university on ‘student engagement’ was also reviewed” (Data collection section, p. 270) |
Did author(s) provide an explanation of why the selected data collection methods were chosen? | No | No description is available to indicate why the interview method was chosen. However, it indicated why telephone interviews were conducted For student engagement data, it says “Case studies often also include the use of archival data or physical artifacts in addition to personal interviews.” (Data collection section p. 270) | |
Did author(s) provide any description of the data collection procedure? (e.g., If interviews were used, are there details on how they were conducted) | Yes | “Interviews ranged from 60 to 90 min and were conducted via telephone, since the participants were part of an online class and were geographically dispersed.” (Data collection section p. 270) | |
Data analysis Description of analytic procedure to reach findings | Did author(s) specify the data analysis method(s) they used? (e.g., constant comparison, thematic analysis, cross-case analysis…) | Yes | Within-case and cross-case analysis “After the interviews were transcribed, they were initially reviewed. This within-case analysis provided initial familiarity with the data. The data was then coded for themes.” (Data analysis section. p. 270) “This necessitated the use of the second coder. Then, cross-case analysis was used to facilitate insights beyond preliminary impressions. Themes began to emerge as the data was coded and analyzed. Enumeration and categorization of the data was also performed at this stage.” (Data analysis section. p. 270) |
Did author(s) provide justification or explanation of why they chose the data analysis method(s)? | No | No description is available | |
Did author(s) provide description of data analysis procedure? (e.g., If thematic analysis is used, is it clear how the categories/themes were derived from the data?) | No | Themes, categorization and enumeration were listed. However, the authors did not provide the detailed process of identification of themes, and they only provided the general description | |
Findings Description of “what” author(s) found in the study. These findings will serve as data in qualitative synthesis | Did the authors' reported results address the research questions? | Yes | “Three critical components emerged as important elements among the participants regarding the establishment of social presence within the online experience in this class: course engagement, projection of self, and connection to other members of the class and the instructor.” (p. 270) The authors also identified three sub-factors: time considerations, technical issues, and interest in being innovative, and these sub-factors are the same for all components |
Did author(s) provide verbatim evidence representing the participant’s voices? (e.g., direct quotations or transcript excerpts) | Yes | Included keywords, quotes (pp. 270–272) | |
Value of the research Description of scholarly contribution to the relevant filed and an implication for both practice and research | Did the author(s) describe implications of the study findings? | Yes | "This study demonstrates that there may be some significant potential in using video for online discussions as a means to establishing social presence for students.” (Future Research section, p. 273) “Further research in this area can establish appropriate parameters for the use of this technology.” (Future research section. p. 273) |
Did author(s) state the contribution to the field? Or did author(s) state how the study fits into the existing literature? | No | In the literature review, it says “most research on this topic has been with traditional college students” (p. 268). However, the authors do not clarify whether a goal of the study is to fill the gap in the literature Last sentence in the limitations of the study simply says, “the study does contribute to the overall body of knowledge in the field of online education.” | |
Trustworthiness and reliability Descriptions for checking trustworthiness/reliability may include but are not limited to inter-rater reliability, member checking, researcher bias, role of researchers, piloting, data triangulation, etc | Did the author(s) discuss methods used to enhance the quality of data collection instruments? (e.g., piloting interview protocols, getting expert feedback on survey questions…) | Yes | "The protocol was reviewed by two subject matter experts in order to establish the validity of the questions used in the process" (Data collection section, p. 270) |
Did author(s) describe methods used to enhance the reliability of their data analysis? (e.g., inter-rater reliability, member checking, clarifying researcher bias, data triangulation, etc | Yes | "In order to assure maximum reliability in the data, each interview was transcribed by two different researchers and then the transcripts were compared to assure accuracy of the data." (Data collection section, p. 270) |
Rights and permissions
About this article
Cite this article
Maeda, Y., Caskurlu, S., Kozan, K. et al. Development of a critical appraisal tool for assessing the reporting quality of qualitative studies: a worked example. Qual Quant 57, 1011–1031 (2023). https://doi.org/10.1007/s11135-022-01403-y
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11135-022-01403-y