Hostname: page-component-7dd5485656-jtdwj Total loading time: 0 Render date: 2025-10-25T05:35:33.466Z Has data issue: false hasContentIssue false

The inclusion or exclusion of studies based on critical appraisal results in JBI qualitative systematic reviews: An analysis of practices

Published online by Cambridge University Press:  23 October 2025

Romy Menghao Jia*
Affiliation:
JBI, School of Public Health, The University of Adelaide , Adelaide, SA, Australia
Cindy Stern
Affiliation:
JBI, School of Public Health, The University of Adelaide , Adelaide, SA, Australia Health Evidence Synthesis, Recommendations and Impact (HESRI), School of Public Health, The University of Adelaide , Adelaide, SA, Australia
*
Corresponding author: Romy Menghao Jia, romy.jia@adelaide.edu.au
Rights & Permissions [Opens in a new window]

Abstract

Critical appraisal is a core component of JBI qualitative evidence synthesis, offering insights into the quality of included studies and their potential influence on synthesized findings. However, limited guidance exists on whether, when, and how to exclude studies based on appraisal results. This study examined the methods used in JBI qualitative systematic reviews and the implications for synthesized findings. In this study, a systematic analysis of qualitative reviews published between 2018 and 2022 in JBI Evidence Synthesis was conducted. Data on decisions and their justifications were extracted from reviews and protocols. Descriptive and content analysis explored variations in the reported methods. Forty-five reviews were included. Approaches reported varied widely: 24% of reviews included all studies regardless of quality, while others applied exclusion criteria (36%), cutoff scores (11%), or multiple methods (9%). Limited justifications were provided for the approaches. Few reviews cited methodological references to support their decisions. Review authors reported their approach in various sections of the review, with inconsistencies identified in 18% of the sample. In addition, unclear or ambiguous descriptions were also identified in 18% of the included reviews. No clear differences were observed in ConQual scores between reviews that excluded studies and those that did not. Overall, the variability raises concerns about the credibility, transparency, and reproducibility of JBI qualitative systematic reviews. Decisions regarding the inclusion or exclusion of studies based on critical appraisal need to be clearly justified and consistently reported. Further methodological research is needed to support rigorous decision-making and to improve the reliability of synthesized findings.

Information

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2025. Published by Cambridge University Press on behalf of The Society for Research Synthesis Methodology

Highlights

What is already known?

  • Critical appraisal is a key step in JBI qualitative evidence synthesis, intended to inform confidence in review findings.

  • Guidance is limited on whether studies should be excluded based on appraisal results and what criteria should be used for exclusion.

What is new?

  • JBI qualitative systematic reviews use varied approaches to handle critical appraisal results.

  • Approaches and justifications are not always clearly reported or aligned across protocols and final reports.

  • Few reviews cited references to support their appraisal-related decisions, and ConQual scores did not differ markedly by exclusion strategy.

  • There is a need for methodological development to support transparency and reproducibility.

Potential impact for RSM readers

  • Our findings highlight that it is important for reviewers to make informed and deliberate decisions about how to handle critical appraisal results.

  • This study promotes greater transparency in qualitative systematic review methods.

1 Introduction

Qualitative systematic reviews play a crucial role in evidence-based healthcare by synthesizing high-quality information regarding the perspectives and experiences of patients, caregivers, providers, and other stakeholders by addressing questions related to the usability, meaningfulness, feasibility, and appropriateness of interventions and practices, factors that are essential for implementing holistic and person-centred care. Review findings are used increasingly to support clinical decision-making, health policy development, practice guidelines, and implementation strategies.Reference Lockwood, Munn and Porritt1, Reference Carroll2 Different methodologies exist for the synthesis of qualitative findings, such as thematic synthesis, narrative synthesis, meta-ethnography, and meta-aggregation.Reference Booth, Noyes and Flemming3 The JBI meta-aggregative approach was developed in the early 2000s, led by Professor Alan Pearson.Reference Pearson4 The approach is grounded in the philosophical principles of pragmatism and aims to generate meaningful recommendations to inform practice.Reference Porritt, Evans, Bennett, ELC, Porritt, Pilla and Jordan5

One fundamental step of JBI qualitative systematic review is critical appraisal.Reference Lockwood, Munn and Porritt1, Reference Porritt, Evans, Bennett, ELC, Porritt, Pilla and Jordan5, Reference Munn, Porritt, Lockwood, Aromataris and Pearson6 Critical appraisal provides an in-depth understanding of the quality of the included studies. In JBI qualitative reviews, critical appraisal is closely linked to the level of confidence that can be placed in synthesized findings.Reference Munn, Porritt, Lockwood, Aromataris and Pearson6, Reference Munthe-Kaas, Bohren and Glenton7 The process is especially critical in JBI meta-aggregative reviews as the findings are intended to inform policy or practice,Reference Porritt, Evans, Bennett, ELC, Porritt, Pilla and Jordan5 where for methodologies focused on generating theoretical insights such as meta-ethnography, the practice remains contentious.Reference Carroll and Booth8, Reference Soundy and Heneghan9 Different standards, criteria, checklists, and tools have been developed to assist reviewers in assessing the quality of primary qualitative studies, such as but not limited to the JBI Checklist for Qualitative Research,Reference Lockwood, Munn and Porritt1 the CochrAne qualitative MEthodological LimitatiOns Tool (CAMELOT),Reference Munthe-Kaas, Noyes and Noyes10 and the Evaluation Tool For Qualitative Studies (ETQS).Reference Long, Godfrey, Randall, Brettle and Grant11 The number of available appraisal tools has grown rapidly in recent years, with more than 100 instruments now identified for qualitative research.Reference Munthe-Kaas, Glenton, Booth, Noyes and Lewin12, Reference Majid and Vanstone13 Critical appraisal using standardized tools is generally superior to ad hoc or unclear assessments, as it offers review authors a systematic and transparent approach to evaluating study quality.Reference Munthe-Kaas, Noyes and Noyes10 The approach to assessing the quality of qualitative studies generally involves determining risks to study rigour or trustworthiness, by looking at the dependability, credibility, and transferability of qualitative studies (Table 1).Reference Porritt, Gomersall and Lockwood14 Most assessments use a series of criteria that can be scored as being ‘yes’, ‘no’, ‘unclear’, and in some instances ‘not applicable’.

Table 1 Criteria for assessing the quality of qualitative research.Reference Porritt, Gomersall and Lockwood14

There is ongoing debate regarding which criteria or checklist to use to evaluate qualitative research and whether studies deemed to be of lower methodological quality should be excluded based on appraisal results in qualitative reviews and, if so, what criteria or methods should guide the exclusions, resulting in a lack of agreement in the field.Reference Carroll and Booth8 Historically, one approach has been to exclude studies with lower methodological quality to ensure the findings of a review are based on the ‘best available evidence’.Reference Porritt, Gomersall and Lockwood14, Reference Aromataris and Lockwood15According to the current JBI guidance, the decision whether to exclude studies based on critical appraisal results is up to the review team.Reference Porritt, Evans, Bennett, ELC, Porritt, Pilla and Jordan5 Exclusion decisions can be made based on whether a study meets a predetermined proportion of all criteria or whether it satisfies specific criteria. Reviewers can also weight certain criteria differently. Similarly, Cochrane guidance for Cochrane and Campbell reviews states that an assessment of methodological strengths and limitations, whether using a tool or not, can be employed to exclude studies in the sampling stage in the review process.Reference Munthe-Kaas, Noyes and Noyes10 Review authors may also choose not to exclude studies based on critical appraisal results because qualitative studies with methodological limitations can provide valuable insights, particularly in contexts where evidence is limited, or contribute perspectives from populations that are often underrepresented in the literature.Reference Ames, Booth, Noyes and Noyes16 The ENTREQ statement, a reporting guide for qualitative syntheses, in item 13, states that appraisal results should be presented, indicating which articles, if any, were weighted or excluded based on the assessment, along with the rationale.Reference Tong, Flemming, McInnes, Oliver and Craig17 While current guidance allows the review team discretion, it is unclear how this is currently undertaken and what implications it has on synthesized results.

The objective of this study was to investigate current practices on the inclusion or exclusion of studies based on critical appraisal results in JBI qualitative systematic reviews. The JBI approach was chosen for this study because appraisal plays a particularly crucial role in JBI meta-aggregation, as the methodology explicitly focuses on informing practice and critical appraisal results directly influence the level of confidence in synthesized findings. This study aims to provide insights that may inform future methodological development within the JBI framework and potentially across the broader evidence synthesis community.

The specific questions addressed are as follows:

  • What methods are applied in JBI qualitative systematic reviews to determine the inclusion or exclusion of studies based on critical appraisal results?

  • What are the justifications provided for applying these methods?

  • Is there a difference in the confidence of synthesized findings between reviews that applied exclusion based on critical appraisal results and those that included all studies regardless of quality?

2 Methods

This study included systematic reviews that exclusively included qualitative evidence, reported using the JBI meta-aggregative approach for qualitative synthesis and published in the peer-reviewed journal JBI Evidence Synthesis (previously JBI Database of Systematic Reviews and Implementation Reports) between 2018 and 2022. The timeframe was selected to capture the most recent 5 years of publications available at the time of data collection (early 2023). Reviews that included evidence types other than qualitative evidence were excluded to ensure methodological consistency and comparability across the sample.

A full list of qualitative reviews published during the time period was obtained from the JBI Evidence Synthesis editorial office (CS is the Deputy Editor-in-Chief of the journal). As part of the editorial process, authors are required to specify the review type (e.g. qualitative), which is then verified by editors, copy editors, and proofreaders during production. While we acknowledge that qualitative systematic reviews using the JBI meta-aggregative approach are also published in other journals, this study focused exclusively on JBI reviews published in JBI Evidence Synthesis, as these were required to adhere to JBI standards of conduct and reporting.Reference Aromataris, Stern and Lockwood18

2.1 Data extraction

Full-text reports of all included reviews were retrieved by RJ. Data extraction was conducted using a standardized data extraction tool (Appendix I, Supplementary Materials). A pilot extraction of five reviews was conducted by both authors independently (RJ and CS) prior to formal data extraction. Upon completion of the pilot, the results and the tool were discussed, and significant modifications were made to the tool. A second round of piloting was then conducted independently by both authors on the same five reviews using the revised tool. Following the second pilot, only minor adjustments were needed. Subsequently, data were extracted from 40 reviews using MS Excel (RJ and CS, 20 reviews each). All data were reviewed by RJ for consistency prior to analysis, with minor modifications made as necessary.

The extracted data included title, year published, country of corresponding author, methods used to address critical appraisal results and their justifications, and the ConQual score. Methods and justifications were extracted verbatim from the review to mitigate any potential for reinterpretation from the authors during data extraction. In addition, a priori protocols for each included review were identified and retrieved where methods pertinent to handling of critical appraisal results were extracted to ensure comprehensive data extraction. References provided in reviews and protocols to support their specified methods and justifications were also extracted.

2.2 Data analysis and presentation

Descriptive analysis and basic content analysis were conducted (RJ) to examine the characteristics of the included reviews and methods regarding the inclusion or exclusion of studies based on the methodological quality of studies and justifications.

Specifically, for the characteristics of the included reviews, data were analysed descriptively. Frequency distributions were presented to describe the year and geographical distribution of studies and reporting of methods.

To analyse the methods employed and their justifications, an inductive content analysis was conducted, an approach well suited for exploring phenomena with limited or fragmented prior knowledge.Reference Elo and Kyngäs19 In this study, while critical appraisal in qualitative evidence synthesis is an established step, there is very limited research specifically examining how review authors make inclusion or exclusion decisions based on appraisal results and how they justify these decisions. Choosing an inductive approach allowed us to explore and categorize patterns in the methods and justifications. The coding process started with reading and re-reading of the extracted data, during which notes were written to describe the methods detailed in each review. The coding framework was developed based on the initial notes. The data were then coded against the framework, with each review being assigned a single code of the method. The frequency of methods was analysed.

During the analysis, it was observed that some review authors described their methods and justification across multiple sections within the review or in both the review and its a priori protocol. In some instances, this resulted in unnecessary repetition, while in others, it led to inconsistencies in the reported information. For example, some protocols indicated that a specific method would be applied, but the corresponding review described a deviation from the methods stipulated a priori. Additionally, where methods for handling of the results of appraisal were described in multiple sections of a review, inconsistencies were identified between the methods reported and appendices, the methods and the results, and the results and the appendices. These inconsistencies were identified in nine reviews, and after discussion between the authors, the following principles were established to handle data:

  • If inconsistencies are evident between data extracted from the protocol and review, details extracted from the review will be analysed.

  • If inconsistencies are evident between data extracted from the methods and results in the review, details extracted from the results will be analysed.

  • If inconsistencies are evident between data extracted from the methods/results and appendices, details extracted from the methods/results will be analysed.

These principles assumed that, compared to the protocol, the review itself offered a more accurate and comprehensive account of the actual conduct regarding appraisal results, acknowledging that the results section of the completed review is considered the most reliable source of information on the applied method.

For reviews where the stated method was inconsistent, only the justifications provided for the analysed method as per the principles defined above were examined. Conversely, when the stated method was consistent throughout the review sections and protocol, justifications provided in all relevant sections (protocol or multiple sections of the review) were included in the analysis.

The analysis of justifications for each method was conducted following the coding of the methods. Reviews were categorized based on their assigned code of method, and justifications were then analysed within each group. Where possible, an inductive content analysis was applied to identify themes in the justifications for each method group. When justification data were limited and insufficient for a method group, verbatim justifications from individual reviews were presented. The frequency of each category was recorded to indicate how many reviews provided justifications.

The association between the confidence of synthesized findings and the method applied was analysed using extracted ConQual scores. In JBI qualitative systematic reviews, synthesized findings are required to be assigned a level of confidence using the ConQual approach,Reference Munn, Porritt, Lockwood, Aromataris and Pearson6 a process providing essential information for knowledge users when considering using synthesized findings to inform decisions. The JBI ConQual approach is a ranking system for synthesized findings that evaluates the dependability of included studies and credibility of extracted findings included that inform synthesized findings. Credibility is assessed during the data extraction stage by evaluating the congruence between the author’s interpretation and the supporting data. Dependability, on the other hand, is determined using the critical appraisal results from five questions (Q2, Q3, Q4, Q6, and Q7) in the JBI critical appraisal checklist.Reference Lockwood, Munn and Porritt1 This study calculates the number and percentage of synthesized findings classified as high, moderate, low, and very low confidence for each method to examine potential differences in ConQual scores between reviews that applied exclusion criteria and those that did not.

Results are presented in tables and figures with explanations provided to describe the findings.

3 Results

3.1 Study inclusion

Forty-seven qualitative systematic reviews were published in JBI Evidence Synthesis between 2018 and 2022. Two reviews were excluded because their protocols were unavailable despite two attempts to contact the authors. As a result, 45 reviews were included in this study (Appendix II, Supplementary Materials).

3.2 Characteristics of included reviews

3.2.1 Year and country of publication

Most qualitative reviews were conducted by authors in Canada (12, 27%), Australia (10, 22%), and the UK (8, 18%). The number of published qualitative reviews ranged from 6 in 2021 to 13 in 2022, with other years falling in between (Figure 1).

Figure 1 The number of qualitative reviews published by country and year.

3.2.2 Reporting of methods regarding the inclusion or exclusion of studies based on methodological quality

All 45 included reviews utilized the JBI Checklist for Qualitative Research, in compliance with current JBI guidance for the conduct of JBI qualitative syntheses (Table 2).Reference Lockwood, Munn and Porritt1 Of these, 44 (98%) outlined their method to the inclusion or exclusion of studies based on methodological quality, while one (2%) did not specify their method in either their review or protocol.Reference Forough, Wong and Lau20 Among the 44 (98%) reviews that specified their method, all (98%) provided this information in the review, with 20 (45%) also specifying it in the protocol. Notably, among the 20 that reported the method in their protocol, the majority were published in 2022 (Figure 2). In contrast, none of the reviews published in 2018 reported their methods for handling the results of appraisal in their protocol (Figure 2).

Table 2 Methods regarding the inclusion or exclusion of studies based on methodological quality.

Figure 2 The number of reviews specified a priori methods in the protocol over time.

Of the 44 (98%) reviews that specified their method in the review, 35 (78%) included this information in the methods section, 29 (64%) in the results section, and 18 (40%) in an Supplementary Materials, with 25 (56%) reviews reporting it in more than one section (Figure 3). Among the 29 (64%) reviews that specified their methods in the results section, the majority provided this information in the methodological quality section (21 reviews, 47%), followed by the study inclusion section (9 reviews, 20%). Two reviews (4%) reported their method in both the study inclusion and methodological quality sections, while one review (2%) presented it under the results header instead of subsections. Inconsistent reporting across sections was identified in eight reviews (18%).

Figure 3 Methods reported in review sections and in results sections.

3.3 Manuscript sections detailing inclusion or exclusion of studies based on methodological quality and justifications

As shown in Table 2, of the 45 included reviews, one (2%) did not report any methods to study inclusion based on methodological quality in either the protocol or review. Eleven reviews (24%) opted to include all studies regardless of the methodological quality. Twenty-five (56%) reviews applied an exclusion-based method, which involved designating certain items on the appraisal checklist as essential (16 reviews, 36%), applying a cutoff score or percentage (5 reviews 11%), or employing a combination of multiple methods (4 reviews, 9%). Eight reviews (24%) mentioned their methods without sufficient clarity (Appendix III, Supplementary Materials).

3.4 Include all studies regardless of methodology quality

Of the 11 (24%) reviews that opted to include all studies regardless of methodological quality, eight reviews provided justifications for using this method. One review provided only a referenceReference Lockwood, Porrit, Munn, Aromataris and Munn21 alongside its choice.Reference Tuomikoski, Parisod, Lotvonen and Valimaki22 These justifications broadly aligned with five themes: ensuring a comprehensive understanding of the phenomenon, mitigating the limited number of studies available in the field, considering the impact of methodological quality during the analysis, acknowledging the limitations of the appraisal process, and accounting for the inherent nature of qualitative research (Table 3).

Table 3 Justifications for including all studies regardless of methodology quality.

Table 4 Exclusion methods.

There were four references provided across four reviews to support these decisions (Appendix IV, Supplementary Materials). JBI guidance was cited in three reviews.[Reference Lockwood, Porrit, Munn, Aromataris and Munn21, Reference Lockwood, Porritt and Munn30] One review cited the work of LeungReference Leung31 and Kisely and KendallReference Kisely and Kendall32 to highlight limitations in the appraisal process, noting that excluding studies based on appraisal might compromise the diverse nature of qualitative studies. Two reviews referenced Dixon-Woods et al.Reference Dixon-Woods, Booth and Sutton33 to support the argument that even studies of lower quality can provide valuable insights, contributing to a richer understanding of the research phenomenon.

3.5 Exclusion methods

Of the 45 included reviews, 25 (56%) opted to apply exclusion-based methods aligned to the results of methodological quality assessment, with the most common being the requirement to fulfil specific items on the appraisal checklist (16 reviews) (Table 4).

3.5.1 Certain items in the appraisal checklist deemed essential

Among the reviews that selected specific items in the appraisal checklist as essential, nearly all (15), with the exception of one,Reference Holopainen and Hakulinen34 identified Q8 as a mandatory criterion. Q2 and Q4 were also frequently deemed essential, while Q1 and Q6 were not required by any of the reviews (Table 5). The number of items deemed essential varied, ranging from 1 to 8 (mean 4.1, median 3.5). The combination of items chosen by reviews can be found in Appendix V, Supplementary Materials.

Table 5 Frequency of items required in the checklist to determine inclusion/exclusion.

For the items deemed essential, most reviews (9 out of 16) did not specify what response (i.e. ‘yes’, ‘no’, ‘unclear’, and ‘not applicable’) constituted meeting the criteria. Five reviews explicitly required a ‘yes’ response, while two reviews also accepted ‘unclear’ as sufficient (Appendix VI, Supplementary Materials). One review justified the acceptance of ‘unclear’, arguing that this was necessary to retain a sufficient number of studies with relevant findings of suitable quality to address the review questions and that the assessment of confidence in synthesized findings would inform readers about the level of confidence in findings.Reference Small, de Boer and Swab35

Six reviews provided justifications for their specified method, which were categorized into six themes (Table 6). Three reviews emphasized the importance of ensuring that participants’ voices were adequately captured.Reference Salmond, Salmond, Ames, Kamienski and Holly36 Reference Kassam, Butcher and Marcellus38 One review emphasized that researcher reflexivity is essential to align with accepted standards for reporting qualitative research,Reference Small, de Boer and Swab35 supported by references.Reference O’Brien, Harris, Beckman, Reed and Cook39, Reference Tong, Sainsbury and Craig40 Ensuring methodological approaches were appropriate was mentioned in one review.Reference Salmond, Salmond, Ames, Kamienski and Holly36 Another review based its justification on adherence to the decisions made in a previous review protocol.Reference Barnes, Jordan and Broom41 One review justified the decision in an attempt to optimize confidence in synthesized findings.Reference Macdonald, Martin-Misener, Weeks, Helwig, Moody and MacLean42 One review stated the limited reporting of certain items (Q1 and Q6) in qualitative research as the reason for not selecting them as mandatory.Reference Small, de Boer and Swab35

Table 6 Justification for certain items in the appraisal checklist deemed essential.

Table 7 Cutoff score details per review.

3.5.2 Cutoff score

Five reviews applied a cutoff score to determine inclusion/exclusion(Table 7). Two reviews required eight items, one review required seven items, and another required five items to be satisfied for inclusion. One review, instead of specifying the number of required items, set a minimum overall score of 5 based on its scoring system. In terms of the required responses to these questions, four reviews specified that only ‘yes’ responses were acceptable, while the other review did not specify it. Three reviews explicitly stated that all items were weighted equally, while two did not provide this information. One review also specified the score for ‘unclear’ and ‘not applicable’ responses, being 0 for ‘unclear’ and subtracting ‘not applicable’ responses from the total score. Two reviews provided justifications for applying a cutoff score, both emphasizing the importance of ensuring that only high-quality studies inform synthesized findings.

3.5.3 Multiple methods utilized

Four reviews employed multiple methods (Table 8). Two reviews utilized a cutoff score and team discussions. Both required studies to meet ‘yes’ responses for eight items, and for studies that did not meet this criterion, it was stated that reviewers will meet and discuss their inclusion. One of the two reviews also reported that they decided to include a study that did not reach the cutoff because ‘it provided good findings’.Reference Parsons, Gaudine, Patrick and Busby49

Table 8 Multiple methods reported per review.

Table 9 The number and percentage of high-, moderate-, low-, and very low-confidence synthesized findings for each method identified.

Note: Percentages may not total 100% due to rounding.

Another two reviews required a certain percentage of items to be met while also setting additional eliminatory criteria (Table 8). Only one review provided a justification for using multiple methods, explaining that this decision was informed by an evaluation of the first five included studies conducted by three reviewers.

3.6 ConQual score by methods

There were no apparent differences in ConQual scores between reviews that applied exclusion criteria and reviews that included all studies regardless of methodology quality. There are 158 synthesized findings in total across the 36 reviews with a clearly categorized method, excluding those that did not report their method (one review) or provided unclear descriptions (eight reviews) (Table 2). As shown in Table 9, reviews using the exclusion methods based on certain items in the checklist had the highest proportion of high-confidence synthesized findings (15%). For moderate-confidence synthesized findings, the exclusion method using multiple methods had the highest proportion (58%). Reviews that employed multiple methods showed the highest percentage of low-confidence synthesized findings (42%), while those using the cutoff score method reported the highest proportion of very low-confidence findings (20%).

4 Discussion

The findings of this study demonstrate variations in methods applied in JBI qualitative systematic reviews regarding the inclusion or exclusion of studies based on methodological quality. About one-third of the reviews chose to include studies regardless of methodological quality. This approach was justified by providing a comprehensive understanding of the phenomenon or the limited number of available studies.Reference Tuomikoski, Parisod, Lotvonen and Valimaki23 Reference Small, Porr, Swab and Murray25 Some reviews suggested that the quality concerns of the included studies could be addressed in the analysis process, either by the elimination of unsupported findings during data extraction or via ConQual, which informs the readers of the confidence they can place on synthesized findings.Reference Bayo, Alobo, Sauve, Feyissa and Tolu.24, Reference Keeping-Burke, McCloskey, Donovan, Yetman and Goudreau53 It is unsurprising that authors make connections between critical appraisal and ConQual, as the answers for five questions in the critical appraisal tool inform the ConQual assessment.

It is worth noting that some authors also outlined methodological concerns for excluding studies based on the critical appraisal results due to limitations in the appraisal process. They suggested that appraisal is based on reviewers’ interpretation of the included data, and in-depth methodological details are often limited due to journal length and editorial requirements.Reference Tuomikoski, Parisod, Lotvonen and Valimaki23 There were also reviewers who argued that poor-quality qualitative studies can provide valuable findingsReference Bayo, Alobo, Sauve, Feyissa and Tolu.24, Reference Ramasamy, Rillotta and Alexander28 and expressed concern that using the checklist might compromise the diverse nature of qualitative research, suggesting further development of the appraisal framework.Reference McCloskey, Keeping-Burke, Witherspoon, Cook and Morris27 This is an area of ongoing debate historically as appraisal is a convention for quantitative studies, which prioritize standardization and objectivity, whereas qualitative research is context-dependent, interpretative, and grounded in naturalistic inquiry,Reference Leung31, Reference Schwandt54 making it less suited for rigid evaluation criteria. These concerns raise valid questions about the suitability of using appraisals as a basis for excluding studies in qualitative reviews.

In contrast, the majority of reviews applied exclusion parameters, with the most common method being the selection of specific checklist items that were deemed essential. Although the items selected and their combinations varied, most of them reflected an emphasis on ensuring participants’ voices were captured. One review justified this decision by referencing a previous review,Reference Barnes, Jordan and Broom41 highlighting the limited methodological research and guidance available in this area. Interestingly, instead of leaving the methodological concerns to be addressed in ConQual, there was a review that chose criteria required by ConQual as mandatory to optimize confidence in synthesized findings.Reference Macdonald, Martin-Misener, Weeks, Helwig, Moody and MacLean42 However, it remains unclear whether selecting criteria aligned with ConQual necessarily results in high-confidence findings as anticipated, which warrants further investigation in future research. In addition to Q8, questions that assess the congruity between the research methodology and the research question (Q2), as well as the representation and analysis of data (Q4), were also commonly applied.

Methods using cutoff scores primarily aimed to include only high-quality studies, and a small subset of reviews adopted multiple methods. An interesting justification involved reviewers evaluating the first few included studies to inform their methods;Reference Didier, Dzemaili and Perrenoud52 however, limited details on the implementation of these methods were reported. One review appeared to comment on their method based on appraisal results, suggesting that inclusion decisions may have been made after evaluating appraisal results rather than following a predefined strategy.Reference Jessen-Winge, Petersen and Morville55

The findings indicate no clear differences in ConQual scores between reviews that applied exclusion criteria based on methodological quality and those that included all studies regardless of quality. Our initial hypothesis was that excluding lower-quality studies would lead to higher ConQual scores. However, confirming this would require investigating critical appraisal results and reintroducing the excluded studies in reviews, reassessing the synthesized findings, and comparing the resulting ConQual scores, which is beyond the scope of this study. Future research is encouraged to explore this issue further. For example, a Study Within a Review (SWAR) could be conducted to examine how different approaches to handling appraisal results influence the ConQual score within the same review.Reference Devane, Burke and Treweek56 The findings of this study also found a lack of clarity in the reporting of methods used in JBI qualitative systematic reviews related to the inclusion or exclusion of studies based on critical appraisal results. Although most reviews (44 of 45) outlined a method, there were eight cases (Appendix III, Supplementary Materials) where the reporting was unclear, and the specific method applied could not be determined. Among the 36 where the method could be identified, only 17 provided justifications, many of which were brief and minimal. This aligns with observations in the existing literature, which have similarly highlighted the limited reporting of methods in qualitative reviews.Reference Dixon-Woods, Booth and Sutton33 Inconsistencies were found between protocols and different sections of individual reviews. Only a few reviews provided justifications for their chosen methods, and the justifications often lacked sufficient detail. Even fewer reviews cited references to support their decisions, highlighting a lack of methodological research and guidance in this area.

Overall, there are substantial inconsistencies in the approaches taken and the rationales provided. At this stage, it may not be possible to make a definitive recommendation about whether studies should always be included or excluded based on appraisal results, under what conditions studies should be excluded, and suggesting criteria or decision rules for applying exclusion thresholds or essential appraisal items. And it remains uncertain whether consistency in the methodological choices is needed—or even reasonable to expect—given the unique nature of qualitative evidence, the ongoing debate around appraisal in qualitative research, and the unclear implications for confidence in synthesized findings. We recommend targeted methodological research to examine the implications of different appraisal-related decisions and theoretical development on the alignment between these decisions and the philosophical underpinnings of the methodology, which is critical for determining whether greater methodological consistency is desirable or necessary. In addition, clear, consistent, and transparent reporting is essential. Poor reporting potentially impacts the transparency and reproducibility of the review and makes it difficult to assess how appraisal decisions are being applied in practice. This study has several limitations that should be acknowledged. First, the scope of the review was restricted to publications up to 2022. Additional reviews have since been published. However, there were no substantive changes in JBI methodological guidance on the inclusion or exclusion of studies based on critical appraisal outcomes until late 2024.Reference Porritt, Evans, Bennett, ELC, Porritt, Pilla and Jordan5 Consequently, the evidence synthesized from this period reflects current methodological practice and remains applicable, ensuring that the findings of this study are both timely and relevant. According to the 2024 updated guidance,Reference Porritt, Evans, Bennett, ELC, Porritt, Pilla and Jordan5 reviewers are expected to provide a detailed rationale regarding which criteria were used and why particular criteria were selected as key markers of study quality. Future research is encouraged to examine how reviews published after this update address these requirements. Second, the analysis was limited to JBI qualitative systematic reviews, meaning the findings may not be generalizable to reviews conducted using other methodologies or frameworks. Future work is encouraged to explore practices across different qualitative synthesis methodologies. Third, the limited and often unclear reporting of methods made it challenging to determine with certainty the methods applied in some reviews. Finally, the principles used to address inconsistencies between protocols and different sections of reviews were based on the authors’ assumption that the review itself provides the most accurate account of how critical appraisal results were addressed, with the results section being considered the most reliable source. This assumption, while practical, may not fully capture the complexities of the decision-making processes documented across various sections or in the protocol.

5 Conclusion

This study demonstrates that JBI qualitative systematic reviews employ a variety of methods to determine the inclusion or exclusion of studies based on critical appraisal results. These methods range from including all studies irrespective of their methodological quality to applying exclusion criteria based on specific appraisal items, cutoff scores, or a combination of methods. The findings highlight the challenges in addressing critical appraisal results within qualitative evidence synthesis and the limited and inconsistent reporting of these methods. Additionally, no clear differences in ConQual scores were found between reviews that applied exclusion criteria and those that included all studies regardless of methodological quality. Future research is encouraged to investigate this issue further to better understand its implications. Addressing these gaps in future research in this area is essential for ensuring more robust and transparent qualitative systematic reviews.

Acknowledgements

We acknowledge Dr Edoardo Aromataris for his valuable contributions during the conceptualization and final review of this study.

Author contributions

Romy Menghao Jia: conceptualization, methodology, data extraction, formal analysis, writing—original draft, writing—reviewing and editing, and project administration.

Cindy Stern: conceptualization, methodology, data extraction, writing—reviewing and editing, and supervision.

Competing interest

The authors declare none.

Data availability statement

The data used for analysis are publicly available at https://doi.org/10.25909/30125557.v2.

Supplementary material

To view supplementary material for this article, please visit http://doi.org/10.1017/rsm.2025.10042.

References

Lockwood, C, Munn, Z, Porritt, . Qualitative research synthesis: methodological guidance for systematic reviewers utilizing meta-aggregation. JBI Evid Implement. 2015;13(3):179187.Google ScholarPubMed
Carroll, C. Qualitative evidence synthesis to improve implementation of clinical guidelines. BMJ. 2017;356:j80.10.1136/bmj.j80CrossRefGoogle ScholarPubMed
Booth, A, Noyes, J, Flemming, K, et al. Structured methodology review identified seven (RETREAT) criteria for selecting qualitative evidence synthesis approaches. J Clin Epidemiol. 2018;99:4152.10.1016/j.jclinepi.2018.03.003CrossRefGoogle ScholarPubMed
Pearson, A. Balancing the evidence: incorporating the synthesis of qualitative data into systematic reviews. JBI Evid Implement. 2004;2(2):4564.Google Scholar
Porritt, K, Evans, C, Bennett, C et al. Systematic reviews of qualitative evidence. In: ELC, Aromataris, Porritt, K, Pilla, B, Jordan, Z, eds. JBI Manual for Evidence Synthesis. JBI; 2024.Google Scholar
Munn, Z, Porritt, K, Lockwood, C, Aromataris, E, Pearson, . Establishing confidence in the output of qualitative research synthesis: the ConQual approach. BMC Med Res Methodol. 2014;14:17.10.1186/1471-2288-14-108CrossRefGoogle ScholarPubMed
Munthe-Kaas, H, Bohren, MA, Glenton, C, et al. Applying GRADE-CERQual to qualitative evidence synthesis findings—paper 3: how to assess methodological limitations. Implement Sci. 2018;13(1):9. https://doi.org/10.1186/s13012-017-0690-9.CrossRefGoogle ScholarPubMed
Carroll, C, Booth, . Quality assessment of qualitative evidence for systematic review and synthesis: is it meaningful, and if so, how should it be performed? Res Synth Methods. 2015;6(2):149154.10.1002/jrsm.1128CrossRefGoogle ScholarPubMed
Soundy, A, Heneghan, NR. Meta-ethnography. Int Rev Sport Exerc Psychol. 2022;15(1):266286.10.1080/1750984X.2021.1966822CrossRefGoogle Scholar
Munthe-Kaas, HBA, Noyes, . Chapter 7. Assessing methodological strengths and limitations of qualitative studies. Draft version (April 2024). In: Noyes, JHA, ed. Cochrane-Campbell Handbook for Qualitative Evidence Synthesis, Version 1. Cochrane; 2024.Google Scholar
Long, AF, Godfrey, M, Randall, T, Brettle, A, Grant, M. Developing evidence based social care policy and practice. Part 3: feasibility of undertaking systematic reviews in social care. 2002. University of Leeds (Nuffield Institute for Health) and University of Salford (Health Care Practice R&D Unit).Google Scholar
Munthe-Kaas, HM, Glenton, C, Booth, A, Noyes, J, Lewin, . Systematic mapping of existing tools to appraise methodological strengths and limitations of qualitative research: first stage in the development of the CAMELOT tool. BMC Med Res Methodol. 2019;19(1):113.10.1186/s12874-019-0728-6CrossRefGoogle ScholarPubMed
Majid, U, Vanstone, . Appraising qualitative research for evidence syntheses: a compendium of quality appraisal tools. Qual Health Res. 2018;28(13):21152131.10.1177/1049732318785358CrossRefGoogle ScholarPubMed
Porritt, K, Gomersall, J, Lockwood, . Study selection and critical appraisal: the steps following the literature search in a systematic review. Am J Nursing. 2014;114(6):4752.10.1097/01.NAJ.0000450430.97383.64CrossRefGoogle Scholar
Aromataris, E, Lockwood, . Decisions, decisions: working towards the best available evidence. JBI Evid Synth. 2020;18(8):16091610.10.11124/JBIES-20-00279CrossRefGoogle ScholarPubMed
Ames, H, Booth, A, Noyes, . Chapter 6: selecting studies and sampling draft version (May 2024). In: Noyes, JHA, ed. Cochrane-Campbell Handbook for Qualitative Evidence Synthesis, Version 1. Cochrane; 2024.Google Scholar
Tong, A, Flemming, K, McInnes, E, Oliver, S, Craig, . Enhancing transparency in reporting the synthesis of qualitative research: ENTREQ. BMC Med Res Methodol. 2012;12:18.10.1186/1471-2288-12-181CrossRefGoogle ScholarPubMed
Aromataris, E, Stern, C, Lockwood, C, et al. JBI series paper 2: tailored evidence synthesis approaches are required to answer diverse questions: a pragmatic evidence synthesis toolkit from JBI. J Clin Epidemiol. 2022;150:196202.10.1016/j.jclinepi.2022.04.006CrossRefGoogle ScholarPubMed
Elo, S, Kyngäs, . The qualitative content analysis process. J Adv Nurs. 2008;62(1):107115.10.1111/j.1365-2648.2007.04569.xCrossRefGoogle ScholarPubMed
Forough, A, Wong, S, Lau, E, et al. Nurse experiences of medication administration to people with swallowing difficulties living in aged care facilities: a systematic review of qualitative evidence. JBI Database System Rev Implement Rep. 2018;16(1):7186. https://doi.org/10.11124/JBISRIR-2016-003334.CrossRefGoogle ScholarPubMed
Lockwood, C, Porrit, K, Munn, Z, et al. Chapter 2: Systematic reviews of qualitative evidence. In: Aromataris, E, Munn, Z, eds. JBI Manual for Evidence Synthesis. JBI; 2020:2270. https://doi.org/10.46658/JBIMES-20-03.Google Scholar
Tuomikoski, A-M, Parisod, H, Lotvonen, S, Valimaki, . Experiences of people with progressive memory disorders participating in non-pharmacological interventions: a qualitative systematic review. JBI Evid Synth. 2022;20(8):18711926. https://doi.org/10.11124/JBIES-21-00199 CrossRefGoogle Scholar
Bayo, P, Alobo, G, Sauve, C, Feyissa, G, Tolu., Mothers’ perceptions of the practice of kangaroo mother care for preterm neonates in sub-Saharan Africa: a systematic review of qualitative evidence. JBI Evid Synthesis. 2022;20(2):297347. https://doi.org/10.11124/JBIES-20-00435.CrossRefGoogle ScholarPubMed
Small, S, Porr, C, Swab, M, Murray, . Experiences and cessation needs of indigenous women who smoke during pregnancy: a systematic review of qualitative evidence. JBI Database Syst Rev Implement Rep. 2018;16(2):385452. https://doi.org/10.11124/JBISRIR-2017-003377.CrossRefGoogle ScholarPubMed
Cramm, H, Godfrey, C, Murphy, S, McKeown, S, Dekel, . Experiences of children growing up with a parent who has military-related post-traumatic stress disorder: a qualitative systematic review. JBI Evid Synth. 2022;20(7):16381740. https://doi.org/10.11124/JBIES-20-00229.CrossRefGoogle ScholarPubMed
McCloskey, R, Keeping-Burke, L, Witherspoon, R, Cook, J, Morris, . Experiences of faculty and staff nurses working with nursing students during clinical placement in residential aged care facilities: a systematic review of qualitative evidence. JBI Evid Synth. 2022;20(5):11761208. https://doi.org/10.11124/JBIES-21-00033.CrossRefGoogle ScholarPubMed
Ramasamy, V, Rillotta, F, Alexander, . Experiences of adults with intellectual disabilities who identify as lesbian, gay, bisexual, or transgender within mainstream community: a systematic review of qualitative studies. JBI Evid Synth. 2021;19(1):59154. https://doi.org/10.11124/JBIES-20-00032.CrossRefGoogle ScholarPubMed
Nixon, A, Leonardi-Bee, J, Wang, H, Chattopadhyay, . Barriers and facilitators to type 2 diabetes management in the Caribbean region: a qualitative systematic review. JBI Evid Synth. 2021;19(5):911965. https://doi.org/10.11124/JBISRIR-D-19-00424.CrossRefGoogle ScholarPubMed
Wang, H, Swain, S, Luo, J, Blake, H, Chattopadhyay, . Barriers and facilitators to physical activity among ethnic Chinese children: a qualitative systematic review. JBI Evid Synth. 2020;18(12):24452511. https://doi.org/10.11124/JBISRIR-D-19-00154.CrossRefGoogle ScholarPubMed
Lockwood, C, Porritt, K, Munn, Z, et al. Chapter 2: Systematic reviews of qualitative evidence. In: Aromataris E, Munn Z, eds. Joanna Briggs Institute Reviewer’s Manual [Internet]. Joanna Briggs Institute; 2017.Google Scholar
Leung, L. Validity, reliability, and generalizability in qualitative research. J Family Med Prim Care. 2015;4(3):324327.10.4103/2249-4863.161306CrossRefGoogle ScholarPubMed
Kisely, S, Kendall, . Critically appraising qualitative research: a guide for clinicians more familiar with quantitative techniques. Austral Psychiatry. 2011;19(4):364367.10.3109/10398562.2011.562508CrossRefGoogle ScholarPubMed
Dixon-Woods, M, Booth, A, Sutton, AJ. Synthesizing qualitative research: a review of published reports. Qual Res. 2007;7(3):375422.10.1177/1468794107078517CrossRefGoogle Scholar
Holopainen, A, Hakulinen, . New parents’ experiences of postpartum depression: a systematic review of qualitative evidence. JBI Database Syst Rev Implement Rep. 2019;17(9):17311769. https://doi.org/10.11124/JBISRIR-2017-003909.CrossRefGoogle ScholarPubMed
Small, S, de Boer, C, Swab, . Barriers to and facilitators of labor market engagement for individuals with chronic physical illnesses in their experiences with work disability policy: a qualitative systematic review. JBI Evid Synth. 2022;20(2):348536. https://doi.org/10.11124/JBIES-20-00482.CrossRefGoogle ScholarPubMed
Salmond, E, Salmond, S, Ames, M, Kamienski, M, Holly, . Experiences of compassion fatigue in direct care nurses: a qualitative systematic review. JBI Database Syst. Rev Implement Rep. 2019;17(5):682753. https://doi.org/10.11124/JBISRIR-2017-003818.CrossRefGoogle ScholarPubMed
Koto, Y, Ueki, S, Yamakawa, M, Sakai, . Experiences of patients with lysosomal storage disorders who are receiving enzyme-replacement therapy and the experiences of their family members: a qualitative systematic review. JBI Evid Synth. 2022;20(6):14741510. https://doi.org/10.11124/JBIES-21-00074.CrossRefGoogle ScholarPubMed
Kassam, S, Butcher, D, Marcellus, . Experiences of nurses caring for involuntary migrant maternal women: a qualitative systematic review. JBI Evid Synth. 2022;20(11):26092655. https://doi.org/10.11124/JBIES-21-00181.CrossRefGoogle ScholarPubMed
O’Brien, BC, Harris, IB, Beckman, TJ, Reed, DA, Cook, DA. Standards for reporting qualitative research: a synthesis of recommendations. Acad Med. 2014;89(9):12451251.10.1097/ACM.0000000000000388CrossRefGoogle ScholarPubMed
Tong, A, Sainsbury, P, Craig, . Consolidated criteria for reporting qualitative research (COREQ): a 32-item checklist for interviews and focus groups. Int J Qual Health Care. 2007;19(6):349357.10.1093/intqhc/mzm042CrossRefGoogle ScholarPubMed
Barnes, S, Jordan, Z, Broom, . Health professionals’ experiences of grief associated with the death of pediatric patients: a systematic review. JBI Evid Synth. 2020;18(3):459515. https://doi.org/10.11124/JBISRIR-D-19-00156.CrossRefGoogle ScholarPubMed
Macdonald, M, Martin-Misener, R, Weeks, L, Helwig, M, Moody, E, MacLean, . Experiences and perceptions of spousal/partner caregivers providing care for community-dwelling adults with dementia: a qualitative systematic review. JBI Evid Synth. 2020;18(4):647703. https://doi.org/10.11124/JBISRIR-2017-003774.Google ScholarPubMed
Salmond, E, Ames, M, Kamienski, M, Watkins, AV, Holly, . Experiences of compassion fatigue in direct care nurses: a qualitative systematic review protocol. JBI Evid. Synth. 2017;15(7):18051811.Google ScholarPubMed
Min, M, Hancock, D, Aromataris, E, Crotti, T, Boros, . Experiences of living with juvenile idiopathic arthritis: a qualitative systematic review. JBI Evid Synth. 2022;20(1):60120. https://doi.org/10.11124/JBIES-21-00139.CrossRefGoogle ScholarPubMed
Pike, A, Dobbin-Williams, K, Swab, . Experiences of adults living with an implantable cardioverter defibrillator for cardiovascular disease: a systematic review of qualitative evidence. JBI Evid Synth. 2020;18(11):22312301. https://doi.org/10.11124/JBISRIR-D-19-00239.CrossRefGoogle ScholarPubMed
Tanywe, A, Green, H, Fernandez, . Perceptions and practices of community members relating to trachoma in Africa: a qualitative systematic review. JBI Evid Synth. 2022;20(10):24452474. https://doi.org/10.11124/JBIES-21-00401.CrossRefGoogle Scholar
Carrier, J, Edwards, D, Harden, . Men’s perceptions of the impact of the physical consequences of a radical prostatectomy on their quality of life: a qualitative systematic review. JBI Database Syst Rev Implement Rep. 2018;16(4):892972. https://doi.org/10.11124/JBISRIR-2017-003566.CrossRefGoogle ScholarPubMed
Lloyd, A, Bannigan, K, Sugavanam, T, Freeman, . Experiences of stroke survivors, their families and unpaid carers in goal setting within stroke rehabilitation: a systematic review of qualitative evidence. JBI Database Syst Rev Implement Rep. 2018;16(6):14181453. https://doi.org/10.11124/JBISRIR-2017-003499.CrossRefGoogle ScholarPubMed
Parsons, K, Gaudine, A, Patrick, L, Busby, . Nurse leaders’ experiences of upwards violence in the workplace: a qualitative systematic review. JBI Evid Synth. 2022;20(5):12431274. https://doi.org/10.11124/JBIES-21-00078.CrossRefGoogle ScholarPubMed
Parsons, K, Gaudine, A, Swab, . Experiences of older adults accessing specialized health care services in rural and remote areas: a qualitative systematic review. JBI Evidence Synthesis. 2021. https://doi.org/10.11124/JBIES-20-00048.CrossRefGoogle Scholar
McCloskey, R, Furlong, K, Hansen, . Patient, family and nurse experiences with patient presence during handovers in acute care hospital settings: a systematic review of qualitative evidence. JBI Database Syst Rev Implement Rep. 2019;17(5):754792. https://doi.org/10.11124/JBISRIR-2017-003737.CrossRefGoogle ScholarPubMed
Didier, A, Dzemaili, S, Perrenoud, B, et al. Patients’ perspectives on interprofessional collaboration between health care professionals during hospitalization: a qualitative systematic review. JBI Evid Synth. 2020;18(6):12081270. https://doi.org/10.11124/JBISRIR-D-19-00121.CrossRefGoogle ScholarPubMed
Keeping-Burke, L, McCloskey, R, Donovan, C, Yetman, L, Goudreau, . Nursing students’ experiences with clinical placement in residential aged care facilities: a systematic review of qualitative evidence. JBI Evid Synth. 2020;18(5):9861018. https://doi.org/10.11124/JBISRIR-D-19-00122.CrossRefGoogle ScholarPubMed
Schwandt, TA. Farewell to criteriology. Qual Inq. 1996;2(1):5872.10.1177/107780049600200109CrossRefGoogle Scholar
Jessen-Winge, C, Petersen, M, Morville, A-L. The influence of occupation on wellbeing, as experienced by the elderly: a systematic review. JBI Database Syst Rev Implement Rep. 2018;16(5):11741189. https://doi.org/10.11124/JBISRIR-2016-003123.CrossRefGoogle ScholarPubMed
Devane, D, Burke, NN, Treweek, S, et al. Study within a review (SWAR). J Evid Based Med. 2022;15(4):328 10.1111/jebm.12505CrossRefGoogle ScholarPubMed
Figure 0

Table 1 Criteria for assessing the quality of qualitative research.14

Figure 1

Figure 1 The number of qualitative reviews published by country and year.

Figure 2

Table 2 Methods regarding the inclusion or exclusion of studies based on methodological quality.

Figure 3

Figure 2 The number of reviews specified a priori methods in the protocol over time.

Figure 4

Figure 3 Methods reported in review sections and in results sections.

Figure 5

Table 3 Justifications for including all studies regardless of methodology quality.

Figure 6

Table 4 Exclusion methods.

Figure 7

Table 5 Frequency of items required in the checklist to determine inclusion/exclusion.

Figure 8

Table 6 Justification for certain items in the appraisal checklist deemed essential.

Figure 9

Table 7 Cutoff score details per review.

Figure 10

Table 8 Multiple methods reported per review.

Figure 11

Table 9 The number and percentage of high-, moderate-, low-, and very low-confidence synthesized findings for each method identified.

Supplementary material: File

Jia and Stern supplementary material

Jia and Stern supplementary material
Download Jia and Stern supplementary material(File)
File 60.6 KB