Hostname: page-component-77f85d65b8-pztms Total loading time: 0 Render date: 2026-03-29T21:49:30.278Z Has data issue: false hasContentIssue false

Synthesis challenges in complex evidence: A critical analysis of systematic reviews of face mask efficacy

Published online by Cambridge University Press:  06 February 2026

Trisha Greenhalgh*
Affiliation:
Nuffield Department of Primary Care Health Sciences, University of Oxford, UK
Sahanika Ratnayake
Affiliation:
Philosophy, University of Manchester – The Victoria University of Manchester Campus, UK
Rebecca Helm
Affiliation:
Law, University of Exeter, UK
Luana Poliseli
Affiliation:
Philosophy, University of Manchester – The Victoria University of Manchester Campus, UK
Jon Williamson
Affiliation:
Philosophy, University of Manchester – The Victoria University of Manchester Campus, UK
*
Corresponding author: Trisha Greenhalgh; Email: trish.greenhalgh@phc.ox.ac.uk
Rights & Permissions [Opens in a new window]

Abstract

The evaluation of the role of face masks in preventing respiratory infections is a paradigm case in synthesising complex evidence (i.e. extensive, diverse, technically specialised, and with multilevel chains of causality). Primary studies have assessed different mask types, diseases, populations, and settings using different research designs. Numerous review teams have attempted to synthesise this literature, in which observational (case–control, cohort, cross-sectional) and ecological studies predominate. Their findings and conclusions vary widely.

This article critically examines how 66 systematic reviews dealt with mask efficacy studies. Risk-of-bias tools produced unreliable assessments when—as was often the case—review teams lacked methodological expertise or topic-specific understanding. This was especially true when datasets were large and heterogeneous, with multiple biases playing out in different ways and requiring nuanced adjustments. In such circumstances, tools were sometimes used crudely and reductively rather than to support close reading of primary studies and guide expert judgments. Various moves by reviewers—excluding observational evidence altogether, assessing risk but not direction of biases, omitting distinguishing details of primary studies, and producing meta-analyses that combined studies of different designs or included studies at critical risk of bias—served to obscure important aspects of heterogeneity, resulting in bland and unhelpful summary statements.

We draw on philosophy to question the formulaic use of generic risk-of-bias tools, especially when the primary evidence demands expert understanding and tailoring of study quality questions to the topic. We call for more rigorous training and oversight of reviewers of complex evidence and for new review methods designed specifically for such evidence.

Information

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Open Practices
Open data
Copyright
© The Author(s), 2026. Published by Cambridge University Press on behalf of The Society for Research Synthesis Methodology
Figure 0

Table 1 Tools used for assessing quality in observational studies

Figure 1

Figure 1 How 66 systematic reviews addressed the quality of primary observational studies of mask efficacy.

Figure 2

Table 2 Scores awarded by systematic reviews to the Doung-Ngern study using robins-I

Supplementary material: File

Greenhalgh et al. supplementary material 1

Greenhalgh et al. supplementary material
Download Greenhalgh et al. supplementary material 1(File)
File 138.6 KB
Supplementary material: File

Greenhalgh et al. supplementary material 2

Greenhalgh et al. supplementary material
Download Greenhalgh et al. supplementary material 2(File)
File 175.2 KB