Hostname: page-component-5f7774ffb-cpcbk Total loading time: 0 Render date: 2026-02-20T13:11:56.903Z Has data issue: false hasContentIssue false

Authorship network bias in meta-analysis

Published online by Cambridge University Press:  18 February 2026

Marvin Rieck
Affiliation:
Department of Biometry and Environmental System Analysis, University of Freiburg , Germany
Anne-Christine Mupepele
Affiliation:
Animal Ecology, Department of Biology, University of Marburg , Germany
Carsten F. Dormann*
Affiliation:
Department of Biometry and Environmental System Analysis, University of Freiburg , Germany
*
Corresponding author: Carsten F. Dormann; Email: carsten.dormann@biom.uni-freiburg.de
Rights & Permissions [Opens in a new window]

Abstract

1. Meta-analyses are a reliable method for a quantitative research synthesis. They are, however, prone to specific biases that can be introduced in the process. Such a bias could exist if primary literature produces similar results if coming from the same authors. Authorship network bias is the non-independence of effect sizes introduced by the overlap of authors of primary studies. If not accounted for, it can severely impact the quality of meta-analysis and the conclusions drawn from it.

2. To account for such non-independence, multilevel models with author clusters as an additional hierarchy level were recently suggested. We propose a new method for the detection of non-independent effect sizes based on authorship networks and for their correction.

3. An analysis of simulated data demonstrates the effectiveness of the here-suggested new method. We further applied our new method to nine exemplary meta-analyses.

4. Our new method for detection and effective correction can be easily integrated in existing meta-analysis workflows, using the functionality already offered by R’s metafor package.

5. Our goal is to enhance the reliability of meta-analyses by highlighting potential authorship network bias and offering a method to address this often-overlooked bias.

Information

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2026. Published by Cambridge University Press on behalf of The Society for Research Synthesis Methodology

Highlights

What is already known?

Authorship network bias is one of many specific biases that can be introduced when conducting a meta-analysis. Multilevel modelling has been proposed as a method for dealing with authorship network bias.

What is new?

We propose a method for the correction of authorship network bias that circumvents the main problems that the multilevel approach faces. Instead of adding an additional hierarchy level in the model, we use pairwise distances of papers from an authorship overlap network to account for the similarity of effect sizes.

Potential impact for RSM readers

The presented method is suitable for any meta-analysis dealing with quantitative outcome measures. It can be readily used by researchers within the framework of R’s metafor package.

1 Introduction

Research syntheses play an important role in scientific knowledge gain. They synthesise primary literature and provide the best evidence, according to evidence-based practice,Reference Higgins, Thomas and Chandler 1 and allow for a generalisation of findings in comparison to primary studies.Reference Borenstein, Hedges, Higgins and Rothstein 2 , Reference Koricheva, Gurevitch and Mengersen 3 Research syntheses need to be based on systematic review standards and are at best based on a quantitative synthesis and include a formal meta-analysis. To achieve a high standard, the primary studies contributing to a meta-analysis must be based on reasonable and objective inclusion criteria and not rely on the meta-analyst’s preferences, ideas, or opinions. Inadequate selection of inclusion criteria and researcher-dependent choices can lead to biased results.Reference Shapiro 4

Even with carefully chosen and unproblematic selection criteria, there are other potentially influential biases specific to meta-analysis, such as indexing bias, search bias, reference bias, and publication bias.Reference Felson 5 To prevent bias from distorting results, methodological approaches have been developed to support the detection of biases, e.g., funnel plots to detect a publication bias.Reference Borenstein, Hedges, Higgins and Rothstein 2

Bias can also be introduced by prior beliefs of researchers that shape every step in research from finding a topic of interest to formulating hypotheses, collecting data, and conducting experiments, the analysis or how results are finally interpreted and communicated. In a crowd-sourced analysis, differences were found between researchers due to quantitatively correct yet subjective methodological choices.Reference Silberzahn, Uhlmann and Martin 6 Similarly, Singh et al. (2013)Reference Singh, Martin and Seena 7 and Munder et al. (2012)Reference Munder, Flückiger, Gerger, Wampold and Barth 8 found significant authorship effects on the outcome of therapy assessments. While other factors, such as conflicts of interest, may play a role, some of this authorship effect can likely be attributed to what the analysts believe to know about the topic beforehand. If such knowledge is correct, it would be labelled as ‘expertise’, otherwise as ‘prejudice’. Detaching oneself from all prior knowledge or belief is undesirable as it would ignore our state of knowledge, but if prior belief leads to a loss of variation within researcher groups, relative to the variation between research groups, it should be accounted for when synthesising data.

In meta-analysis, data (i.e., effect sizes estimated from primary studies) are commonly analysed based on the assumption that data points are independent.Reference Borenstein, Hedges, Higgins and Rothstein 2 , Reference Stevens and Taylor 9 Reference Cheung Mike 11 There are however dependencies based on, e.g., geographic or phylogenetic proximity. If these dependencies are ignored, they can strongly influence the conclusion of meta-analyses.Reference Chamberlain Scott, Hovick Stephen and Dibble Christopher 12 It is, by now, common practice to account for non-independence, e.g., by taking into account phylogenetic proximity when analysing effects across multiple species.Reference Adams 13 Reference Mazel, Davies, Georges, Lavergne, Thuiller and Peres-Neto 15

Non-independence of data can also occur due to methodological differences, different research traditions and schools of thought, influence of heads of labs, supervisors, or eminent scientists and they could all lead to biased results in meta-analyses.Reference Hedges, Elizabeth and Johnson Matthew 10 , Reference Shin 16 Reference Wang, Parpia and Couban 19 There is an increasing awareness of this potential source of bias. Proposed remedies include unspecific suggestions, such as ‘pay[ing] more attention to the issue’,Reference Büchter and Pieper 20 assessing the influence of potentially biased single authors on a study outcome,Reference Singh, Martin and Seena 7 or estimating author(team)-specific outcome measures and comparing them to the overall meta-analysis result.Reference Abou-Setta, Rasheda and Lix 21 These approaches do not take into account a potential bias spillover between papers and authorship network structures. If two papers share one or more authors, they have an authorship overlap and can be considered connected. Integrating an authorship overlap in meta-analysis goes further than previous approaches in which data dependencies have been considered only if they were strictly the same, e.g., the same author, same lab, without considering partial overlap.Reference Shin 16 Hierarchical dependence models have been used to account for data dependencies in general,Reference Stevens and Taylor 9 , Reference Hedges, Elizabeth and Johnson Matthew 10 , Reference Konstantopoulos 22 and such models also have the potential to be used for authorship network bias.Reference Moulin and Amaral 17 This approach is easily incorporated into the existing framework of multilevel models and the interpretation of results is relatively straightforward. It does, however, have some limitations in regard to sample sizes in the higher hierarchy levels and authorship network structure. Here, we present an approach that uses hierarchical models to consider non-independence in authorship as gradual rather than clustered, providing a more versatile and realistic representation of authorship networks.

The goal of this research is to establish a new method to detect and account for an authorship network bias. We first introduce the general concepts behind an authorship overlap and explain the new method involving three steps: 1. detecting potential authorship bias; 2. correcting for the authorship bias; and 3. verifying whether the bias correction was successful. Next, we use simulations of authorship networks to demonstrate that the proposed approach works. At last, we apply the approach to previously published meta-analyses. We also consider potential further developments of the method, discuss its limitations, and compare it with other existing methods as well as providing advice on when to use our suggested approach. Appendix C of the Supplementary Material provides a code-based step-by-step tutorial of how to reanalyse one of the previously published meta-analyses: Gibson et al. (2011).Reference Gibson, Lee and Koh 23

2 Methods

2.1 Authorship networks

Authorship overlap can be represented by networks as well as by a similarity matrix, where 0 indicates no overlap (= no link). Authorship bias correction is useful only under two conditions: first, that there is an authorship network with at least some links and second, that there is variance in effect sizes. The potential for authorship network bias gains relevance with increasing controversy (i.e., variance in effect sizes) about the topic to be analysed, because it is this variance of effect sizes that this method aims to help explain. If experimental or observational consensus is high and results are largely similar, then there is little variance to explain by authorship overlap of studies in the first place, and hence, the idea of more closely related papers leading to more similar results is somewhat obsolete. A more detailed look at which parameters are especially influential is given in Section 3.1.

2.2 Detection of authorship network bias

For the detection of potential authorship network bias, we calculate the distance between primary studies based on authorship overlap. In principle, any kind of distance measure could be used. Here, we use two conceptually complementary measures of paper authorship separation, i.e., geodesic distance (the length of the shortest path between two studies in the authorship network; Figure 1) and Jaccard distance (calculated by subtracting Jaccard similarity from one). Conceptually, geodesic distance reflects the structure of the authorship network well, but does not differentiate between small and large proportions of shared authors. To reflect the proportion of shared authors, Jaccard distance can be used additionally. Because the Jaccard distance alone cannot cover network structures, it complements rather than replaces geodesic distance. A drop of effect size similarity with increasing distance of papers indicates authorship network bias (Figure 2).

Figure 1 Example of simulated authorship network as analysed in Section 3.1. See Table 1 for input and output parameters and their explanation.

Figure 2 Correlograms of simulated network with authorship network bias. The left panel shows raw and adjusted effect sizes (ESs) for the geodesic distance model. Because of too few data, no confidence interval could be calculated at geodesic distance $= 6$ . The right panel shows raw and adjusted ESs for the Jaccard model. Note that Jaccard distance $\neq 1$ can only occur for paper pairs with geodesic distance 1. Jaccard distance was calculated by subtracting Jaccard similarity from one. Note that the CI width is proportional to the number of pairwise distances in the respective distance bin. Wide CIs are less informative of the correction outcome but are also less important for assessment because of their low number.

Effect size similarity is calculated by computing all pairwise differences between two effect sizes y in the meta-analysis. This results in an $n \times n$ matrix of effect size differences, $\Delta _{\text {ES}}$ , which we then range between 0 and 1. For a matrix of effect similarities, $S_{\text {ES}}$ , rather than differences, this value is subtracted from 1:

(1) $$ \begin{align} \Delta_{\text{ES}}(i,j) &= |y_i-y_j| \nonumber\\ S_{\text{ES}}(i,j) &= 1 - \frac{\Delta_{ij}}{\max(\Delta)}. \end{align} $$

Plotting effect size similarity against Jaccard or geodesic distance, which is reflecting the authorship overlap, results in a correlogram (Figure 2). A horizontal line from geodesic distance one to infinity would mean that authorship network distances do not influence similarities between effect sizes. Decreasing effect size similarity with higher geodesic distance values, and/or a high difference of effect sizes between finite and infinite geodesic distances, indicate that effect sizes are influenced by authorship overlap.

2.3 Correcting for authorship network bias

If authorship network bias was detected, the ensuing analysis is equivalent to a phylogenetic meta-analysis.Reference Chamberlain Scott, Hovick Stephen and Dibble Christopher 12 Instead of a phylogeny, we construct a similarity matrix $\Sigma _{\text {AN}}$ , based on (normalised) pairwise geodesic distances $\Delta _{ij}$ between two studies i and j. Next, we compute the effect-size similarity among unconnected studies in the authorship network, i.e., the similarity at distance infinity or authorship network weighting factor $S_\infty ~=~~1-~\hat {\Delta }_{\text {between cluster}}$ . This is then used to construct a weighted authorship network similarity matrix:

(2) $$ \begin{align} W_{\text{AN}}(i,j) = 1 - S_\infty \cdot \frac{\Delta_{ij}}{\max(\Delta)}. \end{align} $$

The value of $ S_\infty $ determines the influence of the geodesic distance matrix: if even independent studies differ little in their effect sizes, then $ S_\infty $ is close to one and authorship overlap bias ( $\frac {\Delta _{ij}}{\max (\Delta )}$ ) cannot have a strong influence on the results. If, however, differences are substantial, then $S_\infty $ is low, and the potential effect of authorship overlap is high, resulting in more similar results among studies of an authorship cluster and an overall wider range of possible values in the similarity matrix $W_{\text {AN}}$ .

The model used to estimate adjusted effect sizes while accounting for an authorship network bias is based on a Generalised Least Squares approach, where the similarity matrix $\Sigma _{\text {AN}}$ is provided as variance–covariance matrix.Reference Ives Anthony and Helmus Matthew 14 In the context of meta-analysis, it is internally multiplied by the study-specific variances, a procedure implemented in the metafor package for R.Reference Viechtbauer 24 The constructed similarity matrix can be used alongside other commonly used specifications of meta-analytical models, such as nesting multiple effect sizes within studies or multiple studies within regions. Even multiple similarity matrices may be used together to account for authorship bias and the phylogenetic relationships between different species.Reference Rios Moura, Oliveira Gonzaga, Silva Pinto, Vasconcellos-Neto and Requena 25 To reduce type I error rates while keeping constant the chance of committing a type II error, the similarity matrix $\Sigma _{\text {AN}}$ should be included in the model even if its contribution to explaining the variance of the fitted model is small. Cinar et al. (2021)Reference Cinar, Nakagawa and Viechtbauer 26 recommend a similar approach to account for dependent effect sizes of multiple species in phylogenetic meta-analyses. If similarity matrices were included in models in which they did not explain any variance at all, there was no effect on either the reported estimates or their confidence intervals (see ?? for the corresponding code). Thus, in case of doubt, the similarity matrix can simply be included in the analysis, as it presents no disadvantage.

To assess the degree to which the similarity matrix $\Sigma _{\text {AN}}$ mitigates authorship network bias, the adjusted effect sizes can be represented by a post-correction correlogram. To do so, the best linear unbiased predictor (BLUP) of the similarity matrix is subtracted from the raw effect sizes. Now, a correlogram can be created to display the adjusted effect sizes in relation to the geodesic distance and to compare it with the raw effect sizes. Ideally, the resulting line should now be horizontal and the effect sizes similar at finite and infinite geodesic distances. An example of an adjusted effect sizes correlogram is shown in Figure 2. See Section 3.1 for detailed analysis.

The geodesic and Jaccard similarity matrices (equation (1)) can now be included in the meta-analytical model. To take into account both the network structure and the number of shared authors as edge weights, the two matrices can be summed and their values normalised. A unique study identifier column is included as a random effect to which the similarity matrix is assigned. The model summary now gives an estimate corrected for similarity of authorship. Comparing this estimate to the estimate from the model without bias correction gives an idea of how influential the matrix is. More important, however, is that we check a new correlogram of adjusted effect sizes for the pattern observed before.

3 Application of the method

3.1 Simulation studies

3.1.1 Simulation of authorship networks

Before applying the method outlined above to real-world meta-analyses, we used simulations to explore the potential bias of an authorship network.

We simulated networks with varying effect-size differences (which determine how much bias there could be) and network connectedness (which sets the degree of non-independence). Additionally, other variables, e.g., number of papers and bias size, were included in the simulation of networks as they potentially influence authorship network bias (Table 1). Simulation aimed at identifying single as well as combinations of variables that may have a particular influence on the authorship bias correction approach (see Section 2 and ?? for the complete code for network simulation; see also Appendix A of the Supplementary Material for how data were simulated conceptually).

Table 1 Parameters used for network simulation, as well as the respective value for the analysed exemplary network

Note: Simulation range gives the values that were used in exploring the parameter’s influence on authorship network bias. Observation ranges are the values found in the published meta-analyses that were re-analysed. For cluster density, the mean cluster densities of the respective networks are given. Note that some parameters were only used to construct networks and not extracted from existing ones.

Authorship networks with the potential to bias the summary effect size are characterised by three main properties: 1. large effect size differences between studies or clusters of studies (potential for large bias); 2. high connectedness: a large proportion of studies have at least some overlap with other studies, enabling potential bias spillover; and 3. a pronounced clustering in the network containing bias spillover in separate clusters, making cluster differences more pronounced than they would be in an evenly connected network.

For our simulation, we tested low to high bias scenarios as well as low to high inter- as well as intra-cluster similarity of effect sizes. Intra-cluster effect size similarity was set by $\sigma _{\text {cluster}}$ , homophily, and cluster density, while inter-cluster effect size similarity was mainly controlled by bias size. After an initial exploration phase, bias origin and bias type were kept at the example values for all further analyses because they best reflect real authorship networks.

Network connectedness is mainly driven by how many papers show no overlap (no. of isolates), the number of clusters, and the connectedness within clusters (cluster density). Cluster size evenness was set to 0.9 for most simulations. This is, with some exceptions, reflective of the meta-analyses we reanalysed.

The degree of clustering depends to some extent on the above parameters but was mainly controlled by the number of clusters, their size, and respective density. Cluster density in real meta-analyses was mostly substantially higher than our simulation values, potentially exacerbating bias risk.

$\sigma _{\text {isolates}}$ was mostly set to lower values than $\sigma _{\text {cluster}}$ to imitate “natural” variability of outcomes from which $\sigma _{\text {cluster}}$ deviates. Network size (number of papers) was set to 400 for sensitivity analysis (Figure 4) which is above network sizes in most real meta-analyses. Prior tests showed a non-significant effect of network size on the methods efficacy.

To assess the likelihood of encountering bias in real authorship networks, three metrics should be assessed: effect size differences $\Delta _{\text {ES}}$ (see the equations in Section 2.2). In the following, we will also use connectedness, calculated as in Krackhardt (1994)Reference Krackhardt, Carley and Prietula 27 and, as a measure of how clustered a network is, the global clustering coefficient (GCC) of Wasserman and Faust (1994).Reference Wasserman and Faust 28 A high GCC means that cluster members are well-connected, facilitating bias spread within clusters. Conversely, in sparsely connected clusters, bias transfer is less likely. However, interpreting GCC requires context: it excludes isolates and two-paper clusters, so a meaningful interpretation can only be achieved if also considering information on the general network structure such as connectedness.

3.1.2 Detection and correction of authorship network bias in simulated networks.

Central to detection and accounting for an authorship network bias is the calculation of the authorship similarity matrix. To do that, we first create a correlogram and extract the mean effect size similarity for the most dissimilar node pairs. In the case of geodesics, this would be at distance infinity; for Jaccard distance, it would be at one. Figure 2 shows the correlograms for geodesic and for Jaccard distances, depicting some within-cluster correlation between geodesic distance and effect size similarity for the raw effect sizes. Importantly, there is a substantial difference between the similarity values of finite and infinite geodesic distances. The latter is also true for the Jaccard correlogram, although no pattern is visible for the within-cluster correlation. Note that the within-cluster correlation for Jaccard applies only to direct neighbours in the network, so patterns in the Jaccard correlogram describe the relationship between the proportion of authorship overlap and effect size similarity, not taking into account indirect neighbours and the general network structure.

A less clear pattern is observable in the geodesic distance plot for adjusted effect sizes and not much change for Jaccard. For both correlograms, however, there is much less difference between unconnected than connected paper pairs. That is, the line from finite to infinite geodesic distance is rather horizontal after authorship similarity correction, meaning the effect size difference is no longer a function of paper similarity.

Because the correlogram relies on averaged effect size similarities for paper distance bins, it may, in some cases, miss small deviant clusters. To assess individual clusters, a forest plot with papers ordered and coloured by authorship cluster membership (Figure 3) can be helpful in identifying such deviants. Figure 3 shows such a forest plot. It can serve as a diagnostic tool as well as a typical visualisation of meta-analytical results showing the influence of the authorship bias at the level of a subset of individual studies, as well as the overall effect. Note that Figures 2 and 3 relate to the example values of Table 1. See Figure 4 for a more comprehensive overview of how the method performs over a wider range of scenarios.

Figure 3 Forest plots of the simulated network analysed in Section 3.1. Top: Studies are ordered and coloured according to their cluster membership to identify deviant clusters. Bottom: Only 20 out of 400 total studies are shown. Grey-shaded diamonds signify effect sizes that were adjusted for the influence of paper proximity in the network. Black diamonds show model results of the models without and with (“adjusted effect”) authorship similarity matrix.

As bias and connectedness increase, the method gets more reliable in correcting biased results (Figure 4). This explains the relatively low success rate at low bias–low connectedness panels, because it is less likely that the main assumption of the method is met. With low connectedness, there is less chance of bias propagation and significant effects are likely to be a result of random generation of data. If, by chance, such a network shows a significant effect, taking into account the authorship overlap is not expected to make a difference. With increasing connectedness both bias propagation and the possibility to correct for it increase as well. The latter is also true for larger bias. If a few clusters show large bias that lead to significant results, this large bias can be accounted for by the proximity of cluster members. For low bias networks, accounting for the proximity of cluster members does not make a difference, because all effect sizes are similar. So the success rate is linked to how likely authorship bias occurrence is for the depicted networks.

3.2 Published meta-analyses

Nine published meta-analyses were re-analysed to assess whether authorship network bias occurred and what its influence on the overall effect was. Four of them were used for the multilevel model approach of Moulin and Amaral (2020).Reference Moulin and Amaral 17 The meta-analyses covered the scientific fields of psychology, biology, ecology, and medicine. They ranged from 10 to 457 primary studies. Data were taken from the respective meta-analyses or, where applicable, from the metafor package.

Figure 4 Influence of bias and connectedness of networks on the proportion of successfully corrected biased results. Proportions refer to corrected effect sizes out of 50 simulated networks for each combination. Proportions of success are the proportions of non-significant results for the full model, where the standard model resulted in a significant effect.

3.2.1 Detection of authorship network bias.

Most constructed networks showed substantial clustering (GCCs from 0.66 to 0.86) but extremely low overall connectedness (Table 2). The meta-analysis of Chen et al. (2014)Reference Chen, Hung and Tsai 29 shows a GCC of zero, because no closed triplets are present in the network. Figure 5 illustrates exemplary sets of diagnostic plots: the resulting networks, correlograms, and forest plots for Bakdash et al. (2021)Reference Bakdash, Marusich, Cox, Geuss, Zaroukian and Morris 30 and Dinu et al. (2017).Reference Dinu, Abbate, Gensini, Casini and Sofi 31 Both networks show small and relatively evenly distributed clusters. In the correlograms, geodesic distance shows an effect on effect size similarity within clusters, albeit weak, and no effect at infinite distance. Jaccard distance seems to have no effect. Diagnostic plots for all analysed meta-analyses can be found in Appendix B of the Supplementary Material.

Table 2 Network characteristics and effect sizes before and after correction for the reanalysed meta-analyses

Figure 5 Diagnostic plots of (a) Bakdash et al. (2021),Reference Bakdash, Marusich, Cox, Geuss, Zaroukian and Morris 30 and (c) Dinu et al. (2017).Reference Dinu, Abbate, Gensini, Casini and Sofi 31 See Appendix B of the Supplementary Material for other meta-analyses.

Taking into account the network metrics and the correlogram patterns, bias risk for the presented meta-analyses can be said to be low.

3.2.2 Correction of authorship network bias.

Correction for the authorship network led to adjustment of effect sizes between 0 and 0.25 in the reanalysed studies (Table 2). In no case did the inclusion of the similarity matrix change the reported effect size enough to lead to a different conclusion (e.g., p-values remained on the same side of the significance thresholds). More important than a change in p-values, however, are changes in effect sizes.Reference Borenstein, Hedges, Higgins and Rothstein 2 In our cases, those were also small for all but one moderator in Besson et al. (2015).Reference Besson, Malgorzata, Senior, Hector and Shinichi 32 The comparison of our proposed method to the multilevel approach used by Moulin and Amaral (2020)Reference Moulin and Amaral 17 showed similar results for Chen et al. (2014),Reference Chen, Hung and Tsai 29 Kredlow et al. (2016),Reference Kredlow, Unger and Otto 33 Mathie et al. (2017),Reference Mathie, Nitish and Legg Lynn 34 and Munkholm et al. (2015).Reference Munkholm, Vinberg and Kessing 35

The resulting forest plots show some considerable differences in single effect sizes for a random sample of studies within the respective meta-analysis, but diamonds at the bottom show no difference in the overall estimate (a) or moderators (c).

4 Discussion

The main parameters determining the likelihood of encountering authorship network bias are the effect size differences of primary studies and the connectedness of the authorship network. Increases in both are required to lead to increased bias risk. The method presented effectively mitigates the bias when it occurs. Based on the meta-analyses included here, authorship network bias affecting the overall result of a meta-analysis is probably rare. However, authorship cluster affiliation has been shown to significantly impact reported results.Reference Khouri, El Amine and Arnaud 36 With uneven authorship cluster sizes, this may lead to substantially biased overall results. For this reason and as there are no negative effects on the model’s estimates when the method is applied unnecessarily, it is recommended to apply it for all potentially affected authorship networks. The main cost is the time it takes to correct author names in the paper database, as not all authors are represented with unique names (such as researcher ID or ORCID). Incorrect naming could lead to missing edges where they should be included and to including edges where there should be none. The latter is of more concern, because it might down-weight the influence of whole clusters on the overall result, hence we analysed the effect of varying numbers of spurious edges, e.g., due to homonymous author names in different network scenarios. The effects of single or few spurious edges have proven to be negligible. See Appendix D of the Supplementary Material for the results of these simulations.

4.1 Similarity measures

For the quantification of overlap in authorship, there is no inherent reason to prefer Simpson, Sørensen, or Jaccard over other methods, such as node-correspondence methodsReference Tantardini, Ieva, Tajoli and Piccardi 37 or edit distances.Reference Marzal and Vidal 38 When choosing a similarity measure, or when testing multiple, the resulting correlograms should be used to determine how well they capture the suspected bias in the network. This being said, we recommend using inverse geodesic distance as the standard similarity measure because of its ability to capture network structures exceptionally well. It may be complemented by adding other measures or even be replaced should the analyst find a better fitting alternative but it is a very good starting point in all cases.

For the simulated data, and by design, the model with geodesic similarity was clearly the better fit. In the re-analysed meta-analyses, neither geodesic distance, nor Jaccard, nor their combination explained much variance of effect sizes. Conceptually, geodesic similarity covers the idea of the authorship network bias better. It does seem sensible to also account for the proportion of shared authors, however. So, the heuristic combination of two measures seems a plausible way of analysing data. There are other ways of weighting network edges.Reference Hunter, Handcock, Butts, Goodreau and Morris 39 Reference Krivitsky, Hunter, Morris and Klumb 41 However, for this set of meta-analyses, it seems that the effect of the edge weight does not carry much information. For larger effect size differences, there might be a larger effect and we would still recommend checking whether including weights changes the effect sizes at all.

The similarity matrix can be specified further. It seems plausible that not every author has the same influence on bias propagation. Depending on multiple aspects, authors’ ideas or suggestions might be taken more or less seriously by collaborators. Seniority and other character traits might influence this phenomenon but are somewhat difficult to quantify objectively. Other, more measurable things, such as academic rank or the respective H-index of the authors, could be included as a way of weighting edges in the network. Giving more weight to first, second, or last author (field specific) might also be an option. There are many ways to include other parameters in this method, which can be justified or refuted by assessing their effect on the results. Supplementary Material SC 7 includes a section on further weighting the included similarity matrices by taking into account the number of shared authors or some assigned measure such as H-index.

4.2 Influential parameters

For the range of connectedness values observed in our case studies, problematic levels of authorship network bias are unlikely. But even when effect size differences are small, with high connectedness (and high GCC), the results might still be changed considerably under authorship network bias. What a considerable change means is dependent on the respective outcome measure used. In any case, the importance of the similarity matrix can be assessed by calculating an $I^2$ equivalent for multilevel models, which quantifies how much of the total variance is explained by the similarity matrix.Reference Viechtbauer 42 The bias risk decreases with diminishing effect size differences. Looking at GCC alone does not have much explanatory power for discovering problematic network structures. Only considering connectedness is slightly more helpful, but we recommend their joint interpretation to get an idea of the probability that we encounter authorship network bias.

No single diagnostic measure is sufficient for checking bias patterns. The correlogram is useful but only reflects mean effect size similarity for each distance category, potentially missing outlier clusters. With enough data, this is less problematic, but relying solely on it risks underestimating bias. Another approach is to colour network nodes by reported effect sizes (as in Figure 1) or to colour studies and effect sizes according to authorship cluster membership in a forest plot (see Figure 3 and Supplementary Material SC 2 and SC 7 for code) and, most reliably, to compute the similarity matrix for inclusion in the model.

4.3 Limitations and extensions

How informative the authorship network is, depends on which primary studies are included. This means the method can never fully capture all possible bias leakage, for example, from connections outside the depicted networks, but also due to mentoring, supervision, and funding review. Moulin and Amaral (2020)Reference Moulin and Amaral 17 proposed an extension to their method, also applicable here: with unique author identifiers and lifetime collaboration networks, collaborations beyond the meta-analysis could be incorporated.

Our method can reliably detect and correct bias in high-bias–high-connectedness networks with diminishing success rate as both decrease. It performs worst with little effect size variance within sparse networks. In these cases, the potential for influential spurious edges is also highest. They have, however, been shown to have little to no effect on simulated networks within a realistic range of wrongfully added connections.

4.4 (Dis)advantages relative to other methods

The issue of non-independence of effect sizes or studies and its adverse influence on meta-analyses is not new. Initially, concerns focused on nested effects and hierarchical data structures.Reference Cheung Mike 11 , Reference Konstantopoulos 22 The former study highlighted that hierarchy can also arise purely from meta-analytical data, where effect sizes nest within studies, which in turn nest within researchers or labs. Moulin and Amaral (2020)Reference Moulin and Amaral 17 extended this by adding another clustering level to account for authorship networks.

A different approach was chosen here, due to key advantages of modelling a network over a multilevel structure. Using a similarity matrix instead of assigning clusters better captures relationships within clusters—more distant papers within clusters might exhibit less effect size similarity, a pattern lost in standard multilevel models. Nestedness is also not always clear-cut, whereas a similarity matrix works regardless of cluster structure. Additionally, this method imposes no cluster size restrictions; even two-paper clusters can add explanatory power.

Multilevel models, by contrast, are recommended to have sample sizes of at least 20 data points per cluster per level.Reference Maas Cora and Hox Joop 43 , Reference Bell, Ferron and Kromrey Jeffrey 44 These recommendations would often be ignored, which may lead to substantially biased estimatesReference Maas Cora and Hox Joop 43 , Reference Austin 45 or clustering would be ignored altogether. Bell et al. (2008)Reference Bell, Ferron and Kromrey Jeffrey 44 further show that a high number of isolates inflates confidence intervals and increases type II error risk—an issue exacerbated in authorship networks due to even higher proportions of isolates. With a similarity matrix, confidence intervals only expand when the overall effect changes, avoiding this issue.

In comparison with Moulin and Amaral (2020),Reference Moulin and Amaral 17 visual assessment of networks is more directly linked to potential bias propagation, when vertices represent papers instead of authors. In the latter case, strong clustering might just be a result of one or few papers being authored by many people, not requiring any overlap.

Using both methods on the same published meta-analyses showed no relevant differences in the reported results. Differences are not to be expected in cases with large clusters, high intra-cluster connectivity, and clear separation between clusters. The latter two seem to be commonly encountered in real meta-analyses. For networks with potentially problematic properties for multilevel models, such as small and gradually connected rather than clearly separated clusters, our proposed method may prove advantageous. When there is no clear indication for the use of one of the methods, they may both be applied and results compared.

4.5 When to use this method

Given the relatively minor changes in effect sizes observed in the re-analysed meta-analyses, the proposed method may be perceived as addressing a largely hypothetical concern. Authorship network bias appears to be small and rare; however, it is important to acknowledge that only nine meta-analyses were examined. While its impact on effect sizes seems minimal, the underlying concept of bias propagation remains a plausible concern in real-world meta-analyses. This, along with the potential influence of authorship cluster membership studies’ results,Reference Khouri, El Amine and Arnaud 36 provides sufficient justification for considering this method as a standard measure. Although such bias may not significantly affect most cases’ overall results, every measure to reduce the likelihood of Type I errors should be considered, particularly when it does not increase the risk of Type II errors. Therefore, when a meta-analysis network exhibits authorship overlap, incorporating the authorship network provides a straightforward approach to mitigating the risk of biased results.

This method gains in importance when dealing with clustered networks, in which single or few author clusters account for large proportions of primary studies (e.g., right panels in Figure 1). Assessment of a coloured forest plot (Figure 3) is a helpful tool to assess the potential impact these clusters may have on the result of the meta-analysis and can help in deciding whether correction is needed. However, because of the above points, we recommend routinely implementing this method (or other ways of mitigating authorship bias), even when diagnostic measures indicate low bias risk. Ignoring authorship network bias may lead (and has led) to obscured overall effects due to a single deviant cluster.Reference Zarychanski, Abou-Setta and Turgeon 46

4.6 Implementation

We present a user-friendly approach for identifying and addressing authorship network bias, outlining the fundamental concepts, assumptions, and comprehensive example analyses. Since all steps were performed using existing functions in the R environment, researchers can readily implement our method in their own meta-analyses. It is seamlessly integrated into the analyst’s workflow and does not require any additional change in model structure. It may be used in basic random effects models, hierarchical dependence models, correlated effects models, and the combination thereof,Reference Pustejovsky James and Elizabeth 47 as long as no other approach of dealing with authorship network bias has been implemented. The only modification lies in providing an additional random effect term and a corresponding similarity matrix. A detailed tutorial, along with the complete code for all studies and simulations referenced in this work, is provided in Appendix E of the Supplementary Material.

4.7 Conclusion

Meta-analysis is generally seen as a reliable method to come to broader conclusions and generalise knowledge better than any single study could. As such, meta-analysis is often trusted more than each single primary study included, because of a broader perspective on topics, larger sample sizes, and higher statistical power.Reference Feinstein 48 This trust is only justified if meta-analyses are conducted avoiding the many potential pitfalls of that method. Criticism has been voiced about the general idea of meta-analysis,Reference Shapiro 4 the inevitable complexity, which cannot be steered through without making mistakes,Reference Wely 49 and its wrong use and interpretation of its results.Reference Shapiro 4 While Borenstein et al.Reference Borenstein, Hedges, Higgins and Rothstein 2 agree with these points and highlight the problems that arise from a misuse of meta-analytical methods, they also argue that all of them can be countered with increasing education on proper implementation of methods. One such problem is the introduction of biases specific to meta-analysis. The present work helps in bringing a potentially influential bias to the reader’s attention and proposes a way to account for it. This work, by drawing further attention to the authorship network bias, along with a diagnostic tool and a method to remedy it, aims at further improving the quality of meta-analysis and reducing the likelihood of reporting biased results.

Acknowledgements

We declare that this manuscript is original, has not been published before, and is not currently considered for publication elsewhere.

Author contributions

A.-C.M. and C.F.D. conceived the idea; M.R. developed the methodology and carried out the formal analysis; M.R. prepared the original draft; all authors contributed critically to the review and editing of the manuscript and gave final approval for publication.

Competing interest statement

The authors declare that no competing interests exist.

Data availability statement

Data and Code are available at https://doi.org/10.5281/zenodo.17712269.

Funding statement

C.F.D. acknowledges partial funding by the Deutsche Forschungsgemeinschaft (DFG) through SFB 1597 ‘SmallData’.

Supplementary material

The supplementary material for this article can be found at https://doi.org/10.1017/rsm.2025.10063.

Footnotes

This article was awarded Open Data and Open Materials badges for transparent practices. See the Data availability statement for details.

References

Higgins, JPT, Thomas, J, Chandler, J, et al. Cochrane Handbook for Systematic Reviews of Interventions. Version 6.4 (Updated August 2023). 2nd ed. John Wiley & Sons; 2023.Google Scholar
Borenstein, M, Hedges, LV, Higgins, JPT, Rothstein, HR. Introduction to Meta-Analysis. John Wiley & Sons; 2009.10.1002/9780470743386CrossRefGoogle Scholar
Koricheva, J, Gurevitch, J, Mengersen, KL. Handbook of Meta-Analysis in Ecology and Evolution. Princeton University Press; 2013.Google Scholar
Shapiro, S. Meta-analysis/Shmeta-analysis. Am J Epidemiol. 1994;140:771778.10.1093/oxfordjournals.aje.a117324CrossRefGoogle ScholarPubMed
Felson, DT. Bias in meta-analytic research. J Clin Epidemiol. 1992;45:885892.10.1016/0895-4356(92)90072-UCrossRefGoogle ScholarPubMed
Silberzahn, R, Uhlmann, EL, Martin, DP, et al. Many analysts, one data set: making transparent how variations in analytic choices affect results. Adv Methods Pract Psychol Sci. 2018;1:337356.Google Scholar
Singh, JP, Martin, G, Seena, F. Authorship bias in violence risk assessment? A systematic review and meta-analysis. PLoS One. 2013;8:e72484.10.1371/journal.pone.0072484CrossRefGoogle ScholarPubMed
Munder, T, Flückiger, C, Gerger, H, Wampold, BE, Barth, J. Is the allegiance effect an epiphenomenon of true efficacy differences between treatments? A meta-analysis. J Couns Psychol. 2012;59:631637.10.1037/a0029571CrossRefGoogle Scholar
Stevens, JR, Taylor, AM. Hierarchical dependence in meta-analysis. J Educ Behav Stat. 2009;34:4673.10.3102/1076998607309080CrossRefGoogle Scholar
Hedges, LV, Elizabeth, T, Johnson Matthew, C. Robust variance estimation in meta-regression with dependent effect size estimates. Res Synth Methods. 2010;1:3965.10.1002/jrsm.5CrossRefGoogle ScholarPubMed
Cheung Mike, W-L. A guide to conducting a meta-analysis with non-independent effect sizes. Neuropsychol Rev. 2019;29:387396.10.1007/s11065-019-09415-6CrossRefGoogle ScholarPubMed
Chamberlain Scott, A, Hovick Stephen, M, Dibble Christopher, J, et al. Does phylogeny matter? Assessing the impact of phylogenetic information in ecological meta-analysis. Ecol Lett. 2012;15:627636.10.1111/j.1461-0248.2012.01776.xCrossRefGoogle ScholarPubMed
Adams, DC. Phylogenetic meta-analysis. Evolution. 2008;62:567572.10.1111/j.1558-5646.2007.00314.xCrossRefGoogle ScholarPubMed
Ives Anthony, R, Helmus Matthew, R. Generalized linear mixed models for phylogenetic analyses of community structure. Ecol Monogr. 2011;81:511525.10.1890/10-1264.1CrossRefGoogle Scholar
Mazel, F, Davies, TJ, Georges, D, Lavergne, S, Thuiller, W, Peres-Neto, PR. Improving phylogenetic regression under complex evolutionary models. Ecology. 2016;97:286293.10.1890/15-0086.1CrossRefGoogle ScholarPubMed
Shin, I-S. Same Author and Same Data Dependence in Meta-Analysis. The Florida State University; 2009.Google Scholar
Moulin, TC, Amaral, OB. Using collaboration networks to identify authorship dependence in meta-analysis results. Res Synth Methods. 2020;11:655668.10.1002/jrsm.1430CrossRefGoogle ScholarPubMed
Abbas-Aghababazadeh, F, Xu, W, Haibe-Kains, B. The impact of violating the independence assumption in meta-analysis on biomarker discovery. Front Genet. 2023;13:1027345.10.3389/fgene.2022.1027345CrossRefGoogle ScholarPubMed
Wang, Y, Parpia, S, Couban, R, et al. Compelling evidence from meta-epidemiological studies demonstrates overestimation of effects in randomized trials that fail to optimize randomization and blind patients and outcome assessors. J Clin Epidemiol. 2024;165:111211.10.1016/j.jclinepi.2023.11.001CrossRefGoogle ScholarPubMed
Büchter, RB, Pieper, D. Most overviews of Cochrane reviews neglected potential biases from dual authorship. J Clin Epidemiol. 2016;77:9194.10.1016/j.jclinepi.2016.04.008CrossRefGoogle ScholarPubMed
Abou-Setta, AM, Rasheda, R, Lix, LM, et al. Can authorship bias be detected in meta-analysis? Can J Anesth/J Can Anesth. 2019;66:287292.10.1007/s12630-018-01268-6CrossRefGoogle ScholarPubMed
Konstantopoulos, S. Fixed effects and variance components estimation in three-level meta-analysis: three-level meta-analysis. Res Synth Methods. 2011;2:6176.10.1002/jrsm.35CrossRefGoogle ScholarPubMed
Gibson, L, Lee, TM, Koh, LP, et al. Primary forests are irreplaceable for sustaining tropical biodiversity. Nature. 2011;478:378381.10.1038/nature10425CrossRefGoogle ScholarPubMed
Viechtbauer, W. Conducting meta-analyses in R with the metafor package. J Stat Softw. 2010;36:148.10.18637/jss.v036.i03CrossRefGoogle Scholar
Rios Moura, R, Oliveira Gonzaga, M, Silva Pinto, N, Vasconcellos-Neto, J, Requena, GS. Assortative mating in space and time: patterns and biases. Ecol Lett. 2021;24:10891102.10.1111/ele.13690CrossRefGoogle ScholarPubMed
Cinar, O, Nakagawa, S, Viechtbauer, W. Phylogenetic multilevel meta-analysis: a simulation study on the importance of modelling the phylogeny. Methods Ecol Evol. 2021;13:383395.10.1111/2041-210X.13760CrossRefGoogle Scholar
Krackhardt, D. Graph theoretical dimensions of informal organizations. In: Carley, KM, Prietula, MJ, eds. Computational Organization Theory. Lawrence Erlbaum Associates, Inc.; 1994:89111.Google Scholar
Wasserman, S, Faust, K. Social Network Analysis: Methods and Applications. Cambridge University Press; 1994.10.1017/CBO9780511815478CrossRefGoogle Scholar
Chen, Y-R, Hung, KW, Tsai, JC, et al. Efficacy of eye-movement desensitization and reprocessing for patients with posttraumatic-stress disorder: a meta-analysis of randomized controlled trials. PLoS One. 2014;9:e103676.10.1371/journal.pone.0103676CrossRefGoogle ScholarPubMed
Bakdash, JZ, Marusich, LR, Cox, KR, Geuss, MN, Zaroukian, EG, Morris, KM. The validity of situation awareness for performance: a meta-analysis. Theor Issues Ergon Sci. 2021;23:221244.10.1080/1463922X.2021.1921310CrossRefGoogle Scholar
Dinu, M, Abbate, R, Gensini, GF, Casini, A, Sofi, F. Vegetarian, vegan diets and multiple health outcomes: a systematic review with meta-analysis of observational studies. Crit Rev Food Sci Nutr. 2017;57:36403649.10.1080/10408398.2016.1138447CrossRefGoogle ScholarPubMed
Besson, AA, Malgorzata, L, Senior, AM, Hector, KL, Shinichi, N. Effect of maternal diet on offspring coping styles in rodents: a systematic review and meta-analysis. Biol Rev. 2015;91:10651080.10.1111/brv.12210CrossRefGoogle ScholarPubMed
Kredlow, MA, Unger, LD, Otto, MW. Harnessing reconsolidation to weaken fear and appetitive memories: a meta-analysis of post-retrieval extinction effects. Psychol Bull. 2016;142:314336.10.1037/bul0000034CrossRefGoogle ScholarPubMed
Mathie, RT, Nitish, R, Legg Lynn, A, et al. Randomised, double-blind, placebo-controlled trials of non-individualised homeopathic treatment: systematic review and meta-analysis. Syst Rev. 2017;6(1):63.10.1186/s13643-017-0445-3CrossRefGoogle ScholarPubMed
Munkholm, K, Vinberg, M, Kessing, LV. Peripheral blood brain-derived neurotrophic factor in bipolar disorder: a comprehensive systematic review and meta-analysis. Mol Psychiatry. 2015;21:216228.10.1038/mp.2015.54CrossRefGoogle ScholarPubMed
Khouri, C, El Amine, B, Arnaud, C, et al. Co-authorship group significantly impacts reported arterial blood pressure variations in rodents exposed to intermittent hypoxia: a meta-research study. Cardiovasc Res. 2023;119:e128e130.10.1093/cvr/cvad048CrossRefGoogle ScholarPubMed
Tantardini, M, Ieva, F, Tajoli, L, Piccardi, C. Comparing methods for comparing networks. Sci Rep. 2019;9:17557.10.1038/s41598-019-53708-yCrossRefGoogle ScholarPubMed
Marzal, A, Vidal, E. Computation of normalized edit distance and applications. IEEE Trans Pattern Anal Mach Intell. 1993;15:926932.10.1109/34.232078CrossRefGoogle Scholar
Hunter, DR, Handcock, MS, Butts, CT, Goodreau, SM, Morris, M. ergm: a package to fit, simulate and diagnose exponential-family models for networks. Journal of Statistical Software. 2008;24(3):129.10.18637/jss.v024.i03CrossRefGoogle ScholarPubMed
Handcock, MS, Hunter, DR, Butts, CT, Goodreau, SM, Krivitsky, PN, Morris, M. ergm: fit, simulate and diagnose exponential-family models for networks. 2023. The Statnet Project. R package version 4.5.0. https://statnet.org.Google Scholar
Krivitsky, PN, Hunter, DR, Morris, M, Klumb, C. ergm 4: new features for analyzing exponential-family random graph models. Journal of Statistical Software. 2023;105(6):144.10.18637/jss.v105.i06CrossRefGoogle Scholar
Viechtbauer, W. ${\mathrm{I}}^2$ for multilevel and multivariate models. 2024. https://www.metafor-project.org/doku.php/tips:i2_multilevel_multivariate. Accessed July 21, 2025.Google Scholar
Maas Cora, JM, Hox Joop, J. Sufficient sample sizes for multilevel modeling. Methodology. 2005;1:8692.10.1027/1614-2241.1.3.86CrossRefGoogle Scholar
Bell, BA, Ferron, JM, Kromrey Jeffrey, D. Cluster size in multilevel models: the impact of sparse data structures on point and interval estimates in two-level models. In JSM Proceedings, Survey Research Methods Section. Alexandria, VA: American Statistical Association. 2008;11221129. Retrieved from http://www.asasrms.org/Proceedings/y2008f.html.Google Scholar
Austin, PC. Estimating multilevel logistic regression models when the number of clusters is low: a comparison of different statistical software procedures. The International Journal of Biostatistics. 2010;6:Article no. 16.10.2202/1557-4679.1195CrossRefGoogle Scholar
Zarychanski, R, Abou-Setta, AM, Turgeon, AF, et al. Association of Hydroxyethyl Starch Administration with Mortality and Acute Kidney Injury in critically ill patients requiring volume resuscitation: a systematic review and meta-analysis. JAMA. 2013;309:678.10.1001/jama.2013.430CrossRefGoogle ScholarPubMed
Pustejovsky James, E, Elizabeth, T. Meta-analysis with robust variance estimation: expanding the range of working models. Prev Sci. 2021;23:425438.10.1007/s11121-021-01246-3CrossRefGoogle ScholarPubMed
Feinstein, AR. Meta-analysis: statistical alchemy for the 21st century. J Clin Epidemiol. 1995;48:7179.10.1016/0895-4356(94)00110-CCrossRefGoogle ScholarPubMed
Wely, M. The good, the bad and the ugly: meta-analyses. Hum Reprod. 2014;29:16221626.10.1093/humrep/deu127CrossRefGoogle ScholarPubMed
R Core Team. R: A Language and Environment for Statistical Computing. R Foundation for Statistical Computing; 2023.Google Scholar
Oksanen, J, Simpson, GL, Guillaume Blanche, F, et al. vegan: community ecology package. 2022. R package version 2.6-4.Google Scholar
Wickham, H, François, R, Henry, L, Müller, K, Vaughan, D. dplyr: a grammar of data manipulation. 2023. R package version 1.1.2.Google Scholar
Wickham, H. stringr: simple, consistent wrappers for common string operations. 2022. R package version 1.5.0.Google Scholar
Bjornstad, ON. ncf: spatial covariance functions. 2022. R package version 1.3-2.Google Scholar
Csárdi, G, Nepusz, T, Traag, V, et al. igraph: network analysis and visualization in R. 2024. R package version 2.0.1.1.Google Scholar
Carlen, J. Creating co-author networks in R. 2019. https://datalab.ucdavis.edu/2019/08/27/creating-co-author-networks-in-r/. Accessed January 01, 2024.Google Scholar
Figure 0

Figure 1 Example of simulated authorship network as analysed in Section 3.1. See Table 1 for input and output parameters and their explanation.

Figure 1

Figure 2 Correlograms of simulated network with authorship network bias. The left panel shows raw and adjusted effect sizes (ESs) for the geodesic distance model. Because of too few data, no confidence interval could be calculated at geodesic distance $= 6$. The right panel shows raw and adjusted ESs for the Jaccard model. Note that Jaccard distance $\neq 1$ can only occur for paper pairs with geodesic distance 1. Jaccard distance was calculated by subtracting Jaccard similarity from one. Note that the CI width is proportional to the number of pairwise distances in the respective distance bin. Wide CIs are less informative of the correction outcome but are also less important for assessment because of their low number.

Figure 2

Table 1 Parameters used for network simulation, as well as the respective value for the analysed exemplary network

Figure 3

Figure 3 Forest plots of the simulated network analysed in Section 3.1. Top: Studies are ordered and coloured according to their cluster membership to identify deviant clusters. Bottom: Only 20 out of 400 total studies are shown. Grey-shaded diamonds signify effect sizes that were adjusted for the influence of paper proximity in the network. Black diamonds show model results of the models without and with (“adjusted effect”) authorship similarity matrix.

Figure 4

Figure 4 Influence of bias and connectedness of networks on the proportion of successfully corrected biased results. Proportions refer to corrected effect sizes out of 50 simulated networks for each combination. Proportions of success are the proportions of non-significant results for the full model, where the standard model resulted in a significant effect.

Figure 5

Table 2 Network characteristics and effect sizes before and after correction for the reanalysed meta-analyses

Figure 6

Figure 5 Diagnostic plots of (a) Bakdash et al. (2021),30 and (c) Dinu et al. (2017).31 See Appendix B of the Supplementary Material for other meta-analyses.

Supplementary material: File

Rieck et al. supplementary material

Rieck et al. supplementary material
Download Rieck et al. supplementary material(File)
File 991.4 KB