ArticlePDF Available

Recommendations for examining and interpreting funnel plot asymmetry in meta-analyses of randomised controlled trials

Authors:

Abstract and Figures

Funnel plots, and tests for funnel plot asymmetry, have been widely used to examine bias in the results of meta-analyses. Funnel plot asymmetry should not be equated with publication bias, because it has a number of other possible causes. This article describes how to interpret funnel plot asymmetry, recommends appropriate tests, and explains the implications for choice of meta-analysis modelThe 1997 paper describing the test for funnel plot asymmetry proposed by Egger et al 1 is one of the most cited articles in the history of BMJ.1 Despite the recommendations contained in this and subsequent papers,2 3 funnel plot asymmetry is often, wrongly, equated with publication or other reporting biases. The use and appropriate interpretation of funnel plots and tests for funnel plot asymmetry have been controversial because of questions about statistical validity,4 disputes over appropriate interpretation,3 5 6 and low power of the tests.2This article recommends how to examine and interpret funnel plot asymmetry (also known as small study effects2) in meta-analyses of randomised controlled trials. The recommendations are based on a detailed MEDLINE review of literature published up to 2007 and discussions among methodologists, who extended and adapted guidance previously summarised in the Cochrane Handbook for Systematic Reviews of Interventions.7What is a funnel plot?A funnel plot is a scatter plot of the effect estimates from individual studies against some measure of each study’s size or precision. The standard error of the effect estimate is often chosen as the measure of study size and plotted on the vertical axis8 with a reversed scale that places the larger, most powerful studies towards the top. The effect estimates from smaller studies should scatter more widely at the bottom, with the spread narrowing among larger studies.9 In the absence of bias and between study heterogeneity, the scatter will be due to sampling variation alone and the plot will resemble a symmetrical inverted funnel (fig 1⇓). A triangle centred on a fixed effect summary estimate and extending 1.96 standard errors either side will include about 95% of studies if no bias is present and the fixed effect assumption (that the true treatment effect is the same in each study) is valid. The appendix on bmj.com discusses choice of axis in funnel plots.View larger version:In a new windowDownload as PowerPoint SlideFig 1 Example of symmetrical funnel plot. The outer dashed lines indicate the triangular region within which 95% of studies are expected to lie in the absence of both biases and heterogeneity (fixed effect summary log odds ratio±1.96×standard error of summary log odds ratio). The solid vertical line corresponds to no intervention effectImplications of heterogeneity, reporting bias, and chance Heterogeneity, reporting bias, and chance may all lead to asymmetry or other shapes in funnel plots (box). Funnel plot asymmetry may also be an artefact of the choice of statistics being plotted (see appendix). The presence of any shape in a funnel plot is contingent on the studies having a range of standard errors, since otherwise they would lie on a horizontal line.Box 1: Possible sources of asymmetry in funnel plots (adapted from Egger et al1)Reporting biasesPublication bias: Delayed publication (also known as time lag or pipeline) bias Location biases (eg, language bias, citation bias, multiple publication bias)Selective outcome reportingSelective analysis reportingPoor methodological quality leading to spuriously inflated effects in smaller studiesPoor methodological designInadequate analysisFraudTrue heterogeneitySize of effect differs according to study size (eg, because of differences in the intensity of interventions or in underlying risk between studies of different sizes)ArtefactualIn some circumstances, sampling variation can lead to an association between the intervention effect and its standard errorChanceAsymmetry may occur by chance, which motivates the use of asymmetry testsHeterogeneityStatistical heterogeneity refers to differences between study results beyond those attributable to chance. It may arise because of clinical differences between studies (for example, setting, types of participants, or implementation of the intervention) or methodological differences (such as extent of control over bias). A random effects model is often used to incorporate heterogeneity in meta-analyses. If the heterogeneity fits with the assumptions of this model, a funnel plot will be symmetrical but with additional horizontal scatter. If heterogeneity is large it may overwhelm the sampling error, so that the plot appears cylindrical.Heterogeneity will lead to funnel plot asymmetry if it induces a correlation between study sizes and intervention effects.5 For example, substantial benefit may be seen only in high risk patients, and these may be preferentially included in early, small studies.10 Or the intervention may have been implemented less thoroughly in larger studies, resulting in smaller effect estimates compared with smaller studies.11Figure 2⇓ shows funnel plot asymmetry arising from heterogeneity that is due entirely to there being three distinct subgroups of studies, each with a different intervention effect.12 The separate funnels for each subgroup are symmetrical. Unfortunately, in practice, important sources of heterogeneity are often unknown.View larger version:In a new windowDownload as PowerPoint SlideFig 2 Illustration of funnel plot asymmetry due to heterogeneity, in the form of three distinct subgroups of studies. Funnel plot including all studies (top left) shows clear asymmetry (P<0.001 from Egger test for funnel plot asymmetry). P values for each subgroup are all >0.49.Differences in methodological quality may also cause heterogeneity and lead to funnel plot asymmetry. Smaller studies tend to be conducted and analysed with less methodological rigour than larger studies,13 and trials of lower quality also tend to show larger intervention effects.14 15Reporting biasReporting biases arise when the dissemination of research findings is influenced by the nature and direction of results. Statistically significant “positive” results are more likely to be published, published rapidly, published in English, published more than once, published in high impact journals, and cited by others.16 17 18 19 Data that would lead to negative results may be filtered, manipulated, or presented in such a way that they become positive.14 20 Reporting biases can have three types of consequence for a meta-analysis:A systematic review may fail to locate an eligible study because all information about it is suppressed or hard to find (publication bias) A located study may not provide usable data for the outcome of interest because the study authors did not consider the result sufficiently interesting (selective outcome reporting) A located study may provide biased results for some outcome—for example, by presenting the result with the smallest P value or largest effect estimate after trying several analysis methods (selective analysis reporting).These biases may cause funnel plot asymmetry if statistically significant results suggesting a beneficial effect are more likely to be published than non-significant results. Such asymmetry may be exaggerated if there is a further tendency for smaller studies to be more prone to selective suppression of results than larger studies. This is often assumed to be the case for randomised trials. For instance, it is probably more difficult to make a large study disappear without trace, while a small study can easily be lost in a file drawer.21 The same may apply to specific outcomes—for example, it is difficult not to report on mortality or myocardial infarction if these are outcomes of a large study. Smaller studies have more sampling error in their effect estimates. Thus even though the risk of a false positive significant finding is the same, multiple analyses are more likely to yield a large effect estimate that may seem worth publishing. However, biases may not act this way in real life; funnel plots could be symmetrical even in the presence of publication bias or selective outcome reporting19 22—for example, if the published findings point to effects in different directions but unreported results indicate neither direction. Alternatively, bias may have affected few studies and therefore not cause glaring asymmetry.ChanceThe role of chance is critical for interpretation of funnel plots because most meta-analyses of randomised trials in healthcare contain few studies.2 Investigations of relations across studies in a meta-analysis are seriously prone to false positive findings when there is a small number of studies and heterogeneity across studies,23 and this may affect funnel plot symmetry.Interpreting funnel plot asymmetryAuthors of systematic reviews should distinguish between possible reasons for funnel plot asymmetry (box 1). Knowledge of the intervention, and the circumstances in which it was implemented in different studies, can help identify causes of asymmetry in funnel plots, which should also be interpreted in the context of susceptibility to biases of research in the field of interest. Potential conflicts of interest, whether outcomes and analyses have been standardised, and extent of trial registration may need to be considered. For example, studies of antidepressants generate substantial conflicts of interest because the drugs generate vast sales revenues. Furthermore, there are hundreds of outcome scales, analyses can be very flexible, and trial registration was uncommon until recently.24 Conversely, in a prospective meta-analysis where all data are included and all analyses fully standardised and conducted according to a predetermined protocol, publication or reporting biases cannot exist. Reporting bias is therefore more likely to be a cause of an asymmetric plot in the first situation than in the second.Terrin et al found that researchers were poor at identifying publication bias from funnel plots.5 Including contour lines corresponding to perceived milestones of statistical significance (P=0.01, 0.05, 0.1, etc) may aid visual interpretation.25 If studies seem to be missing in areas of non-significance (fig 3⇓, top) then asymmetry may be due to reporting bias, although other explanations should still be considered. If the supposed missing studies are in areas of higher significance or in a direction likely to be considered desirable to their authors (fig 3⇓, bottom), asymmetry is probably due to factors other than reporting bias. View larger version:In a new windowDownload as PowerPoint SlideFig 3 Contour enhanced funnel plots. In the top diagram there is a suggestion of missing studies in the middle and right of the plot, broadly in the white area of non-significance, making publication bias plausible. In the bottom diagram there is a suggestion of missing studies on the bottom left hand side of the plot. Since most of this area contains regions of high significance, publication bias is unlikely to be the underlying cause of asymmetryStatistical tests for funnel plot asymmetryA test for funnel plot asymmetry (sometimes referred to as a test for small study effects) examines whether the association between estimated intervention effects and a measure of study size is greater than might be expected to occur by chance. These tests typically have low power, so even when a test does not provide evidence of asymmetry, bias cannot be excluded. For outcomes measured on a continuous scale a test based on a weighted linear regression of the effect estimates on their standard errors is straightforward.1 When outcomes are dichotomous and intervention effects are expressed as odds ratios, this corresponds to an inverse variance weighted linear regression of the log odds ratio on its standard error.2 Unfortunately, there are statistical problems because the standard error of the log odds ratio is mathematically linked to the size of the odds ratio, even in the absence of small study effects.2 4 Many authors have therefore proposed alternative tests (see appendix on bmj.com).4 26 27 28Because it is impossible to know the precise mechanism(s) leading to funnel plot asymmetry, simulation studies (in which tests are evaluated on large numbers of computer generated datasets) are required to evaluate test characteristics. Most have examined a range of assumptions about the extent of reporting bias by selectively removing studies from simulated datasets.26 27 28 After reviewing the results of these studies, and based on theoretical considerations, we formulated recommendations on testing for funnel plot asymmetry (box 2). The appendix describes the proposed tests, explains the reasons that some were not recommended, and discusses funnel plots for intervention effects measured as risk ratios, risk differences, and standardised mean differences. Our recommendations imply that tests for funnel plot asymmetry should be used in only a minority of meta-analyses.29Box 2: Recommendations on testing for funnel plot asymmetryAll types of outcomeAs a rule of thumb, tests for funnel plot asymmetry should not be used when there are fewer than 10 studies in the meta-analysis because test power is usually too low to distinguish chance from real asymmetry. (The lower the power of a test, the higher the proportion of “statistically significant” results in which there is in reality no association between study size and intervention effects). In some situations—for example, when there is substantial heterogeneity—the minimum number of studies may be substantially more than 10Test results should be interpreted in the context of visual inspection of funnel plots— for example, are there studies with markedly different intervention effect estimates or studies that are highly influential in the asymmetry test? Even if an asymmetry test is statistically significant, publication bias can probably be excluded if small studies tend to lead to lower estimates of benefit than larger studies or if there are no studies with significant resultsWhen there is evidence of funnel plot asymmetry, publication bias is only one possible explanation (see box 1)As far as possible, testing strategy should be specified in advance: choice of test may depend on the degree of heterogeneity observed. Applying and reporting many tests is discouraged: if more than one test is used, all test results should be reported Tests for funnel plot asymmetry should not be used if the standard errors of the intervention effect estimates are all similar (the studies are of similar sizes)Continuous outcomes with intervention effects measured as mean differencesThe test proposed by Egger et al may be used to test for funnel plot asymmetry.1 There is no reason to prefer more recently proposed tests, although their relative advantages and disadvantages have not been formally examined. General considerations suggest that the power will be greater than for dichotomous outcomes but that use of the test with substantially fewer than 10 studies would be unwiseDichotomous outcomes with intervention effects measured as odds ratiosThe tests proposed by Harbord et al26 and Peters et al27 avoid the mathematical association between the log odds ratio and its standard error when there is a substantial intervention effect while retaining power compared with alternative tests. However, false positive results may still occur if there is substantial between study heterogeneityIf there is substantial between study heterogeneity (the estimated heterogeneity variance of log odds ratios, τ2, is >0.1) only the arcsine test including random effects, proposed by Rücker et al, has been shown to work reasonably well.28 However, it is slightly conservative in the absence of heterogeneity and its interpretation is less familiar than for other tests because it is based on an arcsine transformation.When τ2 is <0.1, one of the tests proposed by Harbord et al,26 Peters et al,27 or Rücker et al28 can be used. Test performance generally deteriorates as τ2 increases.Funnel plots and meta-analysis modelsFixed and random effects modelsFunnel plots can help guide choice of meta-analysis method. Random effects meta-analyses weight studies relatively more equally than fixed effect analyses by incorporating the between study variance into the denominator of each weight. If effect estimates are related to standard errors (funnel plot asymmetry), the random effects estimate will be pulled more towards findings from smaller studies than the fixed effect estimate will be. Random effects models can thus have undesirable consequences and are not always conservative.30The trials of intravenous magnesium after myocardial infarction provide an extreme example of the differences between fixed and random effects analyses that can arise in the presence of funnel plot asymmetry.31 Beneficial effects on mortality, found in a meta-analysis of small studies,32 were subsequently contradicted when the very large ISIS-4 study found no evidence of benefit.33 A contour enhanced funnel plot (fig 4⇓) gives a clear visual impression of asymmetry, which is confirmed by small P values from the Harbord and Peters tests (P<0.001 and P=0.002 respectively).View larger version:In a new windowDownload as PowerPoint SlideFig 4 Contour enhanced funnel plot for trials of the effect of intravenous magnesium on mortality after myocardial infarctionFigure 5⇓ shows that in a fixed effect analysis ISIS-4 receives 90% of the weight, and there is no evidence of a beneficial effect. However, there is clear evidence of between study heterogeneity (P<0.001, I2=68%), and in a random effects analysis the small studies dominate so that intervention appears beneficial. To interpret the accumulated evidence, it is necessary to make a judgment about the validity or relevance of the combined evidence from the smaller studies compared with that from ISIS-4. The contour enhanced funnel plot suggests that publication bias does not completely explain the asymmetry, since many of the beneficial effects reported from smaller studies were not significant. Plausible explanations for these results are that methodological flaws in the smaller studies, or changes in the standard of care (widespread adoption of treatments such as aspirin, heparin, and thrombolysis), led to apparent beneficial effects of magnesium. This belief was reinforced by the subsequent publication of the MAGIC trial, in which magnesium added to these treatments which also found no evidence of benefit on mortality (odds ratio 1.0, 95% confidence interval 0.8 to 1.1).34View larger version:In a new windowDownload as PowerPoint SlideFig 5 Comparison of fixed and random effects meta-analytical estimates of the effect of intravenous magnesium on mortality after myocardial infarctionWe recommend that when review authors are concerned about funnel plot asymmetry in a meta-analysis with evidence of between study heterogeneity, they should compare the fixed and random effects estimates of the intervention effect. If the random effects estimate is more beneficial, authors should consider whether it is plausible that the intervention is more effective in smaller studies. Formal investigations of heterogeneity of effects may reveal explanations for funnel plot asymmetry, in which case presentation of results should focus on these. If larger studies tend to be methodologically superior to smaller studies, or were conducted in circumstances more typical of the use of the intervention in practice, it may be appropriate to include only larger studies in the meta-analysis.Extrapolation of a funnel plot regression lineAn assumed relation between susceptibility to bias and study size can be exploited by extrapolating within a funnel plot. When funnel plot asymmetry is due to bias rather than substantive heterogeneity, it is usually assumed that results from larger studies are more believable than those from smaller studies because they are less susceptible to methodological flaws or reporting biases. Extrapolating a regression line on a funnel plot to minimum bias (maximum sample size) produces a meta-analytical estimate that can be regarded as corrected for such biases.35 36 37 However, because it is difficult to distinguish between asymmetry due to bias and asymmetry due to heterogeneity or chance, the broad applicability of such approaches is uncertain. Further approaches to adjusting for publication bias are described and discussed in the appendix.DiscussionReporting biases are one of a number of possible explanations for the associations between study size and effect size that are displayed in asymmetric funnel plots. Examining and testing for funnel plot asymmetry, when appropriate, is an important means of addressing bias in meta-analyses, but the multiple causes of asymmetry and limited power of asymmetry tests mean that other ways to address reporting biases are also of importance. Searches of online trial registries can identify unpublished trials, although they do not currently guarantee access to trial protocols and results. When there are no registered but unpublished trials, and the outcome of interest is reported by all trials, restricting meta-analyses to registered trials should preclude publication bias. Recent comparisons of results of published trials with those submitted for regulatory approval have also provided clear evidence of reporting bias.38 39 Methods for dealing with selective reporting of outcomes have been described elsewhere. 40Our recommendations apply to meta-analyses of randomised trials, and their applicability in other contexts such as meta-analyses of epidemiological or diagnostic test studies is unclear.41 The performance of tests for funnel plot asymmetry in these contexts is likely to differ from that in meta-analyses of randomised trials. Further factors, such as confounding and precision of measurements, may cause a relation between study size and effect estimates in observational studies. For example, large studies based on routinely collected data might not fully control confounding compared with smaller, purpose designed studies that collected a wide range of potential confounding variables. Alternatively, larger studies might use self reported exposure levels, which are more error prone, while smaller studies used precise measuring instruments. However, simulation studies have usually not considered such situations. An exception is for diagnostic studies, where large imbalances in group sizes and substantial odds ratios lead to poor performance of some tests: that proposed by Deeks et al was designed for use in this context.4Summary points Inferences on the presence of bias or heterogeneity should consider different causes of funnel plot asymmetry and should not be based on visual inspection of funnel plots aloneThey should be informed by contextual factors, including the plausibility of publication bias as an explanation for the asymmetryTesting for funnel plot asymmetry should follow the recommendations detailed in this articleThe fixed and random effects estimates of the intervention effect should be compared when funnel plot asymmetry exists in a meta-analysis with between study heterogeneityNotesCite this as: BMJ 2011;342:d4002FootnotesContributors: All authors contributed to the drafting and editing of the manuscript. DA, JC, JD, RMH, JPTH, JPAI, DRJ, DM, JP, GR, JACS, AJS and JT contributed to the chapter in the Cochrane Handbook for Systematic Reviews of Interventions on which our recommendations on testing for funnel plot asymmetry are based. JACS will act as guarantor.Funding: Funded in part by the Cochrane Collaboration Bias Methods Group, which receives infrastructure funding as part of a commitment by the Canadian Institutes of Health Research (CIHR) and the Canadian Agency for Drugs and Technologies in Health (CADTH) to fund Canadian based Cochrane entities. This supports dissemination activities, web hosting, travel, training, workshops and a full time coordinator position. JPTH was funded by MRC Grant U.1052.00.011. DGA is supported by Cancer Research UK. GR was supported by a grant from Deutsche Forschungsgemeinschaft (FOR 534 Schw 821/2-2).Competing interests. JC, JJD, SD, RMH, JPAI, DRJ, PM, JP, GR, GS, JACS and AJS are all authors on papers proposing tests for funnel plot asymmetry, but have no commercial interests in the use of these tests. All authors have completed the ICJME unified disclosure form at www.icmje.org/coi_disclosure.pdf (available on request from the corresponding author) and declare that they have no financial or non-financial interests that may be relevant to the submitted work.Provenance and peer review: Not commissioned; externally peer reviewed.References↵Egger M, Davey Smith G, Schneider M, Minder C. Bias in meta-analysis detected by a simple, graphical test. BMJ1997;315:629-34.OpenUrlFREE Full Text↵Sterne JAC, Gavaghan D, Egger M. Publication and related bias in meta-analysis: power of statistical tests and prevalence in the literature. J Clin Epidemiol2000;53:1119-29.OpenUrlCrossRefMedlineWeb of Science↵Lau J, Ioannidis JP, Terrin N, Schmid CH, Olkin I. The case of the misleading funnel plot. BMJ2006;333:597-600.OpenUrlFREE Full Text↵Deeks JJ, Macaskill P, Irwig L. The performance of tests of publication bias and other sample size effects in systematic reviews of diagnostic test accuracy was assessed. J Clin Epidemiol2005;58:882-93.OpenUrlCrossRefMedlineWeb of Science↵Terrin N, Schmid CH, Lau J. In an empirical evaluation of the funnel plot, researchers could not visually identify publication bias. J Clin Epidemiol2005;58:894-901.OpenUrlCrossRefMedlineWeb of Science↵Ioannidis JP. Interpretation of tests of heterogeneity and bias in meta-analysis. J Eval Clin Pract 2008;14:951-7.OpenUrlMedlineWeb of Science↵Sterne JAC, Egger M, Moher D. Addressing reporting biases. In: Higgins JPT, Green S, eds. Cochrane handbook for systematic reviews of interventions. Wiley, 2008.↵Sterne JAC, Egger M. Funnel plots for detecting bias in meta-analysis: guidelines on choice of axis. J Clin Epidemiol2001;54:1046-55.OpenUrlCrossRefMedlineWeb of Science↵Begg CB, Berlin JA. Publication bias: a problem in interpreting medical data. J R Statist Soc A1988;151:419-63.OpenUrlCrossRef↵Davey Smith G, Egger M. Who benefits from medical interventions? Treating low risk patients can be a high risk strategy. BMJ1994;308:72-4.OpenUrlFREE Full Text↵Stuck AE, Siu AL, Wieland GD, Adams J, Rubenstein LZ. Comprehensive geriatric assessment: a meta-analysis of controlled trials. Lancet1993;342:1032-6.OpenUrlCrossRefMedlineWeb of Science↵Peters JL, Sutton AJ, Jones DR, Abrams KR, Rushton L, Moreno SG. Assessing publication bias in meta-analyses in the presence of between-study heterogeneity. J R Statist Soc A2010;173:575-91.OpenUrlCrossRef↵Egger M, Jüni P, Bartlett C, Holenstein F, Sterne J. How important are comprehensive literature searches and the assessment of trial quality in systematic reviews? Empirical study. Health Technol Assess2003;7:1-68.OpenUrlMedline↵Ioannidis JP. Why most discovered true associations are inflated. Epidemiology2008;19:640-8.OpenUrlCrossRefMedlineWeb of Science↵Wood L, Egger M, Gluud LL, Schulz KF, Jüni P, Altman DG, et al. Empirical evidence of bias in treatment effect estimates in controlled trials with different interventions and outcomes: meta-epidemiological study. BMJ2008;336:601-5.OpenUrlFREE Full Text↵Hopewell S, Clarke M, Stewart L, Tierney J. Time to publication for results of clinical trials. Cochrane Database Syst Rev2007;2:MR000011.OpenUrlMedline↵Hopewell S, Loudon K, Clarke MJ, Oxman AD, Dickersin K. Publication bias in clinical trials due to statistical significance or direction of trial results. Cochrane Database Syst Rev2009;1:MR000006.OpenUrlMedline↵Song F, Parekh S, Hooper L, Loke YK, Ryder J, Sutton J, et al. Dissemination and publication of research findings: an updated review of related biases. Health Technol Assess2010;14:iii,ix-iii,193.OpenUrlMedlineWeb of Science↵Dwan K, Altman DG, Arnaiz JA, Bloom J, Chan AW, Cronin E, et al. Systematic review of the empirical evidence of study publication bias and outcome reporting bias. PLoS ONE2008;3:e3081.OpenUrlCrossRefMedline↵Turner EH, Matthews AM, Linardatos E, Tell RA, Rosenthal R. Selective publication of antidepressant trials and its influence on apparent efficacy. N Engl J Med2008;358:252-60.OpenUrlCrossRefMedline↵Rosenthal R. The “file drawer” problem and tolerance for null results. Psychol Bull1979;86:638-41.OpenUrlCrossRefWeb of Science↵Chan AW, Hrobjartsson A, Haahr MT, Gotzsche PC, Altman DG. Empirical evidence for selective reporting of outcomes in randomized trials: comparison of protocols to published articles. JAMA2004;291:2457-65.OpenUrlFREE Full Text↵Higgins JP, Thompson SG. Controlling the risk of spurious findings from meta-regression. Stat Med2004;23:1663-82.OpenUrlCrossRefMedlineWeb of Science↵Ioannidis JP. Effectiveness of antidepressants: an evidence myth constructed from a thousand randomized trials? Philos Ethics Humanit Med2008;3:14.OpenUrlCrossRefMedline↵Peters J, Sutton AJ, Jones DR, Abrams KR, Rushton L. Contour-enhanced meta-analysis funnel plots help distinguish publication bias from other causes of asymmetry. J Clin Epidemiol2008;61:991-6.OpenUrlCrossRefMedlineWeb of Science↵Harbord RM, Egger M, Sterne JA. A modified test for small-study effects in meta-analyses of controlled trials with binary endpoints. Stat Med2006;25:3443-57.OpenUrlCrossRefMedlineWeb of Science↵Peters JL, Sutton AJ, Jones DR, Abrams KR, Rushton L. Comparison of two methods to detect publication bias in meta-analysis. JAMA2006;295:676-80.OpenUrlFREE Full Text↵Rücker G, Schwarzer G, Carpenter J. Arcsine test for publication bias in meta-analyses with binary outcomes. Stat Med2008;27:746-63.OpenUrlCrossRefMedlineWeb of Science↵Ioannidis JP, Trikalinos TA. The appropriateness of asymmetry tests for publication bias in meta-analyses: a large survey. CMAJ2007;176:1091-6.OpenUrlFREE Full Text↵Poole C, Greenland S. Random-effects meta-analyses are not always conservative. Am J Epidemiol1999;150:469-75.OpenUrlFREE Full Text↵Egger M, Davey Smith G. Misleading meta-analysis. Lessons from an “effective, safe, simple” intervention that wasn’t. BMJ1995;310:752-4.OpenUrlFREE Full Text↵Teo KK, Yusuf S, Collins R, Held PH, Peto R. Effects of intravenous magnesium in suspected acute myocardial infarction: overview of randomised trials. BMJ1991;303:1499-503.OpenUrlFREE Full Text↵ISIS-4 (Fourth International Study of Infarct Survival) Collaborative Group. ISIS-4: a randomised factorial trial assessing early oral captopril, oral mononitrate, and intravenous magnesium sulphate in 58,050 patients with suspected acute myocardial infarction. Lancet1995;345:669-85.OpenUrlCrossRefMedlineWeb of Science↵Early administration of intravenous magnesium to high-risk patients with acute myocardial infarction in the Magnesium in Coronaries (MAGIC) Trial: a randomised controlled trial. Lancet2002;360:1189-96.OpenUrlCrossRefMedlineWeb of Science↵Shang A, Huwiler-Muntener K, Nartey L, Jüni P, Dörig S, Stene JA, et al. Are the clinical effects of homoeopathy placebo effects? Comparative study of placebo-controlled trials of homoeopathy and allopathy. Lancet2005;366:726-32.OpenUrlCrossRefMedlineWeb of Science↵Moreno SG, Sutton AJ, Ades AE, Stanley TD, Abrams KR, Peters JL, et al. Assessment of regression-based methods to adjust for publication bias through a comprehensive simulation study. BMC Med Res Methodol2009;9:2.OpenUrlCrossRefMedline↵Rucker G, Schwarzer G, Carpenter JR, Binder H, Schumacher M. Treatment-effect estimates adjusted for small-study effects via a limit meta-analysis. Biostatistics2011;12:122-42.OpenUrlFREE Full Text↵Moreno SG, Sutton AJ, Turner EH, Abrams KR, Cooper NJ, Palmer TM, et al. Novel methods to deal with publication biases: secondary analysis of antidepressant trials in the FDA trial registry database and related journal publications. BMJ2009;339:b2981.OpenUrlFREE Full Text↵Eyding D, Lelgemann M, Grouven U, Härter M, Kromp M, Kaiser T, et al. Reboxetine for acute treatment of major depression: systematic review and meta-analysis of published and unpublished placebo and selective serotonin reuptake inhibitor controlled trials. BMJ2010;341:c4737. OpenUrlFREE Full Text↵Kirkham JJ, Dwan KM, Altman DG, Gamble C, Dodd S, Smyth R, et al. The impact of outcome reporting bias in randomised controlled trials on a cohort of systematic reviews. BMJ2010;340:c365. OpenUrlFREE Full Text↵Egger M, Schneider M, Davey Smith G. Spurious precision? Meta-analysis of observational studies. BMJ1998;316:140-4.OpenUrlFREE Full Text
Content may be subject to copyright.
Recommendations for examining and interpreting
funnel plot asymmetry in meta-analyses of randomised
controlled trials
Funnel plots, and tests for funnel plot asymmetry, have been widely used to examine bias in the
results of meta-analyses. Funnel plot asymmetry should not be equated with publication bias,
because it has a number of other possible causes.This article describes how to interpret funnel
plot asymmetry, recommends appropriate tests, and explains the implications for choice of
meta-analysis model
Jonathan A C Sterne professor1, Alex J Sutton professor2, John P A Ioannidis professor and director 3,
Norma Terrin associate professor 4, David R Jones professor 2, Joseph Lau professor 4, James
Carpenter reader 5, Gerta Rücker research assistant 6, Roger M Harbord research associate 1,
Christopher H Schmid professor 4, Jennifer Tetzlaff research coordinator 7, Jonathan J Deeks
professor8, Jaime Peters research fellow 9, Petra Macaskill associate professor10, Guido Schwarzer
research assistant 6, Sue Duval assistant professor 11, Douglas G Altman professor 12, David Moher
senior scientist 7, Julian P T Higgins senior statistician 13
1School of Social and Community Medicine, University of Bristol, Bristol BS8 2PS, UK; 2Department of Health Sciences, University of Leicester,
Leicester, UK; 3Stanford Prevention Research Center, Stanford University School of Medicine, Stanford, CA, USA; 4Institute for Clinical Research
and Health Policy Studies, Tufts Medical Center, Boston, MA, USA; 5Medical Statistics Unit, London School of Hygiene and Tropical Medicine,
London, UK ; 6Institute of Medical Biometry and Medical Informatics, University Medical Center Freiburg, Germany; 7Clinical Epidemiology Program,
Ottawa Hospital Research Institute, Ottawa, Ontario, Canada; 8School of Health and Population Sciences, University of Birmingham, Birmingham,
UK; 9Peninsula Medical School, University of Exeter, Exeter, UK;10School of Public Health, University of Sydney, NSW, Australia;11University of
Minnesota School of Public Health, Minneapolis, MN, USA; 12Centre for Statistics in Medicine, University of Oxford, Oxford, UK; 13MRC Biostatistics
Unit, Cambridge, UK
The 1997 paper describing the test for funnel plot asymmetry
proposed by Egger et al 1is one of the most cited articles in the
history of BMJ.1Despite the recommendations contained in this
and subsequent papers,2 3 funnel plot asymmetry is often,
wrongly, equated with publication or other reporting biases.
The use and appropriate interpretation of funnel plots and tests
for funnel plot asymmetry have been controversial because of
questions about statistical validity,4disputes over appropriate
interpretation,356and low power of the tests.2
This article recommends how to examine and interpret funnel
plot asymmetry (also known as small study effects2) in
meta-analyses of randomised controlled trials. The
recommendations are based on a detailed MEDLINE review of
literature published up to 2007 and discussions among
methodologists, who extended and adapted guidance previously
summarised in the Cochrane Handbook for Systematic Reviews
of Interventions.7
What is a funnel plot?
A funnel plot is a scatter plot of the effect estimates from
individual studies against some measure of each study’s size or
precision. The standard error of the effect estimate is often
chosen as the measure of study size and plotted on the vertical
axis8with a reversed scale that places the larger, most powerful
studies towards the top. The effect estimates from smaller studies
should scatter more widely at the bottom, with the spread
narrowing among larger studies.9In the absence of bias and
between study heterogeneity, the scatter will be due to sampling
variation alone and the plot will resemble a symmetrical inverted
funnel (fig 1). A triangle centred on a fixed effect summary
estimate and extending 1.96 standard errors either side will
Correspondence to: J A C Sterne jonathan.sterne@bristol.ac.uk
Technical appendix (see http://www.bmj.com/content/342/bmj.d4002/suppl/DC1)
Reprints: http://journals.bmj.com/cgi/reprintform Subscribe: http://resources.bmj.com/bmj/subscribers/how-to-subscribe
BMJ 2011;342:d4002 doi: 10.1136/bmj.d4002 Page 1 of 8
ResearchMethods & Reporting
RESEARCH METHODS & REPORTING
include about 95% of studies if no bias is present and the fixed
effect assumption (that the true treatment effect is the same in
each study) is valid. The appendix on bmj.com discusses choice
of axis in funnel plots.
Implications of heterogeneity, reporting
bias, and chance
Heterogeneity, reporting bias, and chance may all lead to
asymmetry or other shapes in funnel plots (box). Funnel plot
asymmetry may also be an artefact of the choice of statistics
being plotted (see appendix). The presence of any shape in a
funnel plot is contingent on the studies having a range of
standard errors, since otherwise they would lie on a horizontal
line.
Heterogeneity
Statistical heterogeneity refers to differences between study
results beyond those attributable to chance. It may arise because
of clinical differences between studies (for example, setting,
types of participants, or implementation of the intervention) or
methodological differences (such as extent of control over bias).
A random effects model is often used to incorporate
heterogeneity in meta-analyses. If the heterogeneity fits with
the assumptions of this model, a funnel plot will be symmetrical
but with additional horizontal scatter. If heterogeneity is large
it may overwhelm the sampling error, so that the plot appears
cylindrical.
Heterogeneity will lead to funnel plot asymmetry if it induces
a correlation between study sizes and intervention effects.5For
example, substantial benefit may be seen only in high risk
patients, and these may be preferentially included in early, small
studies.10 Or the intervention may have been implemented less
thoroughly in larger studies, resulting in smaller effect estimates
compared with smaller studies.11
Figure 2 shows funnel plot asymmetry arising from
heterogeneity that is due entirely to there being three distinct
subgroups of studies, each with a different intervention effect.12
The separate funnels for each subgroup are symmetrical.
Unfortunately, in practice, important sources of heterogeneity
are often unknown.
Differences in methodological quality may also cause
heterogeneity and lead to funnel plot asymmetry. Smaller studies
tend to be conducted and analysed with less methodological
rigour than larger studies,13 and trials of lower quality also tend
to show larger intervention effects.14 15
Reporting bias
Reporting biases arise when the dissemination of research
findings is influenced by the nature and direction of results.
Statistically significant “positive” results are more likely to be
published, published rapidly, published in English, published
more than once, published in high impact journals, and cited
by others.16-19 Data that would lead to negative results may be
filtered, manipulated, or presented in such a way that they
become positive.14 20
Reporting biases can have three types of consequence for a
meta-analysis:
A systematic review may fail to locate an eligible study
because all information about it is suppressed or hard to
find (publication bias)
A located study may not provide usable data for the
outcome of interest because the study authors did not
consider the result sufficiently interesting (selective
outcome reporting)
A located study may provide biased results for some
outcome—for example, by presenting the result with the
smallest P value or largest effect estimate after trying
several analysis methods (selective analysis reporting).
These biases may cause funnel plot asymmetry if statistically
significant results suggesting a beneficial effect are more likely
to be published than non-significant results. Such asymmetry
may be exaggerated if there is a further tendency for smaller
studies to be more prone to selective suppression of results than
larger studies. This is often assumed to be the case for
randomised trials. For instance, it is probably more difficult to
make a large study disappear without trace, while a small study
can easily be lost in a file drawer.21 The same may apply to
specific outcomes—for example, it is difficult not to report on
mortality or myocardial infarction if these are outcomes of a
large study.
Smaller studies have more sampling error in their effect
estimates. Thus even though the risk of a false positive
significant finding is the same, multiple analyses are more likely
to yield a large effect estimate that may seem worth publishing.
However, biases may not act this way in real life; funnel plots
could be symmetrical even in the presence of publication bias
or selective outcome reporting19 22—for example, if the published
findings point to effects in different directions but unreported
results indicate neither direction. Alternatively, bias may have
affected few studies and therefore not cause glaring asymmetry.
Chance
The role of chance is critical for interpretation of funnel plots
because most meta-analyses of randomised trials in healthcare
contain few studies.2Investigations of relations across studies
in a meta-analysis are seriously prone to false positive findings
when there is a small number of studies and heterogeneity across
studies,23 and this may affect funnel plot symmetry.
Interpreting funnel plot asymmetry
Authors of systematic reviews should distinguish between
possible reasons for funnel plot asymmetry (box 1). Knowledge
of the intervention, and the circumstances in which it was
implemented in different studies, can help identify causes of
asymmetry in funnel plots, which should also be interpreted in
the context of susceptibility to biases of research in the field of
interest. Potential conflicts of interest, whether outcomes and
analyses have been standardised, and extent of trial registration
may need to be considered. For example, studies of
antidepressants generate substantial conflicts of interest because
the drugs generate vast sales revenues. Furthermore, there are
hundreds of outcome scales, analyses can be very flexible, and
trial registration was uncommon until recently.24 Conversely,
in a prospective meta-analysis where all data are included and
all analyses fully standardised and conducted according to a
predetermined protocol, publication or reporting biases cannot
exist. Reporting bias is therefore more likely to be a cause of
an asymmetric plot in the first situation than in the second.
Terrin et al found that researchers were poor at identifying
publication bias from funnel plots.5Including contour lines
corresponding to perceived milestones of statistical significance
(P=0.01, 0.05, 0.1, etc) may aid visual interpretation.25 If studies
seem to be missing in areas of non-significance (fig 3, top) then
asymmetry may be due to reporting bias, although other
explanations should still be considered. If the supposed missing
Reprints: http://journals.bmj.com/cgi/reprintform Subscribe: http://resources.bmj.com/bmj/subscribers/how-to-subscribe
BMJ 2011;342:d4002 doi: 10.1136/bmj.d4002 Page 2 of 8
RESEARCH METHODS & REPORTING
Box 1: Possible sources of asymmetry in funnel plots (adapted from Egger et al1)
Reporting biases
Publication bias:
Delayed publication (also known as time lag or pipeline) bias
Location biases (eg, language bias, citation bias, multiple publication bias)
Selective outcome reporting
Selective analysis reporting
Poor methodological quality leading to spuriously inflated effects in smaller studies
Poor methodological design
Inadequate analysis
Fraud
True heterogeneity
Size of effect differs according to study size (eg, because of differences in the intensity of interventions or in underlying
risk between studies of different sizes)
Artefactual
In some circumstances, sampling variation can lead to an association between the intervention effect and its standard
error
Chance
Asymmetry may occur by chance, which motivates the use of asymmetry tests
studies are in areas of higher significance or in a direction likely
to be considered desirable to their authors (fig 3, bottom),
asymmetry is probably due to factors other than reporting bias.
Statistical tests for funnel plot asymmetry
A test for funnel plot asymmetry (sometimes referred to as a
test for small study effects) examines whether the association
between estimated intervention effects and a measure of study
size is greater than might be expected to occur by chance. These
tests typically have low power, so even when a test does not
provide evidence of asymmetry, bias cannot be excluded. For
outcomes measured on a continuous scale a test based on a
weighted linear regression of the effect estimates on their
standard errors is straightforward.1When outcomes are
dichotomous and intervention effects are expressed as odds
ratios, this corresponds to an inverse variance weighted linear
regression of the log odds ratio on its standard error.2
Unfortunately, there are statistical problems because the standard
error of the log odds ratio is mathematically linked to the size
of the odds ratio, even in the absence of small study effects.2 4
Many authors have therefore proposed alternative tests (see
appendix on bmj.com).4 26-28
Because it is impossible to know the precise mechanism(s)
leading to funnel plot asymmetry, simulation studies (in which
tests are evaluated on large numbers of computer generated
datasets) are required to evaluate test characteristics. Most have
examined a range of assumptions about the extent of reporting
bias by selectively removing studies from simulated datasets.26-28
After reviewing the results of these studies, and based on
theoretical considerations, we formulated recommendations on
testing for funnel plot asymmetry (box 2). The appendix
describes the proposed tests, explains the reasons that some
were not recommended, and discusses funnel plots for
intervention effects measured as risk ratios, risk differences,
and standardised mean differences. Our recommendations imply
that tests for funnel plot asymmetry should be used in only a
minority of meta-analyses.29
Funnel plots and meta-analysis models
Fixed and random effects models
Funnel plots can help guide choice of meta-analysis method.
Random effects meta-analyses weight studies relatively more
equally than fixed effect analyses by incorporating the between
study variance into the denominator of each weight. If effect
estimates are related to standard errors (funnel plot asymmetry),
the random effects estimate will be pulled more towards findings
from smaller studies than the fixed effect estimate will be.
Random effects models can thus have undesirable consequences
and are not always conservative.30
The trials of intravenous magnesium after myocardial infarction
provide an extreme example of the differences between fixed
and random effects analyses that can arise in the presence of
funnel plot asymmetry.31 Beneficial effects on mortality, found
in a meta-analysis of small studies,32 were subsequently
contradicted when the very large ISIS-4 study found no evidence
of benefit.33 A contour enhanced funnel plot (fig 4) gives a clear
visual impression of asymmetry, which is confirmed by small
P values from the Harbord and Peters tests (P<0.001 and
P=0.002 respectively).
Figure 5 shows that in a fixed effect analysis ISIS-4 receives
90% of the weight, and there is no evidence of a beneficial
effect. However, there is clear evidence of between study
heterogeneity (P<0.001, I2=68%), and in a random effects
analysis the small studies dominate so that intervention appears
beneficial. To interpret the accumulated evidence, it is necessary
to make a judgment about the validity or relevance of the
combined evidence from the smaller studies compared with that
from ISIS-4. The contour enhanced funnel plot suggests that
publication bias does not completely explain the asymmetry,
since many of the beneficial effects reported from smaller
studies were not significant. Plausible explanations for these
results are that methodological flaws in the smaller studies, or
changes in the standard of care (widespread adoption of
treatments such as aspirin, heparin, and thrombolysis), led to
apparent beneficial effects of magnesium. This belief was
Reprints: http://journals.bmj.com/cgi/reprintform Subscribe: http://resources.bmj.com/bmj/subscribers/how-to-subscribe
BMJ 2011;342:d4002 doi: 10.1136/bmj.d4002 Page 3 of 8
RESEARCH METHODS & REPORTING
Box 2: Recommendations on testing for funnel plot asymmetry
All types of outcome
As a rule of thumb, tests for funnel plot asymmetry should not be used when there are fewer than 10 studies in the
meta-analysis because test power is usually too low to distinguish chance from real asymmetry. (The lower the power
of a test, the higher the proportion of “statistically significant” results in which there is in reality no association between
study size and intervention effects). In some situations—for example, when there is substantial heterogeneity—the
minimum number of studies may be substantially more than 10
Test results should be interpreted in the context of visual inspection of funnel plots— for example, are there studies
with markedly different intervention effect estimates or studies that are highly influential in the asymmetry test? Even
if an asymmetry test is statistically significant, publication bias can probably be excluded if small studies tend to lead
to lower estimates of benefit than larger studies or if there are no studies with significant results
When there is evidence of funnel plot asymmetry, publication bias is only one possible explanation (see box 1)
As far as possible, testing strategy should be specified in advance: choice of test may depend on the degree of
heterogeneity observed. Applying and reporting many tests is discouraged: if more than one test is used, all test
results should be reported
Tests for funnel plot asymmetry should not be used if the standard errors of the intervention effect estimates are all
similar (the studies are of similar sizes)
Continuous outcomes with intervention effects measured as mean differences
The test proposed by Egger et al may be used to test for funnel plot asymmetry.1There is no reason to prefer more
recently proposed tests, although their relative advantages and disadvantages have not been formally examined.
General considerations suggest that the power will be greater than for dichotomous outcomes but that use of the test
with substantially fewer than 10 studies would be unwise
Dichotomous outcomes with intervention effects measured as odds ratios
The tests proposed by Harbord et al26 and Peters et al27 avoid the mathematical association between the log odds
ratio and its standard error when there is a substantial intervention effect while retaining power compared with
alternative tests. However, false positive results may still occur if there is substantial between study heterogeneity
If there is substantial between study heterogeneity (the estimated heterogeneity variance of log odds ratios, τ2, is
>0.1) only the arcsine test including random effects, proposed by Rücker et al, has been shown to work reasonably
well.28 However, it is slightly conservative in the absence of heterogeneity and its interpretation is less familiar than
for other tests because it is based on an arcsine transformation.
When τ2is <0.1, one of the tests proposed by Harbord et al,26 Peters et al,27 or Rücker et al28 can be used. Test
performance generally deteriorates as τ2increases.
reinforced by the subsequent publication of the MAGIC trial,
in which magnesium added to these treatments which also found
no evidence of benefit on mortality (odds ratio 1.0, 95%
confidence interval 0.8 to 1.1).34
We recommend that when review authors are concerned about
funnel plot asymmetry in a meta-analysis with evidence of
between study heterogeneity, they should compare the fixed
and random effects estimates of the intervention effect. If the
random effects estimate is more beneficial, authors should
consider whether it is plausible that the intervention is more
effective in smaller studies. Formal investigations of
heterogeneity of effects may reveal explanations for funnel plot
asymmetry, in which case presentation of results should focus
on these. If larger studies tend to be methodologically superior
to smaller studies, or were conducted in circumstances more
typical of the use of the intervention in practice, it may be
appropriate to include only larger studies in the meta-analysis.
Extrapolation of a funnel plot regression line
An assumed relation between susceptibility to bias and study
size can be exploited by extrapolating within a funnel plot.
When funnel plot asymmetry is due to bias rather than
substantive heterogeneity, it is usually assumed that results from
larger studies are more believable than those from smaller
studies because they are less susceptible to methodological flaws
or reporting biases. Extrapolating a regression line on a funnel
plot to minimum bias (maximum sample size) produces a
meta-analytical estimate that can be regarded as corrected for
such biases.35 36 37 However, because it is difficult to distinguish
between asymmetry due to bias and asymmetry due to
heterogeneity or chance, the broad applicability of such
approaches is uncertain. Further approaches to adjusting for
publication bias are described and discussed in the appendix.
Discussion
Reporting biases are one of a number of possible explanations
for the associations between study size and effect size that are
displayed in asymmetric funnel plots. Examining and testing
for funnel plot asymmetry, when appropriate, is an important
means of addressing bias in meta-analyses, but the multiple
causes of asymmetry and limited power of asymmetry tests
mean that other ways to address reporting biases are also of
importance. Searches of online trial registries can identify
unpublished trials, although they do not currently guarantee
access to trial protocols and results. When there are no registered
but unpublished trials, and the outcome of interest is reported
by all trials, restricting meta-analyses to registered trials should
preclude publication bias. Recent comparisons of results of
published trials with those submitted for regulatory approval
have also provided clear evidence of reporting bias.38 39 Methods
for dealing with selective reporting of outcomes have been
described elsewhere. 40
Our recommendations apply to meta-analyses of randomised
trials, and their applicability in other contexts such as
meta-analyses of epidemiological or diagnostic test studies is
unclear.41 The performance of tests for funnel plot asymmetry
in these contexts is likely to differ from that in meta-analyses
of randomised trials. Further factors, such as confounding and
Reprints: http://journals.bmj.com/cgi/reprintform Subscribe: http://resources.bmj.com/bmj/subscribers/how-to-subscribe
BMJ 2011;342:d4002 doi: 10.1136/bmj.d4002 Page 4 of 8
RESEARCH METHODS & REPORTING
precision of measurements, may cause a relation between study
size and effect estimates in observational studies. For example,
large studies based on routinely collected data might not fully
control confounding compared with smaller, purpose designed
studies that collected a wide range of potential confounding
variables. Alternatively, larger studies might use self reported
exposure levels, which are more error prone, while smaller
studies used precise measuring instruments. However, simulation
studies have usually not considered such situations. An
exception is for diagnostic studies, where large imbalances in
group sizes and substantial odds ratios lead to poor performance
of some tests: that proposed by Deeks et al was designed for
use in this context.4
Contributors: All authors contributed to the drafting and editing of the
manuscript. DA, JC, JD, RMH, JPTH, JPAI, DRJ, DM, JP, GR, JACS,
AJS and JT contributed to the chapter in the Cochrane Handbook for
Systematic Reviews of Interventions on which our recommendations
on testing for funnel plot asymmetry are based. JACS will act as
guarantor.
Funding: Funded in part by the Cochrane Collaboration Bias Methods
Group, which receives infrastructure funding as part of a commitment
by the Canadian Institutes of Health Research (CIHR) and the Canadian
Agency for Drugs and Technologies in Health (CADTH) to fund Canadian
based Cochrane entities. This supports dissemination activities, web
hosting, travel, training, workshops and a full time coordinator position.
JPTH was funded by MRC Grant U.1052.00.011. DGA is supported by
Cancer Research UK. GR was supported by a grant from Deutsche
Forschungsgemeinschaft (FOR 534 Schw 821/2-2).
Competing interests. JC, JJD, SD, RMH, JPAI, DRJ, PM, JP, GR, GS,
JACS and AJS are all authors on papers proposing tests for funnel plot
asymmetry, but have no commercial interests in the use of these tests.
All authors have completed the ICJME unified disclosure form at www.
icmje.org/coi_disclosure.pdf (available on request from the corresponding
author) and declare that they have no financial or non-financial interests
that may be relevant to the submitted work.
Provenance and peer review: Not commissioned; externally peer
reviewed.
1 Egger M, Davey Smith G, Schneider M, Minder C. Bias in meta-analysis detected by a
simple, graphical test. BMJ 1997;315:629-34.
2 Sterne JAC, Gavaghan D, Egger M. Publication and related bias in meta-analysis: power
of statistical tests and prevalence in the literature. J Clin Epidemiol 2000;53:1119-29.
3 Lau J, Ioannidis JP, Terrin N, Schmid CH, Olkin I. The case of the misleading funnel plot.
BMJ 2006;333:597-600.
4 Deeks JJ, Macaskill P, Irwig L. The performance of tests of publication bias and other
sample size effects in systematic reviews of diagnostic test accuracy was assessed. J
Clin Epidemiol 2005;58:882-93.
5 Terrin N, Schmid CH, Lau J. In an empirical evaluation of the funnel plot, researchers
could not visually identify publication bias. J Clin Epidemiol 2005;58:894-901.
6 Ioannidis JP. Interpretation of tests of heterogeneity and bias in meta-analysis. J Eval
Clin Pract 2008;14:951-7.
7 Sterne JAC, Egger M, Moher D. Addressing reporting biases. In: Higgins JPT, Green S,
eds. Cochrane handbook for systematic reviews of interventions . Wiley, 2008.
8 Sterne JAC, Egger M. Funnel plots for detecting bias in meta-analysis: guidelines on
choice of axis. J Clin Epidemiol 2001;54:1046-55.
9 Begg CB, Berlin JA. Publication bias: a problem in interpreting medical data. J R Statist
Soc A 1988;151:419-63.
10 Davey Smith G, Egger M. Who benefits from medical interventions? Treating low risk
patients can be a high risk strategy. BMJ 1994;308:72-4.
11 Stuck AE, Siu AL, Wieland GD, Adams J, Rubenstein LZ. Comprehensive geriatric
assessment: a meta-analysis of controlled trials. Lancet 1993;342:1032-6.
12 Peters JL, Sutton AJ, Jones DR, Abrams KR, Rushton L, Moreno SG. Assessing
publication bias in meta-analyses in the presence of between-study heterogeneity. J R
Statist Soc A 2010;173:575-91.
13 Egger M, Jüni P, Bartlett C, Holenstein F, Sterne J. How important are comprehensive
literature searches and the assessment of trial quality in systematic reviews? Empirical
study. Health Technol Assess 2003;7:1-68.
14 Ioannidis JP. Why most discovered true associations are inflated. Epidemiology
2008;19:640-8.
15 Wood L, Egger M, Gluud LL, Schulz KF, Jüni P, Altman DG, et al. Empirical evidence of
bias in treatment effect estimates in controlled trials with different interventions and
outcomes: meta-epidemiological study. BMJ 2008;336:601-5.
16 Hopewell S, Clarke M, Stewart L, Tierney J. Time to publication for results of clinical trials.
Cochrane Database Syst Rev 2007;2:MR000011.
17 Hopewell S, Loudon K, Clarke MJ, Oxman AD, Dickersin K. Publication bias in clinical
trials due to statistical significance or direction of trial results. Cochrane Database Syst
Rev 2009;1:MR000006.
18 Song F, Parekh S, Hooper L, Loke YK, Ryder J, Sutton J, et al. Dissemination and
publication of research findings: an updated review of related biases. Health Technol
Assess 2010;14:iii,ix-iii,193.
19 Dwan K, Altman DG, Arnaiz JA, Bloom J, Chan AW, Cronin E, et al. Systematic review
of the empirical evidence of study publication bias and outcome reporting bias. PLoS
ONE 2008;3:e3081.
20 Turner EH, Matthews AM, Linardatos E, Tell RA, Rosenthal R. Selective publication of
antidepressant trials and its influence on apparent efficacy. N Engl J Med 2008;358:252-60.
21 Rosenthal R. The “file drawer” problem and tolerance for null results. Psychol Bull
1979;86:638-41.
22 Chan AW, Hrobjartsson A, Haahr MT, Gotzsche PC, Altman DG. Empirical evidence for
selective reporting of outcomes in randomized trials: comparison of protocols to published
articles. JAMA 2004;291:2457-65.
23 Higgins JP, Thompson SG. Controlling the risk of spurious findings from meta-regression.
Stat Med 2004;23:1663-82.
24 Ioannidis JP. Effectiveness of antidepressants: an evidence myth constructed from a
thousand randomized trials? Philos Ethics Humanit Med 2008;3:14.
25 Peters J, Sutton AJ, Jones DR, Abrams KR, Rushton L. Contour-enhanced meta-analysis
funnel plots help distinguish publication bias from other causes of asymmetry. J Clin
Epidemiol 2008;61:991-6.
26 Harbord RM, Egger M, Sterne JA. A modified test for small-study effects in meta-analyses
of controlled trials with binary endpoints. Stat Med 2006;25:3443-57.
27 Peters JL, Sutton AJ, Jones DR, Abrams KR, Rushton L. Comparison of two methods to
detect publication bias in meta-analysis. JAMA 2006;295:676-80.
28 Rücker G, Schwarzer G, Carpenter J. Arcsine test for publication bias in meta-analyses
with binary outcomes. Stat Med 2008;27:746-63.
29 Ioannidis JP, Trikalinos TA. The appropriateness of asymmetry tests for publication bias
in meta-analyses: a large survey. CMAJ 2007;176:1091-6.
30 Poole C, Greenland S. Random-effects meta-analyses are not always conservative. Am
J Epidemiol 1999;150:469-75.
31 Egger M, Davey Smith G. Misleading meta-analysis. Lessons from an “effective, safe,
simple” intervention that wasn’t. BMJ 1995;310:752-4.
32 Teo KK, Yusuf S, Collins R, Held PH, Peto R. Effects of intravenous magnesium in
suspected acute myocardial infarction: overview of randomised trials. BMJ
1991;303:1499-503.
33 ISIS-4 (Fourth International Study of Infarct Survival) Collaborative Group. ISIS-4: a
randomised factorial trial assessing early oral captopril, oral mononitrate, and intravenous
magnesium sulphate in 58,050 patients with suspected acute myocardial infarction. Lancet
1995;345:669-85.
34 Early administration of intravenous magnesium to high-risk patients with acute myocardial
infarction in the Magnesium in Coronaries (MAGIC) Trial: a randomised controlled trial.
Lancet 2002;360:1189-96.
35 Shang A, Huwiler-Muntener K, Nartey L, Jüni P, Dörig S, Stene JA, et al. Are the clinical
effects of homoeopathy placebo effects? Comparative study of placebo-controlled trials
of homoeopathy and allopathy. Lancet 2005;366:726-32.
36 Moreno SG, Sutton AJ, Ades AE, Stanley TD, Abrams KR, Peters JL, et al. Assessment
of regression-based methods to adjust for publication bias through a comprehensive
simulation study. BMC Med Res Methodol 2009;9:2.
37 Rucker G, Schwarzer G, Carpenter JR, Binder H, Schumacher M. Treatment-effect
estimates adjusted for small-study effects via a limit meta-analysis. Biostatistics
2011;12:122-42.
38 Moreno SG, Sutton AJ, Turner EH, Abrams KR, Cooper NJ, Palmer TM, et al. Novel
methods to deal with publication biases: secondary analysis of antidepressant trials in
the FDA trial registry database and related journal publications. BMJ 2009;339:b2981.
39 Eyding D, Lelgemann M, Grouven U, Härter M, Kromp M, Kaiser T, et al. Reboxetine for
acute treatment of major depression: systematic review and meta-analysis of published
and unpublished placebo and selective serotonin reuptake inhibitor controlled trials. BMJ
2010;341:c4737.
40 Kirkham JJ, Dwan KM, Altman DG, Gamble C, Dodd S, Smyth R, et al. The impact of
outcome reporting bias in randomised controlled trials on a cohort of systematic reviews.
BMJ 2010;340:c365.
41 Egger M, Schneider M, Davey Smith G. Spurious precision? Meta-analysis of observational
studies. BMJ 1998;316:140-4.
Accepted: 21 February 2011
Cite this as: BMJ 2011;342:d4002
Reprints: http://journals.bmj.com/cgi/reprintform Subscribe: http://resources.bmj.com/bmj/subscribers/how-to-subscribe
BMJ 2011;342:d4002 doi: 10.1136/bmj.d4002 Page 5 of 8
RESEARCH METHODS & REPORTING
Summary points
Inferences on the presence of bias or heterogeneity should consider different causes of funnel plot asymmetry and
should not be based on visual inspection of funnel plots alone
They should be informed by contextual factors, including the plausibility of publication bias as an explanation for the
asymmetry
Testing for funnel plot asymmetry should follow the recommendations detailed in this article
The fixed and random effects estimates of the intervention effect should be compared when funnel plot asymmetry
exists in a meta-analysis with between study heterogeneity
Figures
Fig 1 Example of symmetrical funnel plot. The outer dashed lines indicate the triangular region within which 95% of studies
are expected to lie in the absence of both biases and heterogeneity (fixed effect summary log odds ratio±1.96×standard
error of summary log odds ratio). The solid vertical line corresponds to no intervention effect
Fig 2 Illustration of funnel plot asymmetry due to heterogeneity, in the form of three distinct subgroups of studies. Funnel
plot including all studies (top left) shows clear asymmetry (P<0.001 from Egger test for funnel plot asymmetry). P values
for each subgroup are all >0.49.
Reprints: http://journals.bmj.com/cgi/reprintform Subscribe: http://resources.bmj.com/bmj/subscribers/how-to-subscribe
BMJ 2011;342:d4002 doi: 10.1136/bmj.d4002 Page 6 of 8
RESEARCH METHODS & REPORTING
Fig 3 Contour enhanced funnel plots. In the top diagram there is a suggestion of missing studies in the middle and right of
the plot, broadly in the white area of non-significance, making publication bias plausible. In the bottom diagram there is a
suggestion of missing studies on the bottom left hand side of the plot. Since most of this area contains regions of high
significance, publication bias is unlikely to be the underlying cause of asymmetry
Fig 4 Contour enhanced funnel plot for trials of the effect of intravenous magnesium on mortality after myocardial infarction
Reprints: http://journals.bmj.com/cgi/reprintform Subscribe: http://resources.bmj.com/bmj/subscribers/how-to-subscribe
BMJ 2011;342:d4002 doi: 10.1136/bmj.d4002 Page 7 of 8
RESEARCH METHODS & REPORTING
Fig 5 Comparison of fixed and random effects meta-analytical estimates of the effect of intravenous magnesium on mortality
after myocardial infarction
Reprints: http://journals.bmj.com/cgi/reprintform Subscribe: http://resources.bmj.com/bmj/subscribers/how-to-subscribe
BMJ 2011;342:d4002 doi: 10.1136/bmj.d4002 Page 8 of 8
RESEARCH METHODS & REPORTING
... It is also difficult to accurately determine the presence or absence of asymmetry or bias, making visual interpretation challenging. Furthermore, the results tend to depend on specific conditions and contexts, making it difficult to draw broad conclusions and limiting the generalizability of the results [13,14]. Therefore, we decided not to evaluate the risk of bias, as we believe that the potential error in assessing the risk of bias would be greater. ...
... Small study effects (potential publication bias) will be assessed by examining asymmetry in contour-enhanced funnel plots with Egger's test of the intercept for pairwise comparisons including ten or more trials [30,31,74]. The analyses will include both studies for which IPD were versus were not available and will, therefore, also be based on the effect size data extracted from the publications [31]. ...
Article
Full-text available
Background Various treatments are recommended as first-line options in practice guidelines for depression, but it is unclear which is most efficacious for a given person. Accurate individualized predictions of relative treatment effects are needed to optimize treatment recommendations for depression and reduce this disorder’s vast personal and societal costs. Aims We describe the protocol for a systematic review and individual participant data (IPD) network meta-analysis (NMA) to inform personalized treatment selection among five major empirically-supported depression treatments. Method We will use the METASPY database to identify randomized clinical trials that compare two or more of five treatments for adult depression: antidepressant medication, cognitive therapy, behavioral activation, interpersonal psychotherapy, and psychodynamic therapy. We will request IPD from identified studies. We will conduct an IPD-NMA and develop a multivariable prediction model that estimates individualized relative treatment effects from demographic, clinical, and psychological participant characteristics. Depressive symptom level at treatment completion will constitute the primary outcome. We will evaluate this model using a range of measures for discrimination and calibration, and examine its potential generalizability using internal-external cross-validation. Conclusions We describe a state-of-the-art method to predict personalized treatment effects based on IPD from multiple trials. The resulting prediction model will need prospective evaluation in mental health care for its potential to inform shared decision-making. This study will result in a unique database of IPD from randomized clinical trials around the world covering five widely used depression treatments, available for future research.
... We had also performed subgroup analysis to explore the diversity among different studies and source of heterogeneity. Publication bias was assessed using Egger's test and funnel plots [23,24] in Stata 12.0. ...
Article
Full-text available
Background In medical education, Lecture Based Learning (LBL) is the most common way of disseminating information. Team Based Learning (TBL), a new teaching method, is a teacher-guided method that employs teams in a class, showing suitability for medical education. Two teaching methods represent distinct educational approaches, each with its own set of advantages and limitations. In this study, we performed a systemic review on the efficacy of TBL pedagogy in medical education. Methods MEDLINE, EMBASE, the Cochrane Library, and Web of Science database were searched through July 2022. Standard mean difference (SMD) and 95% confidence intervals (CIs) were calculated. Results The analysis included 33 studies. Our analysis revealed that students utilizing the TBL method exhibited significantly higher pre-( SMD = 0.51, 95%CI 0.11 to 0.92) /post-test (SMD = 0.96, 95%CI 0.70 to 1.22) scores than students with LBL. Students in TBL classes had better development of scores, retention (SMD = 1.03, 95%CI 0.38 to 1.69), engagement (SMD = 2.26, 95%CI 0.23 to 4.29) and higher satisfactory rate (SMD = 1.08, 95%CI 0.87 to 1.29). However, students required more time to independently complete reading materials and preparatory tasks. Conclusion Our study indicates the gratifying effectiveness of TBL application in medical education. TBL pedagogy is compatible with the present medical education and should be generalized in more classrooms.
... Heterogeneity was considered significant if the Cochran Q test yielded a p value below 0.05 or if the I 2 exceeded 50%. Standard funnel plots and Egger's regression test p value were also calculate to confirm the findings from the Doi plots and LFK index [19]. Moreover, a leave-one-out sensitivity analysis was performed by systematically excluding individual studies to observe any shifts in pooled estimates or heterogeneity, to identify particular studies contributing to heterogeneity. ...
Article
Full-text available
Background/Objective: This study aimed to determine the clinicopathological findings and prognostic value of chemokine receptor 7 (CCR7) expression in patients with breast cancer (BC). Methods: Up to the 25th of March 2025, a search was conducted using five databases: PubMed, Embase, Scopus, Medline, and Web of Science. The methodological standards for the epidemiological research scale were used to assess the quality of the included articles, and Stata software (Stata 19) was used to synthesize the meta-analysis. Results: We considered 12 of 853 studies that included 3119 patients with BC. High CCR7 expression was not associated with age (odds ratio [OR] 0.82, 95% confidence interval [CI] 0.66–1.03); clinicopathological findings, including tumor size (OR 1.062, 95% CI 0.630–1.791); clinical stage (OR 1.753, 95% CI 0.231–13.304); nodal metastasis (OR 1.252, 95% CI 0.571–2.741); or histological differentiation (OR 1.167, 95% CI 0.939–1.450). CCR7 expression did not affect overall survival (hazard ratio 0.996, 95% CI 0.659–1.505). Conclusions: Our quantitative analysis did not reveal an association between CCR7 expression and poor clinicopathological or prognostic features in BC patients. Because of the high heterogeneity and potential publication bias, large high-quality studies are required to further confirm these findings.
Article
Full-text available
Aims The aim of this study is to investigate the prognostic role of T1 mapping techniques in predicting major adverse cardiovascular events (MACE) in patients affected by non‐ischaemic dilated cardiomyopathy (NIDCM) by performing a meta‐analysis of available studies. Methods and results Data from 12 observational studies exploring the prognostic role of native T1 mapping and extracellular volume (ECV) were analysed with random effect generic inverse variance. The primary endpoint was MACE defined as a composite of heart failure or arrhythmic‐related events, expressed as hazard ratio (HR) with 95% confidence interval (CI). Secondary main outcomes were heart failure‐related events, arrhythmic‐related events, and weighted mean difference of native T1 mapping values or ECVs between patients with or without MACE. Overall, 4025 patients with NIDCM were included. The median follow‐up length was 22 (IQR 14–22) months. The primary outcome of MACE occurred in 610 patients with a pooled HR for native T1 mapping values of 1.07 (95% CI 1.04–1.09, I² 31.5%) and a pooled HR for ECV of 1.37 (95% CI 1.29–1.44, I² 0%). HF‐related events occurred in 492 patients, with a pooled HR for T1 mapping of 1.05 (95% CI 1.03–1.07, I² 1%) and a pooled HR for ECVs of 1.43 (95% CI 1.25–1.61, I² 63%). Arrhythmic‐related events occurred in 118 patients, with a pooled HR for T1 mapping values of 1.09 (95% CI 1.07–1.12, I² 0%). The weighted mean difference of native T1 mapping between patients with and without MACE was 30.91 (95% CI 18.45–43.16, I² 16.72%), while the mean difference of ECV was 4.52 (95% CI 2.78–6.26, I² 86%). Conclusions In NIDCM patients, native T1 mapping and ECV were associated with increased risk of the composite primary endpoint of MACE and the secondary endpoint of heart failure and arrhythmic‐related events.
Article
Full-text available
Background: Hypercholesterolemia is a risk factor for cardiovascular disease that can cause death and disability, so it must be managed properly. Garlic flour (GP) is one of the dosage forms of garlic that has the potential to reduce cholesterol. This study is a systematic review and meta-analysis that aims to determine the effectiveness of GP in reducing total cholesterol (TC) and low density lipoprotein (LDL).Subjects and Method: Search articles through the Pubmed database, Google scholar, Proquest, Springer link, and Science direct with the appropriate keywords. Population= patients with hyper­cholesterolemia. Intervention= GP. Comparison= placebo. Outcome= TC and LDL. Inclusion crite­ria included complete articles in English or Indonesian published from 2010-2022, Randomized Con­trolled Trial (RCT) design, subjects >18 years old with hypercholesterolemia, the study had a complete fat profile outcome, duration of intervention 2 weeks-12 months. The flow of article selection is based on the PRISMA flow diagram. Meta-analysis using RevMan 5.3 with effect size standardized mean difference (SMD) and Random Effect Model (REM) analysis model.Results: A total of 10 articles were eligible for a meta-analysis. The GP group had TC levels 1.49 units lower than the placebo group (SMD = -1.49; 95% CI = -1.95 to -0.55; p = 0.005) and LDL levels 0.68 units lower than the placebo group (SMD= -0.68; CI95). %= -1.08 to -0.27; p=0.001). Subgroup analysis based on the length of the intervention showed that administration of GP <12 weeks only significantly reduced TC while administration > 12 weeks significantly reduced TC and LDL.Conclusion: GP is effective in lowering TC and LDL especially at 12 weeks or more.Keywords: garlic powder, hypercholesterolemia, RCTCorrespondence: Zuraida Zulkarnain. Masters Program in Public Health, Universitas Sebelas Maret. Jl. Ir. Sutami 36A, Surakarta, Central Java 57126. Mobile: 081393933862. Email: zuraida.zu@gmail.com.Indonesian Journal of Medicine (2022), 07(01): 102-114https://doi.org/10.26911/theijmed.2022.07.01.11
Chapter
This book presents a system of critical appraisal applicable to clinical, epidemiological and public health studies and to many other fields. It assumes no prior knowledge. The methods are relevant to students, practitioners and policymakers. The book shows how to assess if the results of one study or of many studies show a causal effect. The book discusses study designs: randomised and non-randomised trials, cohort studies, case-control studies, and surveys, showing the presentation of results including person-time and survival analysis, and issues in the selection of subjects. The system shows how to describe a study, how to detect and assess selection biases, observation bias, confounding, and chance variation, and how to assess internal validity and external validity (generalisability). Statistical methods are presented assuming no previous knowledge, and showing applications to each study design. Positive features of causation including strength, dose-response, and consistency are discussed. The book shows how to do systematic reviews and meta-analyses, and discusses publication bias. Systems of assessing all evidence are shown, leading to a general method of critical appraisal based on 20 key questions in five groups, which can be applied to any type of study or any topic. Six chapters show the application of this method to randomised trials, prospective and retrospective cohort studies, and case-control studies. An appendix summarises key statistical methods, each with a worked example. Each main chapter has self-test questions, with answers provided.
Book
Full-text available
The revised edition of the Handbook offers the only guide on how to conduct, report and maintain a Cochrane Review. The second edition of The Cochrane Handbook for Systematic Reviews of Interventions contains essential guidance for preparing and maintaining Cochrane Reviews of the effects of health interventions. Designed to be an accessible resource, the Handbook will also be of interest to anyone undertaking systematic reviews of interventions outside Cochrane, and many of the principles and methods presented are appropriate for systematic reviews addressing research questions other than effects of interventions. This fully updated edition contains extensive new material on systematic review methods addressing a wide-range of topics including network meta-analysis, equity, complex interventions, narrative synthesis, and automation. Also new to this edition, integrated throughout the Handbook, is the set of standards Cochrane expects its reviews to meet. Written for review authors, editors, trainers and others with an interest in Cochrane Reviews, the second edition of The Cochrane Handbook for Systematic Reviews of Interventions continues to offer an invaluable resource for understanding the role of systematic reviews, critically appraising health research studies and conducting reviews.
Article
Full-text available
Background: The tendency for authors to submit, and of journals to accept, manuscripts for publication based on the direction or strength of the study findings has been termed publication bias. Objectives: To assess the extent to which publication of a cohort of clinical trials is influenced by the statistical significance, perceived importance, or direction of their results. Search strategy: We searched the Cochrane Methodology Register (The Cochrane Library [Online] Issue 2, 2007), MEDLINE (1950 to March Week 2 2007), EMBASE (1980 to Week 11 2007) and Ovid MEDLINE In-Process & Other Non-Indexed Citations (March 21 2007). We also searched the Science Citation Index (April 2007), checked reference lists of relevant articles and contacted researchers to identify additional studies. Selection criteria: Studies containing analyses of the association between publication and the statistical significance or direction of the results (trial findings), for a cohort of registered clinical trials. Data collection and analysis: Two authors independently extracted data. We classified findings as either positive (defined as results classified by the investigators as statistically significant (P < 0.05), or perceived as striking or important, or showing a positive direction of effect) or negative (findings that were not statistically significant (P >/= 0.05), or perceived as unimportant, or showing a negative or null direction in effect). We extracted information on other potential risk factors for failure to publish, when these data were available. Main results: Five studies were included. Trials with positive findings were more likely to be published than trials with negative or null findings (odds ratio 3.90; 95% confidence interval 2.68 to 5.68). This corresponds to a risk ratio of 1.78 (95% CI 1.58 to 1.95), assuming that 41% of negative trials are published (the median among the included studies, range = 11% to 85%). In absolute terms, this means that if 41% of negative trials are published, we would expect that 73% of positive trials would be published.Two studies assessed time to publication and showed that trials with positive findings tended to be published after four to five years compared to those with negative findings, which were published after six to eight years. Three studies found no statistically significant association between sample size and publication. One study found no significant association between either funding mechanism, investigator rank, or sex and publication. Authors' conclusions: Trials with positive findings are published more often, and more quickly, than trials with negative findings.
Article
Full-text available
Publication and selection biases in meta-analysis are more likely to affect small studies, which also tend to be of lower methodological quality. This may lead to “small-study effects,” where the smaller studies in a meta-analysis show larger treatment effects. Small-study effects may also arise because of between-trial heterogeneity. Statistical tests for small-study effects have been proposed, but their validity has been questioned. A set of typical meta-analyses containing 5, 10, 20, and 30 trials was defined based on the characteristics of 78 published meta-analyses identified in a hand search of eight journals from 1993 to 1997. Simulations were performed to assess the power of a weighted regression method and a rank correlation test in the presence of no bias, moderate bias or severe bias. We based evidence of small-study effects on P
Article
Full-text available
For any given research area, one cannot tell how many studies have been conducted but never reported. The extreme view of the "file drawer problem" is that journals are filled with the 5% of the studies that show Type I errors, while the file drawers are filled with the 95% of the studies that show nonsignificant results. Quantitative procedures for computing the tolerance for filed and future null results are reported and illustrated, and the implications are discussed. (15 ref) (PsycINFO Database Record (c) 2012 APA, all rights reserved)
Article
58,050 patients entering 1086 hospitals up to 24 h (median 8 h) after the onset of suspected acute myocardial infarction (MI) with no clear contraindications to the study treatments (in particular, no cardiogenic shock or persistent severe hypotension) were randomised in a "2 x 2 x 2 factorial" study. The treatment comparisons were: (i) 1 month of oral captopril (6.25 mg initial dose titrated up to 50 mg twice daily) versus matching placebo; (ii) 1 month of oral controlled-release mononitrate (30 mg initial dose titrated up to 60 mg once daily) versus matching placebo; and (iii) 24 h of intravenous magnesium sulphate (8 mmol initial bolus followed by 72 mmol) versus open control. There were no significant "interactions" between the effects of these three treatments, and the results for each are based on the randomised comparison of about 29,000 active versus 29,000 control allocated patients. Captopril There was a significant 7% (SD 3) proportional reduction in 5-week mortality (2088 [7.19%] captopril-allocated deaths vs 2231 [7.69%] placebo; 2p = 0.02), which corresponds to an absolute difference of 4.9 SD 2.2 fewer deaths per 1000 patients treated for 1 month. The absolute benefits appeared to be larger (perhaps about 10 fewer deaths per 1000) in certain higher-risk groups, such as those presenting with a history of previous MI or with heart failure. The survival advantage appeared to be maintained in the longer term (5.4 [SD 2.8] fewer deaths per 1000 at 12 months). Captopril was associated with an increase of 52 (SD 2) patients per 1000 in hypotension considered severe enough to require termination of study treatment, of 5 (SD 2) per 1000 in reported cardiogenic shock, and of 5 (SD 1) per 1000 in some degree of renal dysfunction. It produced no excess of deaths on days 0-1, even among patients with low blood pressure at entry. Mononitrate There was no significant reduction in 5-week mortality, either overall (2129 [7.34%] mononitrate-allocated deaths vs 2190 [7.54%] placebo) or in any subgroup examined (including those receiving short-term non-study intravenous or oral nitrates at entry). Further follow-up did not indicate any later survival advantage. The only significant side-effect of the mononitrate regimen studied was an increase of 15 (SD 2) per 1000 in hypotension. Those allocated active treatment had somewhat fewer deaths on days 0-1, which is reassuring a bout the safety of using nitrates early in acute MI.(ABSTRACT TRUNCATED AT 400 WORDS)
Article
Publication bias, the phenomenon in which studies with positive results are more likely to be published than studies with negative results, is a serious problem in the interpretation of scientific research. Various hypothetical models have been studied which clarify the potential for bias and highlight characteristics which make a study especially susceptible to bias. Empirical investigations have supported the hypothesis that bias exists and have provided a quantitative assessment of the magnitude of the problem. The use of meta-analysis as a research tool has focused attention on the issue, since naive methodologies in this area are especially susceptible to bias. In this paper we review the available research, discuss alternative suggestions for conducting unbiased meta-analysis and suggest some scientific policy measures which could improve the quality of published data in the long term.
Article
Background The benefits of supplemental administration of intravenous magnesium in patients with ST-elevation myocardial infarction (STEMI) are controversial. Despite promising results from work in animals and the ready availability of this simple, inexpensive treatment, conflicting results have been reported in clinical trials. Our aim was to compare short-term mortality in patients with STEMI who received either intravenous magnesium sulphate or placebo. Methods We did a randomised, double-blind trial in 6213 patients with acute STEMI who were assigned a 2 g intravenous bolus of magnesium sulphate administered over 15 min, followed by a 17 g infusion of magnesium sulphate over 24 h (n=3113), or matching placebo (n=3100). Our primary endpoint was 30-day all-cause mortality. At randomisation, patients were stratified by their eligibility for reperfusion therapy. The first stratum included patients who were aged 65 years or older and eligible for reperfusion therapy, and the second stratum included patients of any age who were not eligible for reperfusion therapy. Analysis was by intention-to-treat. Findings At 30 days, 475 (15·3%) patients in the magnesium group and 472 (15·2%) in the placebo group had died (odds ratio 1·0, 95% Cl 0·9–1·2, p=0·96). No benefit or harm of magnesium was observed in eight prespecified subgroup analyses of patients and in 15 additional exploratory subgroup analyses. After adjustment for factors shown to effect mortality risk in a multivariate regression model, no benefit of magnesium was observed (1·0, 0·8–1·1, p=0·53). Interpretation Early administration of magnesium in high-risk patients with STEMI has no effect on 30-day mortality. In view of the totality of the available evidence, in current coronary care practice there is no indication for the routine administration of intravenous magnesium in patients with STEMI.