Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Systematic review finds that study data not published in full text articles have unclear impact on meta-analyses results in medical research

  • Christine M. Schmucker ,

    schmucker@cochrane.de

    Affiliation Cochrane Germany, Medical Center University of Freiburg, Faculty of Medicine, University of Freiburg, Germany

  • Anette Blümle,

    Affiliation Cochrane Germany, Medical Center University of Freiburg, Faculty of Medicine, University of Freiburg, Germany

  • Lisa K. Schell,

    Affiliation Institute for Quality and Efficiency in Health Care, Cologne, Germany

  • Guido Schwarzer,

    Affiliation Institute for Medical Biometry and Statistics, Faculty of Medicine and Medical Center University of Freiburg, Germany

  • Patrick Oeller,

    Affiliation Cochrane Germany, Medical Center University of Freiburg, Faculty of Medicine, University of Freiburg, Germany

  • Laura Cabrera,

    Affiliation Cochrane Germany, Medical Center University of Freiburg, Faculty of Medicine, University of Freiburg, Germany

  • Erik von Elm,

    Affiliation Cochrane Switzerland, Institute of Social and Preventive Medicine (IUMSP), University Hospital Lausanne, Lausanne, Switzerland

  • Matthias Briel,

    Affiliation Basel Institute for Clinical Epidemiology and Biostatistics, Department of Clinical Research, University of Basel and University Hospital Basel, Switzerland

  • Joerg J. Meerpohl,

    Affiliation Cochrane Germany, Medical Center University of Freiburg, Faculty of Medicine, University of Freiburg, Germany

  • on behalf of the OPEN consortium

    The complete membership of the OPEN consortium can be found in the Acknowledgments.

Abstract

Background

A meta-analysis as part of a systematic review aims to provide a thorough, comprehensive and unbiased statistical summary of data from the literature. However, relevant study results could be missing from a meta-analysis because of selective publication and inadequate dissemination. If missing outcome data differ systematically from published ones, a meta-analysis will be biased with an inaccurate assessment of the intervention effect. As part of the EU-funded OPEN project (www.open-project.eu) we conducted a systematic review that assessed whether the inclusion of data that were not published at all and/or published only in the grey literature influences pooled effect estimates in meta-analyses and leads to different interpretation.

Methods and findings

Systematic review of published literature (methodological research projects). Four bibliographic databases were searched up to February 2016 without restriction of publication year or language. Methodological research projects were considered eligible for inclusion if they reviewed a cohort of meta-analyses which (i) compared pooled effect estimates of meta-analyses of health care interventions according to publication status of data or (ii) examined whether the inclusion of unpublished or grey literature data impacts the result of a meta-analysis.

Seven methodological research projects including 187 meta-analyses comparing pooled treatment effect estimates according to different publication status were identified. Two research projects showed that published data showed larger pooled treatment effects in favour of the intervention than unpublished or grey literature data (Ratio of ORs 1.15, 95% CI 1.04–1.28 and 1.34, 95% CI 1.09–1.66). In the remaining research projects pooled effect estimates and/or overall findings were not significantly changed by the inclusion of unpublished and/or grey literature data. The precision of the pooled estimate was increased with narrower 95% confidence interval.

Conclusions

Although we may anticipate that systematic reviews and meta-analyses not including unpublished or grey literature study results are likely to overestimate the treatment effects, current empirical research shows that this is only the case in a minority of reviews. Therefore, currently, a meta-analyst should particularly consider time, effort and costs when adding such data to their analysis. Future research is needed to identify which reviews may benefit most from including unpublished or grey data.

Introduction

A meta-analysis as part of a systematic review aims to provide a thorough, comprehensive and unbiased statistical summary of data from the literature.[1] However, relevant study-results could be missing from a meta-analysis because of selective publication and inadequate dissemination (non-dissemination or insufficient dissemination). Even the most comprehensive searches are likely to miss study data which are not published at all such as supplemental unpublished data related to published trials, data obtained from the Food and Drug Administration (FDA) or other regulatory websites or postmarketing analyses hidden from the public. In addition, study data that are not published in conventional journals and, therefore, are not indexed in electronic databases are also likely to be not identified. This so called ‘grey literature’ is not controlled by commercial or academic publishers. It includes non-indexed conference abstracts frequently published in journal collections, dissertations, press releases, government reports, policy documents, book chapters or data obtained from trial registers (Table 1). If the results from missing study data (unpublished and/or study data published in the grey literature) differ systematically from the published data available, a meta-analysis may become biased with an inaccurate assessment of the intervention effect.[24]

thumbnail
Table 1. Definitions of unpublished, grey and published study data.

https://doi.org/10.1371/journal.pone.0176210.t001

There is some evidence that indicates that published randomized controlled trials tend to be larger and show an overall greater treatment effect in favor of the intervention than grey literature trials or unpublished data.[58] However, the identification of relevant unpublished study data or data published in the grey literature and their inclusion in meta-analyses can be particularly challenging regarding excessive time, effort and costs. There is also some controversy regarding whether unpublished study data and data published in the grey literature should be included in meta-analyses at all, because they are generally not peer reviewed and their internal validity (risk of bias) may be difficult to assess due to poor reporting of the trials. On the other hand, particularly conference proceedings may take a separate role in the grey literature as they often provide preliminary results or results following intermediate follow-up. A publication by Cook and colleagues showed that 78% of authors of meta-analyses felt that unpublished studies should be included in meta-analyses compared to only 47% of journal editors.[9] Therefore, research is needed to assess the potential impact of inclusion of ‘grey literature’ study data and unpublished data in meta-analyses of health care interventions.

We investigated the impact of study data that were not published in full text articles in scientific journals on pooled effect estimates and the overall interpretation of meta-analyses.

The current review was part of the EU-funded OPEN project (To Overcome failure to Publish nEgative fiNdings; www.open-project.eu) which aimed to investigate non-publication of study data and related dissemination bias through a series of systematic reviews[1014] following a protocol published previously.[15]

Methods

Systematic literature search

We initially searched Medline (Ovid), Embase (Ovid), The Cochrane Library and Web of Science from inception until February 2012. An update search was performed in February 2016. The search strategy was based on combinations of medical subject headings (MeSH) and keywords and was not restricted to specific languages or years of publication. The search strategy used in Medline (Ovid) is presented in S1 Search Strategy. Search strategies for other databases were modified to meet the requirements of each database. The searches were supplemented by checking the bibliographies of any eligible articles for additional references.

Patient involvement

This research is based on empirical work. Therefore, there was no patient involvement in this methodological systematic review of reviews (so called umbrella review).

Study selection

Titles and abstracts were reviewed using pre-defined inclusion criteria. Full papers of all methodological research projects which included a cohort of meta-analyses (i.e., more than one meta-analysis) and (i) compared pooled effect estimates of meta-analyses of health care interventions according to publication status (i.e., published vs. unpublished and/or grey study data) and/or (ii) examined whether the inclusion of unpublished and/or grey study data impact the overall findings of a meta-analysis (i.e., from negatively significant to positively significant; from not clinically relevant to clinically relevant) were obtained for detailed evaluation.

All stages of study selection, data extraction and quality assessment were done independently by two reviewers (study selection and data extraction: PO and LC, quality assessment: CS and LKS). Any disagreement during the selection, extraction, and assessment process were resolved by discussion and consensus or with help of a third reviewer (JJM).

We considered a study ‘published’ when it appeared in a peer-reviewed journal. The definition of unpublished and/or grey literature study data had to be in accordance with the definition of ‘unpublished studies’ and ‘grey literature’ described above (see Introduction).

A meta-analysis was defined as mathematical calculation of a weighted summary estimate of a treatment effect by pooling results of two or more studies.

Outcomes

First, we focused on the extent to which the pooled effect estimate in a meta-analysis changes with the inclusion of unpublished and/or study data published in the grey literature in comparison to published study data. Where possible, we calculated as our primary study outcome a ratio of risk ratios (RRR) or odds ratios (ROR) between the results of published data and the results of unpublished and/or grey literature data and estimate the percentage change (pooled risk ratio from published data divided by pooled risk ratio from unpublished data and/or grey literature data).[15] Thereby, a ratio greater than 1.0 would indicate that published study data showed a greater treatment effect; likewise a ratio below 1.0 would indicate that published data would show a smaller treatment effect. We also intended to calculate a single weighted pooled RRR or ROR to combine ratios from the different methodological research projects to estimate an overall pooled effect, which also takes into account factors such as number of studies, patients and events. For the intended analyses (to calculate a ratio of risk or odds ratios (RRR, ROR) between the results of published study data and unpublished and/or grey study data), the single effect estimates (RR, OR) estimated by the included meta-analyses would be the ‘unit of analyses’.

Second, we aimed to investigate the impact of the inclusion of unpublished or grey literature study data on the interpretation of meta-analyses. This impact can be estimated by calculating the proportion of meta-analyses which showed a change in their interpretation (e.g., from negatively significant to positively significant; from not clinically relevant to clinically relevant).[15]

Data extraction

We extracted main characteristics of (i) the methodological research projects (e.g., baseline data, area of health care, number of meta-analyses included); (ii) the meta-analyses (e.g., purpose and scope of meta-analyses, number of studies and participants included); and (iii) the studies included in meta-analyses (e.g., number of studies depending on publication status). For more detail see our published protocol.[15]

Assessment of risk of bias and generalizability of results

No quality assessment tool exists for these types of methodological research projects. Risk of bias (internal validity) and generalizability (external validity) were therefore assessed according to pre-defined criteria which were developed considering empirical evidence on dissemination bias[10, 16] and internal discussion.[15] The assessment of risk of bias was based on (i) the selection process; i.e., whether and to which extent the search criteria were reported to identify unpublished and/or grey and published study data; (ii) definition of the publication status; i.e., whether explicit criteria were reported for the definition of unpublished or grey literature and published data; (iii) role of confounding factors; i.e., whether the difference of the results between unpublished/grey and published study data may be explained by differences in study designs, type of participants or intervention characteristics and not by a true difference in the results between unpublished/grey literature and published data; therefore we investigated whether analyses were stratified or results adjusted for possible confounders. In addition, we investigated the reliability of the data extraction process; i.e., whether data extraction was performed by two researchers independently. Generalizability assessment was based on (i) the status of the sample of meta-analyses included; i.e., whether a random, consecutive or selected sample was included and (ii) whether the research project selected a broad-ranging sample of meta-analyses that presents the current literature in the field of interest (e.g., in terms of size or diversity of topic).

For data extraction and risk of bias assessment, we relied on information provided in publications of the methodological research projects.

Statistical analysis and data synthesis

The sparse data did not allow us to apply the predefined statistical analyses neither for the main analysis nor the subgroup analyses.[15] Instead, results of this systematic review are presented descriptively using text and tables.

Results

Literature search and selection process

The searches identified 8464 citations, including 3301 duplicates (Fig 1). Among the 5163 unique references screened, 10 references[3, 1725] corresponding to 7 methodological research projects[3, 1924] were eligible for inclusion in this systematic review.

thumbnail
Fig 1. PRISMA statement flow diagram.

(as published by Moher D et al in BMJ 2009;339:b2535).

https://doi.org/10.1371/journal.pone.0176210.g001

Characteristics of included research projects

Main characteristics of the 7 research projects are presented in Tables 2 and 3. In brief, 5 research projects included conventional intervention reviews[3, 2023], 1 research project was solely based on safety aspects,[24] while another research project included individual participant data meta-analyses.[19] Different medical specialties were displayed in 4 research projects[3, 2224], while 3 focused on a single medical field.[1921] In total, 187 meta-analyses with 1617 primary studies (373 unpublished/grey literature studies and 1244 published studies) enrolling a total of 428762 participants (58786 participants in unpublished/grey literature studies and 369976 in published studies) were included. It has to be taken into account that the given numbers of included studies and participants are underestimated because Hart et al[23] and Golder et al[24] did not provide these study characteristics in detail. The publication dates of the latest meta-analyses included in the research projects ranged between 1995[3] and 2014.[24]

thumbnail
Table 2. Main characteristics of the included methodological research projects.

https://doi.org/10.1371/journal.pone.0176210.t002

thumbnail
Table 3. Main characteristics of studies included in analyses.

https://doi.org/10.1371/journal.pone.0176210.t003

Assessment of risk of bias and generalizability of results

Table 4 presents the assessment of risk of bias and generalizability of results for each research project. Regarding risk of bias, each research project reported how unpublished or grey literature study data were identified within meta-analyses. Unpublished or grey literature data (e.g., in terms of conference abstracts, dissertations or editorials) were sufficiently defined in all research projects. The main limitation of the research projects was that most of them (except for Golder et al[24]) did not allow us to judge whether grey literature or unpublished study data in comparison to published data were adequately matched (e.g., in terms of study aim or sample size) or adjusted for confounders.

thumbnail
Table 4. Risk of bias and generalizability of included methodological research projects.

https://doi.org/10.1371/journal.pone.0176210.t004

Generalizability of results was low or unclear in four research projects.[3, 19, 20, 24] It means that the results of these research projects were either based on a selected sample of meta-analyses (e.g., meta-analyses from one research group only were used) or the medical field of interest was not sufficiently represented (e.g., only few rare sorts of cancers or a small range of interventions were considered).

Effect of unpublished or grey literature study data on pooled estimates in meta-analyses

The effects of unpublished or grey literature studies on pooled estimates in meta-analyses are shown in Table 5. One research project (including 467 randomized controlled trials) showed that published studies had a larger pooled treatment effect in favor of the intervention than unpublished studies (ROR 1.15, 95% CI 1.04–1.28).[3] In the remaining research projects pooled effect estimates were not significantly changed by the inclusion of unpublished or grey literature data. However, Egger et al[22] presented the pooled effect estimate across different medical specialties (ROR 1.07, 95% CI 0.98; 1.15)–but also separated effect estimates for selected medical fields. In the field of obstetrics and gynaecology this pooled analysis showed that published results are more positive than unpublished results (ROR 1.34, 95% CI 1.09–1.66). In psychiatry there was a similar trend but pooled estimates did not reach statistical significance (ROR 1.61, 95% CI 0.9–2.9). The combination of estimates across methodological research projects was not possible due to differences in the definitions of effect estimates (some research projects reported hazard ratios, other odds ratios or risk ratios, or even weighted mean differences) and clinical heterogeneity (different aims of the research projects regarding safety and efficacy outcomes).

thumbnail
Table 5. Outcomes of the included methodological research projects.

https://doi.org/10.1371/journal.pone.0176210.t005

Impact of unpublished or grey literature study data on the interpretation of meta-analyses

Five research projects provided additional information on the overall impact of unpublished or grey literature study data on the interpretation of the results. The results are descriptively summarized in Table 5. Hart and colleagues[23] reported that the addition of unpublished data to their sample of meta-analyses caused in 46% lower, in 7% identical and in 46% greater effect estimates than published data. In the research project from Egger et al[22] removal of grey literature data resulted in a change in pooled estimates from a 28% decrease to a 24% increase in benefit. McAuley and colleagues[3] reported that removal of grey literature data changed the estimate by at least 10%. Thereby, significance of the results was affected in 3 out of 41 meta-analyses.

On the other hand, Fergusson and colleagues[21] and Golder and colleagues[24] stated that ‘effect estimates were not substantially changed’[21] or that ‘the direction and magnitude of the difference varies and is not consistent’[24] when unpublished or grey literature data are added.

Discussion

Principal findings

Although it was shown that some case samples of meta-analyses not including grey literature or unpublished data clearly overestimate treatment effects,[68] quantifying this effect by considering all meta-epidemiological studies (so called methodological research projects) reveals that this affects only a minority of reviews. In the majority of meta-analyses over a wide range of medical fields, excluding unpublished trials had no or only a small effect on the pooled estimates of treatment effects. However, in some instances more substantial, statistically significant changes were observed (overestimating the effect between 9 and 60%)[22] There may be a tendency in research areas involving new drugs or technologies to publish the most exciting and positive results more rapidly, and negative ones less quickly, if at all.[10] Also sponsorship of drug and device studies by the manufacturer leads to more favourable results and conclusions than sponsorship by other sources.[26] Consequently, the problem of dissemination bias could be more pronounced in medical areas in which relevant innovations are being developed at quick pace or when trials are published close to drug approval. This assumption, however, could not be proven with the available empirical data.

Our research and other reviews[5, 27] revealed that unpublished trials are often smaller (e.g., Table 3, differences in medians between unpublished or grey literature study data and published data: 11,[22] 534,[21] and 29[3] patients, respectively). Small sample sizes may be one of the reasons that unpublished or grey literature study data are less likely to produce statistically significant results than published data. However, if study size was the only factor impacting on the likelihood of publication this would not result in bias, but a lack of precision with wider confidence intervals of effect estimates.

Methodological research projects included in this review used different statistical methods to determine the contribution of unpublished data in meta-analyses. For example, Egger and colleagues[22] used the statistic chosen by the original reviewers of the meta-analyses to calculate pooled effect estimates separately for unpublished and published trials. Thereafter, weighted averages for all these ratios were calculated using random effects models. McAuley and colleagues[3] chose a fixed effect logistic regression model which requires individual patient data from each trial. This approach ignores heterogeneity between trials and between meta-analyses. In general, too little consideration has been given to appropriate statistical methods for this type of meta-epidemiological research so far. This may lead to an underestimation of the uncertainty of effect estimates due to unpublished data in meta-analyses.[28]

None of the methodological research projects addressed the problem of multiple journal publications.[29] Unaccounted duplicate publication may inflate the number of participants and/or events leading to increased precision and, obviously, causes dissemination bias.

Evidence from a Cochrane review has shown that only about half of all trials reported as abstracts and presented at conferences are subsequently published in full.[16] In addition, it takes, on average, three years for a trial reported as an abstract to be eventually published in a peer-reviewed journal. Therefore, excluding them seems an arbitrary act that may bias the results. On the other hand, McAuley and colleagues showed that the inclusion of abstracts had no relevant impact on pooled estimates of meta-analyses over different medical fields.[3] Moreover, concerns have been raised regarding the methodological and reporting quality in unpublished studies, because grey or unpublished literature is often not peer reviewed. We believe that abstracts may take a separate role in the grey literature as they often provide preliminary study results, results following intermediate follow-up, or unexpected findings. Consequently, when a researcher decides to include unpublished or grey literature study data in meta-analyses, it is important to run sensitivity analyses to identify possible differences between results from unpublished or grey literature studies and from fully published papers. While there is no doubt that studies that have positive results are subsequently published as full-length journal articles more often than studies with negative results,[10] lack of time of the authors may be a major reason for non-publication of research—independent of the direction of results.[30]

Strengths and weaknesses of this review

This systematic review sought to comprehensively synthesize the body of research on the impact of including unpublished study data and data published in the grey literature in meta-analyses. By discussing multiple study characteristics and potential confounders related to unpublished studies and studies published in the grey literature, we could not identify sufficient evidence to conclude whether or to which extent inclusion of unpublished and grey study data have an impact on the pooled effect estimates and the conclusions from meta-analyses. Nevertheless, the available research projects demonstrates that availability of unpublished and grey literature data leads to a more precise risk estimates with narrower 95% confidence intervals, thus representing higher evidence strength according to the GRADE evaluation (Grades of Recommendation, Assessment, Development and Evaluation).[31] In addition, we developed criteria to assess both risk of bias and generalizability for this specific type of empirical research which may be of high value in future methodological research. Our strategy was not focused on the results of single meta-analyses including published and unpublished data, but on meta-epidemiological studies. We expected that theses research projects would allows us to estimate the “average” overestimation of treatment effects due to dissemination bias.

However, we are aware that our findings have several limitations: First, we could not identify sufficient research projects to conclude whether or to which extent inclusion of unpublished and grey study data have an impact on the conclusions from meta-analyses. Second, the risk of bias assessment revealed that the internal validity may be hampered due to the lack of appropriate adjustment for potential confounders between published and unpublished or grey literature data in the identified methodological research projects. Second, our research is mainly limited to selected samples of medical literature (e.g., rare sorts of cancers or a small range of adverse effects), and hence the findings may not be generalizable to other medical fields. However, most medical fields assessed were large and permitted evaluation of a large number of studies.[19, 21] Another weakness of our study relates to its retrospective nature and its reliance on what authors described as comprehensive literature searches. We did not assess whether the sample of trials identified by these authors was in fact complete and whether searches were truly comprehensive. If searches were inadequate, so that many unpublished or grey literature studies with negative results were consciously or unconsciously omitted, then our review may underestimate the impact of dissemination bias. Roughly the same would be true, if predominantly unpublished or grey literature studies with similar results to published studies were identified by inadequate searches. But we could not judge how often this happened. On the other hand, we are concerned about the possibility of dissemination bias (in particular reporting bias), where investigators may have chosen not to write up their results (e.g., for a subgroup of patients) if they did not find any significant differences between published and unpublished study data. We believe that the impact of unpublished or grey literature data on pooled estimates could be assessed more thoroughly if the intention to compare data sources according to publication status was built in at the protocol stage of these meta-analyses.

Time, effort, and cost involved in locating and retrieving unpublished data and grey literature makes its inclusion in reviews challenging. Legal obligation to prospectively register trials and make results available after completion of the trial in many countries (including the United States and Europe), different registries for clinical trials such as the International Clinical Trials Registry Platform (ICTRP) or the database ClinicalTrials.gov, internet-based grey literature resources, journals devoted to negative trials, or efforts taken by various groups, including Cochrane (through trial registries), may further ease the identification and inclusion of unpublished data in meta-analyses.

We acknowledge that more than half of all published systematic reviews are not including meta-analyses.[32] Despite our focus on the impact of unpublished and grey literature study data on pooled effect estimates in meta-analyses, we believe that our findings are also applicable to systematic reviews with qualitative/descriptive summaries.

Comparison with other systematic reviews

We are aware of one methodological Cochrane review which addressed the impact of unpublished and grey literature data in meta-analyses on the basis of meta-epidemiological studies.[5] This review was published in 2007 and concluded that grey literature trials show an overall greater treatment effect than published trials. The authors acknowledged that the evidence is sparse and that more efforts are needed to identify a complete and unbiased set of trials irrespective of whether they have been published or not. In contrast to our review, this methodological review is nearly 10 years old, did not apply methods to address risk of bias and generalizability of the results of the included studies covering the given research question.

Our findings suggest that dissemination bias is a very serious threat to the results of meta-analyses, but not always impacts their results. This finding is supported by other studies (not meeting the inclusion criteria for this review) based on unpublished FDA data and published data.e.g., [6, 33] One of these meta-analyses investigating selective publication of antidepressant trials found a bias toward the publication of positive results, resulting in an effect size nearly one third larger than the effect size derived from unpublished FDA data.[6] Controversially, MacLean and collegues[33] reported that risk ratios for dyspepsia did not significantly or clinically differ using published or unpublished FDA data.

Implications for policy makers and further research

This work has implications for researchers and those who use meta-analyses to help inform clinical and policy decisions. (i) Investigators should ensure a comprehensive systematic literature search to avoid or at least attenuate the effect of dissemination bias. Such searches can be resource-intensive particularly when unpublished and grey literature data need to be identified. If the available resources do not permit comprehensive searches to identify unpublished or grey literature data, we strongly recommend (at least) a search in trial registries (such as the ICTRP and ClinicalTrials.gov) and websites of regulatory authorities which is less resource-intensive than searching for conference proceedings or dissertations, contacting experts, the industry and authors. When including unpublished or grey literature data sensitivity analyses should be carried out taking into account that this research may provide only preliminary results, is usually not peer reviewed and/or at higher risk of bias. It is obvious that even a thorough literature search cannot eliminate dissemination bias. Therefore, it is also of great importance to apply additional methods for detecting, quantifying and adjusting for dissemination bias in meta-analyses.[14] Such methods include graphical methods based on funnel plot asymmetry, statistical methods, such as regression tests, selection models, and a great number of more recent statistical approaches. [2] [3436] However, the empirical research work of Mueller et al 2016 concluded that it remains difficult to advise which method should be used as they are all limited and only few methods have been validated in empirical evaluations using unpublished studies obtained from regulators (e.g., FDA studies).[14] Selective outcome reporting in clinical studies is also an indicator for hidden or missing data, especially when only selective slices of the complete clinical trial are published or when studies show huge drop-out rates without providing reasons for these patients who left the study.[7, 37, 38] Overall, researchers should carefully consider the potential risk of dissemination bias when interpreting their findings. (ii) Those using meta-analyses to assist with clinical and policy decisions should also be aware of dissemination bias, because dissemination bias may have direct impact for patient care.[39] (iii) Major improvements have been made in the accessibility of data by initiatives such as the AllTrials campaign (www.alltrials.net) calling for all trials to be registered and the methods and results to be reported, the European Medicines Agency (EMA) policy on publication of clinical data on request since 2015, the obligatory release of results in trial registries by the European law (Clinical Trial Regulation), the FDA Amendment Act in 2007 and advocacy from the Cochrane Collaboration to fully implement such policies. Although progress has been made, there are still major issues related to unrestricted data access. Even when data are released, they can be incomplete, selective or not in compliance with the results reported in study registers such as ClinicalTrials.gov.[40] [41] [42] Therefore, further action is required to progress toward unrestricted data access. Particularly the full release of clinical study reports (CSR) may contain more information than other unpublished sources and, therefore, may have the potential to overcome existing problems.[43] (iv) Our research indicates that it seems that it will not be possible for a meta-analyst to judge before-hand whether the addition of unpublished and grey literature study data impacts the pooled effect estimates and leads to a change in the overall conclusions. (v) Finally, even the most comprehensive search for grey and unpublished data will not allow a final judgment whether the identified sample is in fact complete and representative for all of the hidden data.

Supporting information

Acknowledgments

The authors thank the members of the OPEN consortium Gerd Antes, Vittorio Bertele, Xavier Bonfill, Marie-Charlotte Bouesseau, Isabelle Boutron, Silvano Gallus, Silvio Garattini, Karam Ghassan, Carlo La Vecchia, Britta Lang, Jasper Littmann, Jos Kleijnen, Michael Kulig, Mario Malicki, Ana Marusic, Katharina Felicitas Mueller, Hector Pardo, Matthias Perleth, Philippe Ravaud, Andreas Reis, Daniel Strech, Ludovic Trinquart, Gerard Urrútia, Elizabeth Wager, Alexandra Wieland, and Robert Wolff.

The authors also thank Edith Motschall for conducting the comprehensive systematic literature search.

Author Contributions

  1. Conceptualization: CS JM.
  2. Formal analysis: GS.
  3. Funding acquisition: JM.
  4. Investigation: CS LKS AB PO LC JM.
  5. Methodology: CS JM.
  6. Supervision: JM.
  7. Writing – original draft: CS.
  8. Writing – review & editing: CS EvE MB JM.

References

  1. 1. Higgins JPT, Green S. Cochrane handbook for systematic reviews of interventions version 5.1.0 [updated march 2011]. The cochrane collaboration, 2011. www.cochrane-handbook.Org.
  2. 2. Sterne JA, Sutton AJ, Ioannidis JP, Terrin N, Jones DR, Lau J, et al. Recommendations for examining and interpreting funnel plot asymmetry in meta-analyses of randomised controlled trials. BMJ. 2011; 343: d4002. pmid:21784880
  3. 3. McAuley L, Pham B, Tugwell P, Moher D. Does the inclusion of grey literature influence estimates of intervention effectiveness reported in meta-analyses? Lancet. 2000; 356: 1228–1231. pmid:11072941
  4. 4. Song F, Eastwood AJ, Gilbody S, Duley L, Sutton AJ. Publication and related biases. Health Technol Assess. 2000; 4: 1–115.
  5. 5. Hopewell S, McDonald S, Clarke MJ, Egger M. Grey literature in meta-analyses of randomized trials of health care interventions. Cochrane Database Syst Rev. 2007; 2: MR000010.
  6. 6. Turner EH, Matthews AM, Linardatos E, Tell RA, Rosenthal R. Selective publication of antidepressant trials and its influence on apparent efficacy. N Engl J Med. 2008; 358: 252–260. pmid:18199864
  7. 7. Eyding D, Lelgemann M, Grouven U, Harter M, Kromp M, Kaiser T, et al. Reboxetine for acute treatment of major depression: Systematic review and meta-analysis of published and unpublished placebo and selective serotonin reuptake inhibitor controlled trials. BMJ. 2010; 341: c4737. pmid:20940209
  8. 8. Driessen E, Hollon SD, Bockting CL, Cuijpers P, Turner EH. Does publication bias inflate the apparent efficacy of psychological treatment for major depressive disorder? A systematic review and meta-analysis of us national institutes of health-funded trials. PLoS One. 2015; 10: e0137864. pmid:26422604
  9. 9. Cook DJ, Guyatt GH, Ryan G, Clifton J, Buckingham L, Willan A, et al. Should unpublished data be included in meta-analyses? Current convictions and controversies. Jama. 1993; 269: 2749–2753. pmid:8492400
  10. 10. Schmucker C, Schell LK, Portalupi S, Oeller P, Cabrera L, Bassler D, et al. Extent of non-publication in cohorts of studies approved by research ethics committees or included in trial registries. PLoS One. 2014; 9: e114023. pmid:25536072
  11. 11. Mueller KF, Briel M, Strech D, Meerpohl JJ, Lang B, Motschall E, et al. Dissemination bias in systematic reviews of animal research: A systematic review. PLoS One. 2014; 9: e116016. pmid:25541734
  12. 12. Mueller KF, Meerpohl JJ, Briel M, Antes G, von Elm E, Lang B, et al. Detecting, quantifying and adjusting for publication bias in meta-analyses: Protocol of a systematic review on methods. Syst Rev. 2013; 2: 60. pmid:23885765
  13. 13. Portalupi S, von Elm E, Schmucker C, Lang B, Motschall E, Schwarzer G, et al. Protocol for a systematic review on the extent of non-publication of research studies and associated study characteristics. Syst Rev. 2013; 2: 2. pmid:23302739
  14. 14. Mueller KF, Meerpohl JJ, Briel M, Antes G, von Elm E, Lang B, et al. Methods for detecting, quantifying, and adjusting for dissemination bias in meta-analysis are described. J Clin Epidemiol. 2016; 80: 25–33. pmid:27502970
  15. 15. Schmucker C, Bluemle A, Briel M, Portalupi S, Lang B, Motschall E, et al. A protocol for a systematic review on the impact of unpublished studies and studies published in the gray literature in meta-analyses. Syst Rev. 2013; 2: 24. pmid:23634657
  16. 16. Scherer RW, Langenberg P, von Elm E. Full publication of results initially presented in abstracts. Cochrane Database Syst Rev. 2007; MR000005. pmid:17443628
  17. 17. McAuley LM, Moher D, Tugwell P. The role of grey literature in meta-analysis [abstract]. Third International Congress on Biomedical Peer Review and Global Communications; 1997 Sept 18–20; Prague, Czech Republic. 1997.
  18. 18. Burdett S, Stewart LA. Publication bias and meta-analysis: A practical example [abstract]. 8th Annual Cochrane Colloquium; 2000 Oct 25–29; Cape Town, South Africa. 2000; 12.
  19. 19. Burdett S, Stewart LA, Tierney JF. Publication bias and meta-analyses: A practical example. Int J Technol Assess Health Care. 2003; 19: 129–134. pmid:12701945
  20. 20. Martin JLR, Perez V, Sacristan M, Alvarez E. Is grey literature essential for a better control of publication bias in psychiatry? An example from three meta-analyses fo schizophrenia. European Psychiatry. 2005; 20: 550–553. pmid:15994063
  21. 21. Fergusson D, Laupacis A, Salmi LR, McAlister FA, Huet C. What should be included in meta-analyses? An exploration of methodological issues using the ispot meta-analyses. Int J Technol Assess Health Care. 2000; 16: 1109–1119. pmid:11155831
  22. 22. Egger M, Juni P, Bartlett C, Holenstein F, Sterne J. How important are comprehensive literature searches and the assessment of trial quality in systematic reviews? Empirical study. Health Technology Assessment. 2003; 7: 1–76.
  23. 23. Hart B, Lundh A, Bero L. Effect of reporting bias on meta-analyses of drug trials: Reanalysis of meta-analyses. BMJ. 2012; 344: d7202. pmid:22214754
  24. 24. Golder S, Loke YK, Wright K, Norman G. Reporting of adverse events in published and unpublished studies of health care interventions: A systematic review. PLoS Med. 2016; 13: e1002127. pmid:27649528
  25. 25. Golder S, Loke YK, Bland M. Unpublished data can be of value in systematic reviews of adverse effects: Methodological overview. J Clin Epidemiol. 2010; 63: 1071–1081. pmid:20457510
  26. 26. Lundh A, Sismondo S, Lexchin J, Busuioc OA, Bero L. Industry sponsorship and research outcome. Cochrane Database Syst Rev. 2012; 12: Mr000033. pmid:23235689
  27. 27. Hopewell S. Impact of grey literature on systematic reviews of randomized trials [phd thesis]. Oxford: Wolfson College, University of Oxford. 2004.
  28. 28. Sterne JA, Juni P, Schulz KF, Altman DG, Bartlett C, Egger M. Statistical methods for assessing the influence of study characteristics on treatment effects in 'meta-epidemiological' research. Stat Med. 2002; 21: 1513–1524. pmid:12111917
  29. 29. Rennie D. Fair conduct and fair reporting of clinical trials. Jama. 1999; 282: 1766–1768. pmid:10568651
  30. 30. Scherer RW, Ugarte-Gil C, Schmucker C, Meerpohl JJ. Authors report lack of time as main reason for unpublished research presented at biomedical conferences: A systematic review. J Clin Epidemiol. 2015; 68: 803–810. pmid:25797837
  31. 31. Guyatt GH, Oxman AD, Vist GE, Kunz R, Falck-Ytter Y, Alonso-Coello P, et al. Grade: An emerging consensus on rating quality of evidence and strength of recommendations. BMJ 2008; 336: 924–926. pmid:18436948
  32. 32. Clarke M, Hopewell S, Chalmers I. Clinical trials should begin and end with systematic reviews of relevant evidence: 12 years and waiting. Lancet. 2010; 376: 20–21. pmid:20609983
  33. 33. MacLean CH, Morton SC, Ofman JJ, Roth EA, Shekelle PG. How useful are unpublished data from the food and drug administration in meta-analysis? J Clin Epidemiol. 2003; 56: 44–51. pmid:12589869
  34. 34. Kepes S, Banks G, McDaniel M, Whetzel D. Publication bias in the organizational sciences. Organizational Research Methods. 2012; 15: 624–662.
  35. 35. Langhorne P. Bias in meta-analysis detected by a simple, graphical test. Prospectively identified trials could be used for comparison with meta-analyses. BMJ. 1998; 316: 471.
  36. 36. David SP, Ware JJ, Chu IM, Loftus PD, Fusar-Poli P, Radua J, et al. Potential reporting bias in fmri studies of the brain. PLoS One. 2013; 8: e70104. pmid:23936149
  37. 37. Loder E, Tovey D, Godlee F. The tamiflu trials. BMJ. 2014; 348: g2630. pmid:24811414
  38. 38. Chan AW, Song F, Vickers A, Jefferson T, Dickersin K, Gotzsche PC, et al. Increasing value and reducing waste: Addressing inaccessible research. Lancet. 2014; 383: 257–266. pmid:24411650
  39. 39. Meerpohl JJ, Schell LK, Bassler D, Gallus S, Kleijnen J, Kulig M, et al. Evidence-informed recommendations to reduce dissemination bias in clinical research: Conclusions from the open (overcome failure to publish negative findings) project based on an international consensus meeting. BMJ Open. 2015; 5: e006666. pmid:25943371
  40. 40. Boutron I, Dechartres A, Baron G, Li J, Ravaud P. Sharing of data from industry-funded registered clinical trials. Jama. 2016; 315: 2729–2730. pmid:27367768
  41. 41. Miller JE, Korn D, Ross JS. Clinical trial registration, reporting, publication and fdaaa compliance: A cross-sectional analysis and ranking of new drugs approved by the fda in 2012. BMJ Open. 2015; 5: e009758. pmid:26563214
  42. 42. Cohen D. Dabigatran: How the drug company withheld important analyses. BMJ. 2014; 349: g4670. pmid:25055829
  43. 43. Maund E, Tendal B, Hrobjartsson A, Jorgensen KJ, Lundh A, Schroll J, et al. Benefits and harms in clinical trials of duloxetine for treatment of major depressive disorder: Comparison of clinical study reports, trial registries, and publications. BMJ. 2014; 348: g3510. pmid:24899650