Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Poor replication validity of biomedical association studies reported by newspapers

Abstract

Objective

To investigate the replication validity of biomedical association studies covered by newspapers.

Methods

We used a database of 4723 primary studies included in 306 meta-analysis articles. These studies associated a risk factor with a disease in three biomedical domains, psychiatry, neurology and four somatic diseases. They were classified into a lifestyle category (e.g. smoking) and a non-lifestyle category (e.g. genetic risk). Using the database Dow Jones Factiva, we investigated the newspaper coverage of each study. Their replication validity was assessed using a comparison with their corresponding meta-analyses.

Results

Among the 5029 articles of our database, 156 primary studies (of which 63 were lifestyle studies) and 5 meta-analysis articles were reported in 1561 newspaper articles. The percentage of covered studies and the number of newspaper articles per study strongly increased with the impact factor of the journal that published each scientific study. Newspapers almost equally covered initial (5/39 12.8%) and subsequent (58/600 9.7%) lifestyle studies. In contrast, initial non-lifestyle studies were covered more often (48/366 13.1%) than subsequent ones (45/3718 1.2%). Newspapers never covered initial studies reporting null findings and rarely reported subsequent null observations. Only 48.7% of the 156 studies reported by newspapers were confirmed by the corresponding meta-analyses. Initial non-lifestyle studies were less often confirmed (16/48) than subsequent ones (29/45) and than lifestyle studies (31/63). Psychiatric studies covered by newspapers were less often confirmed (10/38) than the neurological (26/41) or somatic (40/77) ones. This is correlated to an even larger coverage of initial studies in psychiatry. Whereas 234 newspaper articles covered the 35 initial studies that were later disconfirmed, only four press articles covered a subsequent null finding and mentioned the refutation of an initial claim.

Conclusion

Journalists preferentially cover initial findings although they are often contradicted by meta-analyses and rarely inform the public when they are disconfirmed.

Introduction

Over recent years many scientific editorials have highlighted the "reproducibility crisis" of biomedical findings [15]. Empirical studies have actually observed that many initial findings are refuted by subsequent studies, a trend that holds true at all three levels of biomedical research: i) preclinical studies [6, 7], ii) associations between biomarkers or risk factors and diseases [811], and iii) clinical trials [1216]. These empirical studies, as well as theoretical considerations [1], have highlighted that initial findings on a novel scientific question are especially vulnerable to refutation. However, others have argued that it is a mistake to expect 100% replication [17]: “a single study almost never provides definitive resolution for or against an effect and its explanation.” Scientific progress is a cumulative process that reduces uncertainty by combining uncertain initial findings with validation by subsequent studies. In other words, the low replication validity of many initial studies does not hinder the advancement of knowledge as long as science remains self-corrective [18].

However, an unrecognized consequence of the low replication validity of single studies concerns their media coverage. In the case of attention deficit hyperactivity disorder (ADHD), we previously found that newspapers preferentially report initial findings although most are refuted or strongly attenuated by subsequent studies [19]. We examined the replication validity of ten biomedical studies related to ADHD, which were widely reported by the media during the 1990’s. Among them, seven were found to be initial studies and all seven were either refuted or strongly attenuated by subsequent studies [19]. However, the strength of evidence provided by this preliminary study was limited by its sample size. The present study examined the preferential media coverage of early biomedical findings on a larger scale. Another weakness of our preliminary study was related to the heterogeneity of the ten ADHD studies that were examined: one was a pre-clinical study in mice, another estimated ADHD prevalence, three investigated the effectiveness of medication and five investigated the association of ADHD with biomarkers or risk factors. In the present study we did not consider pre-clinical studies because they are rarely included in meta-analyses, thus precluding assessment of their replication validity. We also excluded clinical trials investigating treatment effectiveness because their publication is influenced by stakeholders, in most cases pharmaceutical companies [20, 21], and because their media coverage is biased by this influence [22, 23]. Because this influence represents a parameter that might confound our analysis, we decided to focus on studies associating biomarkers and risk factors with diseases. This type of biomedical research is crucial for improving the understanding, diagnosis and prevention of diseases and is thus often reported by newspapers.

We recently built a large database of 595 meta-analyses investigating biomarkers and risk factors associated with four psychiatric disorders, four neurological pathologies and four somatic diseases [11]. Starting from the assumption that each meta-analysis provided the best estimate of the ‘true’ effect size, we showed that most effect sizes reported by initial studies were inflated. In the present study we investigated three questions. First, we quantified and characterized the newspaper coverage of the scientific publications included in our database. Second, we tested whether scientific observations reported by newspapers were confirmed, attenuated or refuted by the corresponding meta-analyses. We explored several factors that might influence the percentage of studies covered by newspapers and the replication validity of scientific observations that are covered: the type of study (initial versus subsequent studies and null findings versus "positive" ones), the prestige of the scientific journal that published them, the biomedical domain and the type of association. In particular, as previous studies showed that newspapers preferentially cover studies associating diseases with lifestyle risk factors (e.g. smoking) [24, 25], we considered this type of association to be a category. Third, we analyzed the possible media coverage of subsequent studies contradicting initial claims that had previously been reported by newspapers.

Methods

Scientific studies and their classification

The database described in our previous publication [11] included 663 meta-analyses investigating the association between biomarkers, risk factors or cognitive/behavioral observations and twelve pathologies belonging to three biomedical domains: psychiatry (Attention Deficit Hyperactivity Disorder (ADHD), autism, major depression, and schizophrenia), neurology (Alzheimer's and Parkinson's diseases, epilepsy, multiple sclerosis) and four somatic diseases (breast cancer, glaucoma, psoriasis and rheumatoid arthritis). In a preliminary step of the present study we observed that initial studies related to cognitive/behavioral observations were never covered by newspapers. Therefore, the 68 meta-analyses dealing with cognitive/behavioral observations were not considered in the present study. The 595 meta-analyses considered here were reported by 306 meta-analysis articles and included primary observations reported in 4775 scientific articles. As described previously, initial studies were distinguished from subsequent ones on the basis of their publication date (their references are given in our previous study [11]). From this database we excluded primary articles published before 1981. Indeed, we used the Dow Jones Factiva database to test whether these primary articles were covered by newspapers and the Factiva database starts on January 2nd 1981. Thus our final database includes 5029 scientific articles: 405 initial studies, 4218 subsequent studies and 306 meta-analysis articles.

In our previous study, these 595 meta-analyses were classified into three types: genetic studies, brain imaging studies and "other" studies. This latter type included a wide variety of biochemical, biophysical and epidemiological studies. Previous studies [24, 25] and a preliminary step of the present study led us to subdivide all association studies into two categories: a lifestyle category corresponding to any association between a pathology and a risk factor on which each subject can individually decide to act (e.g. red meat consumption, smoking, non compulsory vaccination) and a non-lifestyle category.

Impact factor of the journals that published scientific studies

We used the Journal Citation Reports from The Web of Science to determine the 2012 impact factor of the journals that published the 5029 articles of our database. In our previous study we observed that studies associating ADHD with a biomarker or a risk factor were rarely covered by newspapers if they were published in a scientific journal with an impact factor below five [19]. Therefore, when testing whether each study of our database was echoed by newspapers, we limited our initial search to articles published in scientific journals with an impact factor equal or superior to five. We observed that among the 91 scientific articles reported by newspapers and related to other pathologies than breast cancer, the lowest impact factor was 6.5. We found no press coverage for the 220 articles published in journals with an impact factor between 5 and 6.5. Therefore, it seems unlikely that the 2342 primary studies not related to breast cancer and published in journals with an impact factor below five attracted journalists' attention. As a result, we did not check the media coverage of these 2342 primary studies. Regarding breast cancer, among the 64 scientific articles reported by newspapers, we found 14 articles with an impact factor between 5 and 6.5. Therefore, we extended our search for a possible media coverage of breast cancer studies with an impact factor between 4 and 5. However, we found no media coverage for these 340 articles. Therefore, we did not check the media coverage of the 339 primary studies related to breast cancer with an impact factor below four.

In order to test the influence of the impact factor on media coverage, we considered three classes. Prestigious journals with an impact factor ≥ 30 (Nature, Science, JAMA, The Lancet and The New England Journal of Medicine) formed the first class. The limit between the two other classes was adjusted in order to keep the number of covered articles roughly equal in both classes. Thus, the second class included articles published in journals with an impact factor between 10 and 30 and the third one all other articles (IF ≤ 10).

Newspaper coverage of scientific studies

As previously described [19], we used the Dow Jones Factiva database to find newspaper articles echoing the scientific studies of our database. Briefly, the following keywords were used in each search: "study OR research* OR scientist*". Moreover, each search also used specific keywords that included the name of the pathology, one or several keywords that characterized the study (e.g. "gene", "milk", "birth weight") and was restricted to one month after the publication date of the study. Some searches, particularly about genetic associations, led to many items. In those cases, we added the name of the authors of the study, or of the university to the keywords. We only considered newspaper articles written in English and published in the general press. Articles published in the specialized press (e.g. Pharma Business Week), as well as articles published by any press agency (e.g. Associated Press), were not taken into account. Sometimes newspaper articles found by this search actually highlighted a biomedical finding that was not included in our database of 595 meta-analyses. For example, some newspaper articles underlined risk factors associated with colon cancer, but did not mention the association of the same risk factor with breast cancer although it was also reported in the corresponding scientific article. In these few cases this media coverage was not considered. This search was performed by one author (EDM). Another author (FG) independently checked the possible media coverage of 20% of scientific studies. Both authors always reached agreement on whether any scientific study was echoed or not. A few minor disagreements occurred regarding the number of newspaper articles reporting some studies and they were resolved by discussion.

Replication validity of scientific observations reported by newspapers

We used our previously published database of 595 meta-analyses to check whether biomedical findings echoed by newspapers were confirmed, or not, by the corresponding meta-analyses. As previously described [11], a primary observation was considered in agreement with the corresponding meta-analysis if it satisfied two criteria: first, both observations were in nominal agreement (i.e. both reported a statistically significant effect in the same direction or the absence of a significant effect) and, second, when both reported a significant effect, the effect size of the primary observation must not be more than twice as large as that of the corresponding meta-analysis (i.e. effect size inflation below 100%). Sometimes single scientific articles reported several associations included in our 595 meta-analyses while newspaper articles that echoed them did not specify which association they highlighted. This is often the case with genetic studies: newspapers reported that new genes have been found to be associated with a disease, whereas the scientific article described significant associations with several single nucleotide polymorphisms. In these cases we considered the replication validity of the most credible association (i.e. with the smallest p value of the statistical test). Thus, we assessed the replication validity of any scientific article covered by newspapers under the most favorable conditions.

Results

Overview of scientific articles reported by newspapers

Among the 4723 primary studies of our database 156 were echoed by at least one newspaper article (Table 1). These 156 studies were covered by 1475 newspaper articles and were related to ten pathologies (ADHD, autism, major depression, schizophrenia, Alzheimer's and Parkinson's diseases, multiple sclerosis, breast cancer, glaucoma and rheumatoid arthritis). The primary studies included in our database and related to epilepsy or psoriasis did not attract newspaper attention. Moreover, 86 newspaper articles covered five out of the 306 meta-analysis articles of our database. The references of these 156 studies and of the corresponding meta-analysis articles are given in Supporting Information (S1 Text). The raw data about the studies covered by newspapers and their corresponding meta-analyses are also given (S1 File).

thumbnail
Table 1. Number of primary articles included in meta-analyses and number of primary articles covered by newspapers classified by domains and association types.

https://doi.org/10.1371/journal.pone.0172650.t001

The number of newspaper articles echoing each of these 161 scientific articles (156 primary and 5 meta-analysis articles) ranged from 1 to 50 and this number was positively correlated with the impact factor of the journal that published them (linear regression, R2 = 0.107, F-test on variance equality: p = 0.003, Pearson's test: p < 0.0001).

Biomedical domain and association type of covered studies

The percentage of primary articles covered by newspapers ranged from 1.74% to 10.5% depending on the biomedical domain and the association type (Table 1). Whatever the domain, lifestyle studies were more often covered than non-lifestyle ones (Table 1). Compared to this major effect of association type, the coverage rate does not strongly differ between biomedical domains (Table 1). Newspapers also covered a larger proportion of meta-analysis articles of the lifestyle category (Table 1). Altogether, newspapers echoed 161 scientific articles, of which 65 (40.4%) were of the lifestyle category. Moreover, lifestyle associations tended to receive a larger newspaper coverage (mean: 11.9 newspaper articles per study) than non-lifestyle associations (mean: 8.2 newspaper articles per study). However, this difference does not reach statistical significance. Overall, this larger press coverage of lifestyle studies was not related to the impact factor of the scientific journals that published them. Indeed, lifestyle scientific studies were published in less prestigious journals (median IF = 10.6) than those of the non-lifestyle category (median IF = 17.6) and this difference is statistically significant (unpaired t test, p = 0.0003).

Media coverage of initial studies versus subsequent ones

Initial studies attract more media attention than subsequent studies. Indeed, 13.1% of initial studies were covered by newspapers whereas only 2.4% of subsequent studies were (raw data are given in Supporting Information: S2 Text). This preferential coverage is largely influenced by the type of association and the impact factor of the scientific journal. Among the 639 primary articles of the lifestyle category, 39 reported initial findings and 600 subsequent observations. Newspapers almost equally covered these initial findings (5 covered articles, 12.8%) and these subsequent observations (58 covered articles, 9.7%). In contrast, among the 4084 primary articles investigating non-lifestyle associations, newspapers preferentially covered initial findings (Fig 1). Indeed, they echoed 48 of the 366 initial non-lifestyle studies (13.1%), but only 45 of the 3718 subsequent ones (1.2%). This difference reaches statistical significance (Chi2 test: X2 = 211.52 p<0.0001). This preferential coverage of initial non-lifestyle studies also holds true when considering articles published in scientific journals with an impact factor below 10 (Chi 2 test: X2 = 17.73 p<0.0001) and with an impact factor between 10 and 30 (Chi 2 test: X2 = 29.11 p<0.0001) (Fig 1). However, for the 83 non-lifestyle articles published in prestigious journals, no significant preferential coverage of initial observations was observed (Chi 2 test: X2 = 3.49 p = 0.062). Finally, the percentage of articles covered by newspapers strongly increased with the impact factor (Fig 1). This effect is statistically significant regarding initial or subsequent non-lifestyle studies (2x3 Chi 2 tests: X2 = 124.4 p<0.0001 or X2 = 612.1 p<0.0001, respectively) and for lifestyle studies (2x3 Fisher test: p<0.0001). Therefore, publishing a study in a prestigious journal considerably increases its chance of being covered by newspapers irrespective of the type of association investigated or its newness.

thumbnail
Fig 1. Preferential coverage of initial findings and influence of the impact factor (IF).

The figure shows the percentage of primary studies that are covered by newspapers depending on the study type (lifestyle versus non-lifestyle). Studies of the lifestyle category described associations linking a pathology to a risk factor on which each subject can act. Regarding non-lifestyle articles, the figure also contrasts initial articles with subsequent ones. Differences in the media coverage between initial studies and subsequent ones were statistically significant (see text) except for studies published in prestigious journals (IF ≥ 30). Raw data are given in Supporting Information (S2 Text).

https://doi.org/10.1371/journal.pone.0172650.g001

Media coverage of "positive" versus null findings

All 53 initial studies covered by newspapers reported a "positive" finding (i.e. a statistically significant effect). It follows that none of the 174 initial studies included in our database and reporting a null effect (i.e. 43% of all initial studies) attracted journalists' attention (Table 2). Among the 103 subsequent studies covered by newspapers, only 14 studies reported a null effect. Regarding the non-lifestyle associations, among the 45 subsequent studies covered by newspapers, only five reported a null finding. Moreover this media coverage was small (10 articles) compared to that of all non-lifestyle studies (742 newspaper articles). The 58 subsequent studies of the lifestyle category were echoed by 733 newspaper articles. Only nine subsequent studies reported the absence of a significant association and these null findings were reported by 65 newspaper articles. Therefore, assuming that, in our database, the percentage of subsequent studies reporting a null finding is at least as large as that of initial studies (43%) we observed that newspapers under-reported them. Indeed, only 13.6% of the subsequent studies covered by newspapers reported a null finding (Table 2). Moreover, the number of newspaper articles covering each null finding was small, particularly regarding non-lifestyle studies (Table 2). Finally, only one of the five meta-analysis articles covered by newspapers reported a null finding. Moreover, this null finding was echoed by only four newspaper articles whereas the four "positive" meta-analyses attracted 82 press articles.

thumbnail
Table 2. Newspaper coverage of primary studies reporting null findings.

https://doi.org/10.1371/journal.pone.0172650.t002

Replication validity of primary studies reported by newspapers

Overall, among the 156 primary studies echoed by newspapers 76 (48.7%) reported a main finding consistent with the corresponding meta-analysis (raw data are given in Supporting Information: S2 Text). Among the 156 primary studies covered by newspapers 53 were initial studies of which only 18 (34.0%) were confirmed by the corresponding meta-analyses. In contrast, of the 103 subsequent studies 58 (56.3%) were confirmed and this difference is statistically significant (Chi2 test: X2 = 6.99 p = 0.0082). Regarding the 63 lifestyle studies echoed by newspapers, 49.2% of their findings were confirmed by the corresponding meta-analyses without any effect of the impact factor (Fig 2). Only 33.3% of the main findings reported by the 48 initial non-lifestyle studies and covered by newspapers were actually validated by the corresponding meta-analyses. This replication rate was not influenced by the impact factor (Fig 2). In contrast the 45 subsequent non-lifestyle studies covered by newspapers were more often consistent with the corresponding meta-analyses than the initial non-lifestyle studies (Fig 2). Their replication rate was indeed larger than that of initial studies whether considering them as a whole (64.4% versus 33.3% for initial studies; Chi2 test, X2 = 9.0 p = 0.0027) or when considering only the 17 studies published in prestigious journals (88.2% versus 40.7% for initial studies; Fisher test, p = 0.0021). The replication rate of subsequent non-lifestyle studies increased with the impact factor (2x3 Fisher test, p = 0.0055).

thumbnail
Fig 2. Replication validity of primary articles reported by newspapers.

The figure shows the percentage of primary articles echoed by newspapers whose main finding was consistent with the corresponding meta-analysis. We considered here the same three categories as in Fig 1: primary articles of the lifestyle category and initial or subsequent non-lifestyle studies. Raw data are given in Supporting Information (S2 Text).

https://doi.org/10.1371/journal.pone.0172650.g002

When we classified the 156 primary studies echoed by newspapers according to the biomedical domain (Fig 3), studies related to psychiatric disorders were less often confirmed by the corresponding meta-analyses than those related to neurological diseases (chi2 test; X2 = 10.94 p = 0.0009) and to somatic diseases (X2 = 6.8 p = 0.0091). Moreover, studies related to psychiatry and covered by newspapers were more often initial findings than those in neurology (X2 = 9.13 p = 0.0025) and for somatic diseases (X2 = 14.11 p = 0.0002). The low replication validity of studies in psychiatry appears thus to be correlated with the fact that newspapers favor initial findings in this domain even more than in other domains. This correlation further supports our previous conclusion [19]: the replication validity of biomedical studies covered by newspapers is poor because journalists favor initial findings over subsequent studies.

thumbnail
Fig 3. Replication validity of primary studies reported by newspapers in three biomedical domains.

The blue bars show the percentage of primary studies covered by newspapers whose main finding was consistent with the corresponding meta-analysis. The red bars show the percentage of initial findings among primary studies echoed by newspapers and related to four psychiatric disorders (PSY), four neurological diseases (NEURO) and four somatic diseases (SOMA). Raw data are given in Supporting Information (S2 Text).

https://doi.org/10.1371/journal.pone.0172650.g003

Analysis of primary studies covered by three newspapers or more

When considering only the 102 primary studies covered by at least three newspaper articles, we observed the same patterns (raw data are given in Supporting Information: S2 Text). Journalists covered initial studies more often (32/405) than subsequent studies (70/4318). In particular, non-lifestyle studies were ten times more often echoed when they reported initial findings than subsequent observations. Of these 102 primary studies, 53 (52.0%) were confirmed by the corresponding meta-analyses. The replication validity of the 28 subsequent non-lifestyle observations was still higher than that of the 29 initial findings (71.4% versus 41.4%; Chi2 test, X2 = 5.22 p = 0.022). Likewise, studies related to psychiatry and covered by newspapers were more often initial findings than those in neurology (Fisher test; p = 0.0072) and for somatic diseases (Chi2 test X2 = 9.59 p = 0.002). They were also confirmed less often (30.4%) than those related to neurological diseases (64%) (chi2 test; X2 = 5.41 p = 0.020) and to somatic diseases (55.6) (X2 = 4.08 p = 0.043).

Lack of media coverage of studies contradicting initial claims

Among the 53 initial studies covered by newspapers, 35 (66%) were disconfirmed by the corresponding meta-analyses. These 35 initial studies were followed by 503 subsequent studies of which 398 reported either the absence of a statistically significant effect or a significant effect in the opposite direction. Only one of these 398 subsequent studies and only one of the 35 corresponding meta-analysis articles were covered by newspapers. In the first case, the initial study published in 1999 by The Lancet (IF = 39) reported that the dopamine transporter level was increased (+70%) in both striatum of ADHD patients [26]. This initial study was covered by 22 newspaper articles. In contrast, according to a subsequent study published in 2009 by JAMA (IF = 30), the dopamine transporter level was significantly decreased (-19%) in the left striatum of ADHD patients, but not in the right [27]. The three newspapers that covered this subsequent study did not mention that it contradicted the initial claim although two newspapers (The Globe and Mail and The Washington Post) had also covered the initial study.

In the second case, the initial study was published in 2003 in Science and claimed that a genetic factor was associated with depression when subjects were exposed to stressful life events [28]. This finding was widely covered by newspapers (50 articles). Two subsequent studies that confirmed the same association were also echoed (7 and 2 newspaper articles). Newspapers never covered the eleven subsequent studies that failed to replicate this genetic association. Finally, the meta-analysis published in 2009 by Risch and coworkers in JAMA (IF = 30) and contradicting "positive" findings [29], was covered by only four newspaper articles and all four mentioned that the initial finding "has not held up to scientific scrutiny". Overall, 234 newspapers articles covered the 35 initial studies that were later disconfirmed by subsequent meta-analyses. However, only four newspaper articles explicitly mentioned the refutation of an initial claim previously covered in the press.

Comments

We explored the replication validity of association studies echoed by the lay press. Altogether, 48.7% of the studies examined were confirmed by subsequent meta-analyses. However, subpopulation analyses show large variations of this replication rate. Newspapers preferentially echo initial studies over subsequent studies investigating the same question and initial findings covered by newspapers are less often validated by replications (34%) than subsequent findings (56.3%). This observation is consistent with several observational studies reporting that many initial or early biomedical findings are disconfirmed by subsequent studies [616]. Thus, the present study further supports our previous conclusion [19]: a major factor that explains the poor replication validity of biomedical studies covered by newspapers is that journalists favor initial findings over replication studies.

In agreement with previous studies [24, 25], we observed that newspapers preferentially reported lifestyle association studies linking a pathology to a risk factor on which each reader can act. Non-lifestyle studies related to brain imaging, genetic factor or other inescapable risk factors were less often echoed. This preferential coverage further supports the view that the first journalists' aim is to attract readers' attention. We also observed that newspapers never covered initial findings reporting null findings and under-reported subsequent null findings, particularly regarding non-lifestyle studies. As a consequence, journalists rarely cover null findings that contradict previously echoed studies. Within our database we observed only one interesting exception to this general rule: four newspaper articles covered the meta-analysis by Risch and coworkers and informed the public that it disconfirmed Caspi's initial study, which received much larger coverage (50 articles). All these observations support the conclusion of a case study: "scientific uncertainty per se is not attractive to journalists" (p. 344) [30].

We observed that findings reported by newspapers and related to psychiatric disorders were less often confirmed by meta-analyses than those related to neurological or somatic diseases. We correlated this observation with the fact that the preferential coverage of initial studies is even stronger regarding psychiatry. However, another factor might contribute to the low replication validity of psychiatric studies covered by newspapers: we previously showed that initial association studies linking biomarkers or risk factors to psychiatric disorders were less often confirmed by corresponding meta-analyses than those related to neurological diseases [11]. Likewise, highly cited early studies investigating the effectiveness of psychiatric treatments are often contradicted by subsequent studies [16].

Limitations

Although single primary studies have often investigated several associations linking a biomarkers or risk factors to a pathology (e.g. several polymorphisms on the same gene), we restricted our replication analysis to only one association per scientific article. Because we selected the most credible association (i.e. with the highest degree of statistical significance) it is likely that the replication rates we reported here are overestimated.

The 4723 primary articles of our database were included in meta-analyses thus enabling us to assess their replication validity. Whether the newspaper coverage of association studies included in meta-analyses is representative of the coverage of all association studies remains an open question.

We only considered newspaper articles published within a month following the publication of each scientific study. Indeed, when newspaper articles were published much later, it was often impossible to identify the scientific study they were reporting on with certainty. Consequently, we did not investigate the long-term impact of scientific publications in the lay press, although there is no doubt that certain scientific studies did have such an impact. For example, Caspi's study published in 2003 about a genetic susceptibility to major depression [28] was still cited in the lay press well after its refutation in 2009 by a meta-analysis [29]. However, in relative terms, it is likely that the scientific publications with the strongest long-term impact also received the widest newspaper coverage shortly following their publication.

Our data support three conclusions. First, newspapers preferentially cover lifestyle associations over non-lifestyle ones. Second, they strongly favor initial findings over subsequent studies when they cover non-lifestyle associations, but they equally cover initial and subsequent lifestyle studies. Third, among non-lifestyle associations covered by newspapers, initial findings are less often confirmed by meta-analyses than subsequent studies. However, our number of observations is too small to know whether this third conclusion also holds true regarding lifestyle studies. Indeed, our database included only five initial lifestyle studies covered by newspapers.

We only studied newspaper articles but ignored television reporting, and this although television is the main source of information on medical sciences for Europeans (European Commission, 2007). Indeed, television programs rarely explicitly report on specific scientific publications [31]. Nevertheless, newspaper articles might indirectly influence the content of television programs by raising the attention of their producers.

Conclusion

Biomedical researches mature from early studies that are highly uncertain to scientific consensuses built on subsequent studies and quantified by meta-analyses. Indeed, recent meta-researches have consistently shown that most early studies report inflated effect size or false positive effects [616]. Our study shows that many biomedical findings reported by newspapers are disconfirmed by subsequent studies. This is partly due to the fact that newspapers preferentially cover "positive" initial studies rather than subsequent observations, in particular those reporting null findings. Our study also suggests that most journalists from the general press do not know or prefer not to deal with the high degree of uncertainty inherent in early biomedical studies. Importantly, such biased newspaper coverage can have important social consequences [32, 33]. For example, a content analysis of newspaper articles covering Caspi's study showed that they emphasized the genetic side of the gene-by-environment interactions and this discourse deflected public attention away from the considerable impact of social inequalities upon health [34]. Therefore, with Susan Watts [35], we advocate that society "needs science journalism to weigh up the values and the vices of new science" (p. 151). In particular, when preparing a report on a scientific study, journalists should always ask scientists whether it is an initial finding and, if so, they should inform the public that this discovery is still tentative and must be validated by subsequent studies. Larsson and coworkers (2003) have identified the obstacles science journalists meet to accurately cope with uncertainty [36]. In particular, most interviewed journalists feel that it is difficult to find scientists who are independent from authors and who are willing to assist them. Thus, scientists, either as independent experts or as authors, are also responsible for improving the informative value of biomedical reporting in the mass media. In particular, they are responsible for the accuracy of the press releases covering their work and published by scientific editors or universities. This is not only a moral duty but also in the interest of science. Indeed, contrary to common beliefs, reception studies show that scientists are viewed by the public as more trustworthy when news coverage of health research acknowledges its uncertainty, especially if this is attributed to scientific authors [37, 38].

Supporting information

S1 Text. References of the 156 studies reported by newspapers and of their corresponding meta-analysis articles.

https://doi.org/10.1371/journal.pone.0172650.s001

(DOCX)

S2 Text. Raw data shown in Figs 13 and about primary studies covered by three newspapers or more.

https://doi.org/10.1371/journal.pone.0172650.s002

(DOCX)

S1 File. Raw data about the 156 studies covered by newspapers and their corresponding meta-analyses.

https://doi.org/10.1371/journal.pone.0172650.s003

(XLSX)

Author Contributions

  1. Conceptualization: EDM AS TB FG.
  2. Data curation: EDM TB FG.
  3. Formal analysis: EDM TB FG.
  4. Investigation: EDM FG.
  5. Methodology: EDM FG.
  6. Project administration: EDM AS TB FG.
  7. Supervision: AS TB FG.
  8. Validation: FG.
  9. Visualization: EDM TB FG.
  10. Writing – original draft: EDM FG.
  11. Writing – review & editing: EDM AS TB FG.

References

  1. 1. Ioannidis JP. Why most published research findings are false. PLoS Med. 2005;2(8):e124. pmid:16060722
  2. 2. Sarewitz D. Beware the creeping cracks of bias. Nature. 2012;485(7397):149. pmid:22575922
  3. 3. Munafo MR, Stothart G, Flint J. Bias in genetic association studies and impact factor. Mol Psychiatry. 2009;14(2):119–20. pmid:19156153
  4. 4. Begley CG, Ioannidis JP. Reproducibility in science: improving the standard for basic and preclinical research. Circ Res. 2015;116(1):116–26. pmid:25552691
  5. 5. Baker M. 1,500 scientists lift the lid on reproducibility. Nature. 2016;533(7604):452–4. pmid:27225100
  6. 6. Prinz F, Schlange T, Asadullah K. Believe it or not: how much can we rely on published data on potential drug targets? Nat Rev Drug Discov. 2011;10(9):712. pmid:21892149
  7. 7. Freedman LP, Cockburn IM, Simcoe TS. The Economics of Reproducibility in Preclinical Research. PLoS Biol. 2015;13(6):e1002165. pmid:26057340
  8. 8. Ioannidis JP, Ntzani EE, Trikalinos TA, Contopoulos-Ioannidis DG. Replication validity of genetic association studies. Nat Genet. 2001;29(3):306–9. pmid:11600885
  9. 9. Trikalinos TA, Ntzani EE, Contopoulos-Ioannidis DG, Ioannidis JP. Establishment of genetic associations for complex diseases is independent of early study findings. Eur J Hum Genet. 2004;12(9):762–9. pmid:15213707
  10. 10. Ioannidis JP, Panagiotou OA. Comparison of effect sizes associated with biomarkers reported in highly cited individual articles and in subsequent meta-analyses. JAMA. 2011;305(21):2200–10. pmid:21632484
  11. 11. Dumas-Mallet E, Button K, Boraud T, Munafo M, Gonon F. Replication Validity of Initial Association Studies: A Comparison between Psychiatry, Neurology and Four Somatic Diseases. PLoS One. 2016;11(6):e0158064. pmid:27336301
  12. 12. Ioannidis J, Lau J. Evolution of treatment effects over time: empirical insight from recursive cumulative metaanalyses. Proc Natl Acad Sci U S A. 2001;98(3):831–6. pmid:11158556
  13. 13. Trikalinos TA, Churchill R, Ferri M, Leucht S, Tuunainen A, Wahlbeck K, et al. Effect sizes in cumulative meta-analyses of mental health randomized trials evolved over time. J Clin Epidemiol. 2004;57(11):1124–30. pmid:15612138
  14. 14. Gan HK, You B, Pond GR, Chen EX. Assumptions of expected benefits in randomized phase III trials evaluating systemic treatments for cancer. J Natl Cancer Inst. 2012;104(8):590–8. pmid:22491345
  15. 15. Pereira TV, Horwitz RI, Ioannidis JP. Empirical evaluation of very large treatment effects of medical interventions. JAMA. 2012;308(16):1676–84. pmid:23093165
  16. 16. Tajika A, Ogawa Y, Takeshima N, Hayasaka Y, Furukawa TA. Replication and contradiction of highly cited research papers in psychiatry: 10-year follow-up. Br J Psychiatry. 2015;207(4):357–62. pmid:26159600
  17. 17. Open-Science-Collaboration. PSYCHOLOGY. Estimating the reproducibility of psychological science. Science. 2015;349(6251):aac4716. pmid:26315443
  18. 18. Bacchetti P, Deeks SG, McCune JM. Breaking free of sample size dogma to perform innovative translational research. Sci Transl Med. 2011;3(87):87ps24. pmid:21677197
  19. 19. Gonon F, Konsman JP, Cohen D, Boraud T. Why most biomedical findings echoed by newspapers turn out to be false: the case of Attention Deficit Hyperactivity Disorder. PLoS One. 2012;7(9):e44275. pmid:22984483
  20. 20. Heres S, Davis J, Maino K, Jetzinger E, Kissling W, Leucht S. Why olanzapine beats risperidone, risperidone beats quetiapine, and quetiapine beats olanzapine: an exploratory analysis of head-to-head comparison studies of second-generation antipsychotics. Am J Psychiatry. 2006;163(2):185–94. pmid:16449469
  21. 21. Turner EH, Matthews AM, Linardatos E, Tell RA, Rosenthal R. Selective publication of antidepressant trials and its influence on apparent efficacy. N Engl J Med. 2008;358(3):252–60. pmid:18199864
  22. 22. Moynihan R, Bero L, Ross-Degnan D, Henry D, Lee K, Watkins J, et al. Coverage by the news media of the benefits and risks of medications. N Engl J Med. 2000;342(22):1645–50. pmid:10833211
  23. 23. Yavchitz A, Boutron I, Bafeta A, Marroun I, Charles P, Mantz J, et al. Misrepresentation of randomized controlled trials in press releases and news coverage: a cohort study. PLoS Med. 2012;9(9):e1001308. pmid:22984354
  24. 24. Bartlett C, Sterne J, Egger M. What is newsworthy? Longitudinal study of the reporting of medical research in two British newspapers. BMJ. 2002;325(7355):81–4. pmid:12114239
  25. 25. Stryker JE. Reporting medical information: effects of press releases and newsworthiness on medical journal articles' visibility in the news media. Prev Med. 2002;35(5):519–30. pmid:12431901
  26. 26. Dougherty DD, Bonab AA, Spencer TJ, Rauch SL, Madras BK, Fischman AJ. Dopamine transporter density in patients with attention deficit hyperactivity disorder. Lancet. 1999;354(9196):2132–3. pmid:10609822
  27. 27. Volkow ND, Wang GJ, Kollins SH, Wigal TL, Newcorn JH, Telang F, et al. Evaluating dopamine reward pathway in ADHD: clinical implications. JAMA. 2009;302(10):1084–91. pmid:19738093
  28. 28. Caspi A, Sugden K, Moffitt TE, Taylor A, Craig IW, Harrington H, et al. Influence of life stress on depression: moderation by a polymorphism in the 5-HTT gene. Science. 2003;301(5631):386–9. pmid:12869766
  29. 29. Risch N, Herrell R, Lehner T, Liang KY, Eaves L, Hoh J, et al. Interaction between the serotonin transporter gene (5-HTTLPR), stressful life events, and risk of depression: a meta-analysis. JAMA. 2009;301(23):2462–71. pmid:19531786
  30. 30. Kitzinger J, Reilly J. The rise and fall of risk reporting: media coverage of human genetics research, 'false memory syndrome' and 'mad cow disease'. Eur J Communication. 1997;12(3):319–50.
  31. 31. Bourdaa M, Konsman JP, Secail C, Venturini T, Veyrat-Masson I, Gonon F. Does television reflect the evolution of scientific knowledge? The case of attention deficit hyperactivity disorder coverage on French TV. Public Understanding of Science. 2015;24(2):200–9. pmid:23825292
  32. 32. Holtzman NA, Bernhardt BA, Mountcastle-Shah E, Rodgers JE, Tambor E, Geller G. The quality of media reports on discoveries related to human genetic diseases. Community Genet. 2005;8(3):133–44. pmid:16113530
  33. 33. Gonon F, Bezard E, Boraud T. What should be said to the lay public regarding ADHD etiology. Am J Med Genet B Neuropsychiatr Genet. 2011;156(8):989–91.
  34. 34. Horwitz AV. Media portrayals and health inequalities: a case study of characterizations of Gene x Environment interactions. J Gerontol B Psychol Sci Soc Sci. 2005;60 Spec No 2:48–52.
  35. 35. Watts S. Society needs more than wonder to respect science. Nature. 2014;508(7495):151. pmid:24717473
  36. 36. Larsson A, Oxman AD, Carling C, Herrin J. Medical messages in the media—barriers and solutions to improving medical journalism. Health Expect. 2003;6(4):323–31. pmid:15040794
  37. 37. Jensen JD, Hurley RJ. Conflicting stories about public scientific controversies: Effects of news convergence and divergence on scientists' credibility. Public Underst Sci. 2012;21(6):689–704. pmid:23832155
  38. 38. Jensen JD, Krakow M, John KK, Liu M. Against conventional wisdom: when the public, the media, and medical practice collide. BMC Med Inform Decis Mak. 2013;13 Suppl 3:S4.