HOME

TheInfoList



OR:

"Why Most Published Research Findings Are False" is a 2005 essay written by
John Ioannidis John P. A. Ioannidis (; el, Ιωάννης Ιωαννίδης, ; born August 21, 1965) is a Greek-American physician-scientist, writer and Stanford University professor who has made contributions to evidence-based medicine, epidemiology, and cl ...
, a professor at the Stanford School of Medicine, and published in ''
PLOS Medicine ''PLOS Medicine'' (formerly styled ''PLoS Medicine'') is a peer-reviewed weekly medical journal covering the full spectrum of the medical sciences. It began operation on October 19, 2004, as the second journal of the Public Library of Science (PLO ...
''. It is considered foundational to the field of
metascience Metascience (also known as meta-research) is the use of scientific methodology to study science itself. Metascience seeks to increase the quality of scientific research while reducing inefficiency. It is also known as "''research on research''" ...
. In the paper, Ioannidis argued that a large number, if not the majority, of published
medical research Medical research (or biomedical research), also known as experimental medicine, encompasses a wide array of research, extending from "basic research" (also called ''bench science'' or ''bench research''), – involving fundamental scientif ...
papers contain results that cannot be replicated. In simple terms, the essay states that scientists use
hypothesis testing A statistical hypothesis test is a method of statistical inference used to decide whether the data at hand sufficiently support a particular hypothesis. Hypothesis testing allows us to make probabilistic statements about population parameters. ...
to determine whether scientific discoveries are significant. "Significance" is formalized in terms of probability, and one formalized calculation ("''
P value In null-hypothesis significance testing, the ''p''-value is the probability of obtaining test results at least as extreme as the result actually observed, under the assumption that the null hypothesis is correct. A very small ''p''-value means ...
''") is reported in the scientific literature as a screening mechanism. Ioannidis posited assumptions about the way people perform and report these tests; then he constructed a statistical model which indicates that most published findings are false positive results.


Argument

Suppose that in a given scientific field there is a known baseline probability that a result is true, denoted by \mathbb(\text). When a study is conducted, the probability that a positive result is obtained is \mathbb(+). Given these two factors, we want to compute the
conditional probability In probability theory, conditional probability is a measure of the probability of an event occurring, given that another event (by assumption, presumption, assertion or evidence) has already occurred. This particular method relies on event B occur ...
\mathbb(\text\mid +), which is known as the positive predictive value (PPV). Bayes' theorem allows us to compute the PPV as:\mathbb(\text \mid +) = where \alpha is the
type I error rate In statistical hypothesis testing, a type I error is the mistaken rejection of an actually true null hypothesis (also known as a "false positive" finding or conclusion; example: "an innocent person is convicted"), while a type II error is the fa ...
and \beta is the type II error rate; the statistical power is 1-\beta. It is customary in most scientific research to desire \alpha = 0.05 and \beta = 0.2. If we assume \mathbb(\text) = 0.1 for a given scientific field, then we may compute the PPV for different values of \alpha and \beta: However, the simple formula for PPV derived from Bayes' theorem does not account for
bias Bias is a disproportionate weight ''in favor of'' or ''against'' an idea or thing, usually in a way that is closed-minded, prejudicial, or unfair. Biases can be innate or learned. People may develop biases for or against an individual, a group ...
in study design or reporting. Some published findings would not have been presented as research findings if not for researcher bias. Let u\in ,1/math> be the probability that an analysis was only published due to researcher bias. Then the PPV is given by the more general expression:\mathbb(\text, +) = The introduction of bias will tend to depress the PPV; in the extreme case when the bias of a study is maximized, \mathbb(\text, +) = \mathbb(\text). Even if a study meets the benchmark requirements for \alpha and \beta, and is free of bias, there is still a 36% probability that a paper reporting a positive result will be incorrect; if the base probability of a true result is lower, then this will push the PPV lower too. Furthermore, there is strong evidence that the average statistical power of a study in many scientific fields is well below the benchmark level of 0.8. Given the realities of bias, low statistical power, and a small number of true hypotheses, Ioannidis concludes that the majority of studies in a variety of scientific fields are likely to report results that are false.


Corollaries

In addition to the main result, Ioannidis lists six corollaries for factors that can influence the reliability of published research. Research findings in a scientific field are less likely to be true, # the smaller the studies conducted. # the smaller the
effect size In statistics, an effect size is a value measuring the strength of the relationship between two variables in a population, or a sample-based estimate of that quantity. It can refer to the value of a statistic calculated from a sample of data, the ...
s. # the greater the number and the lesser the selection of tested relationships. # the greater the flexibility in
designs A design is a plan or specification for the construction of an object or system or for the implementation of an activity or process or the result of that plan or specification in the form of a prototype, product, or process. The verb ''to design'' ...
, definitions, outcomes, and analytical modes. # the greater the financial and other interests and
prejudices Prejudice can be an affective feeling towards a person based on their perceived group membership. The word is often used to refer to a preconceived (usually unfavourable) evaluation or classification of another person based on that person's perc ...
. # the hotter the scientific field (with more scientific teams involved). Ioannidis has added to this work by contributing to a meta-epidemiological study which found that only 1 in 20 interventions tested in Cochrane Reviews have benefits that are supported by high-quality evidence. He also contributed to research suggesting that the quality of this evidence does not seem to improve over time.


Reception

Despite skepticism about extreme statements made in the paper, Ioannidis's broader argument and warnings have been accepted by a large number of researchers. The growth of
metascience Metascience (also known as meta-research) is the use of scientific methodology to study science itself. Metascience seeks to increase the quality of scientific research while reducing inefficiency. It is also known as "''research on research''" ...
and the recognition of a scientific
replication crisis The replication crisis (also called the replicability crisis and the reproducibility crisis) is an ongoing methodological crisis in which the results of many scientific studies are difficult or impossible to reproduce. Because the reproducibi ...
have bolstered the paper's credibility, and led to calls for methodological reforms in scientific research. In commentaries and technical responses, statisticians Goodman and Greenland identified several weaknesses in Ioannidis' model. Ioannidis's use of dramatic and exaggerated language that he "proved" that most research findings' claims are false and that "most research findings are false for ''most research designs'' and for ''most fields''" talics addedwas rejected, and yet they agreed with his paper's conclusions and recommendations. Biostatisticians Jager and Leek criticized the model as being based on justifiable but arbitrary assumptions rather than empirical data, and did an investigation of their own which calculated that the false positive rate in biomedical studies was estimated to be around 14%, not over 50% as Ioannidis asserted. Their paper was published in a 2014 special edition of the journal ''Biostatistics'' along with extended, supporting critiques from other statisticians. Leek summarized the key points of agreement as: when talking about the science-wise false discovery rate one has to bring data; there are different frameworks for estimating the science-wise false discovery rate; and "it is pretty unlikely that most published research is false," but that probably varies by one's definition of "most" and "false". Statistician Ulrich Schimmack reinforced the importance of the empirical basis for models by noting the reported false discovery rate in some scientific fields is not the actual discovery rate because non-significant results are rarely reported. Ioannidis's theoretical model fails to account for that, but when a statistical method ("z-curve") to estimate the number of unpublished non-significant results is applied to two examples, the false positive rate is between 8% and 17%, not greater than 50%.


Causes of high false positive rate

Despite these weaknesses there is nonetheless general agreement with the problem and recommendations Ioannidis discusses, yet his tone has been described as "dramatic" and "alarmingly misleading", which runs the risk of making people unnecessarily skeptical or cynical about science. A lasting impact of this work has been awareness of the underlying drivers of the high false positive rate in clinical medicine and biomedical research, and efforts by journals and scientists to mitigate them. Ioannidis restated these drivers in 2016 as being:{{cite web , last1=Minikel , first1=Eric V. , title=John Ioannidis: The state of research on research , url=https://www.cureffi.org/2016/03/17/john-ioannidis-the-state-of-research-on-research/ , website=www.cureffi.org , archive-url=https://web.archive.org/web/20200117053609/https://www.cureffi.org/2016/03/17/john-ioannidis-the-state-of-research-on-research/ , archive-date=17 January 2020 , date=17 March 2016 * Solo, siloed investigator limited to small sample sizes * No
preregistration Preregistration is the practice of registering the hypotheses, methods, and/or analyses of a scientific study before it is conducted. This can include analyzing primary data or secondary data. Clinical trial registration is similar, although it may ...
of hypotheses being tested * Post-hoc
cherry picking Cherry picking, suppressing evidence, or the fallacy of incomplete evidence is the act of pointing to individual cases or data that seem to confirm a particular position while ignoring a significant portion of related and similar cases or data th ...
of hypotheses with best P values * Only requiring P < .05 * No replication * No
data sharing Data sharing is the practice of making data used for scholarly research available to other investigators. Many funding agencies, institutions, and publication venues have policies regarding data sharing because transparency and openness are consid ...


See also

* Bayes' theorem *
Metascience Metascience (also known as meta-research) is the use of scientific methodology to study science itself. Metascience seeks to increase the quality of scientific research while reducing inefficiency. It is also known as "''research on research''" ...
*
Replication crisis The replication crisis (also called the replicability crisis and the reproducibility crisis) is an ongoing methodological crisis in which the results of many scientific studies are difficult or impossible to reproduce. Because the reproducibi ...
**
Berkeley Initiative for Transparency in the Social Sciences The Berkeley Initiative for Transparency in the Social Sciences, abbreviated BITSS, is an academic initiative dedicated to advancing Transparency (behavior), transparency, reproducibility, and openness in social science research. It was establishe ...
** Data dredging ** Publication bias ** Reproducibility Project


References


Further reading

* Carnegie Mellon University, Statistics Journal Club
''Summary and discussion of: “Why Most Published Research Findings Are False”''
* Applications to Economics
''De Long, J. Bradford; Lang, Kevin. "Are all Economic Hypotheses False?" Journal of Political Economy. 100 (6): 1257–1272, 1992''
* Applications to Social Sciences
''Hardwicke, Tom E.; Wallach, Joshua D.; Kidwell, Mallory C.; Bendixen, Theiss; Crüwell Sophia and Ioannidis, John P. A. "An empirical assessment of transparency and reproducibility-related research practices in the social sciences (2014–2017)." Royal Society Open Science. 7: 190806, 2020.''


External links

* YouTube video(s) from the ''Berkeley Initiative for Transparency in the Social Sciences'', 2016, "Why Most Published Research Findings are False"
Part IPart IIPart III
* YouTube video of John Ioannidis at ''Talks at Google'', 201
"Reproducible Research: True or False?"
2005 essays Academic journal articles Applied probability Criticism of academia Metascience Scientific method