Estimation Statistics
   HOME

TheInfoList



OR:

Estimation statistics, or simply estimation, is a data analysis framework that uses a combination of
effect size In statistics, an effect size is a value measuring the strength of the relationship between two variables in a population, or a sample-based estimate of that quantity. It can refer to the value of a statistic calculated from a sample of data, the ...
s,
confidence interval In frequentist statistics, a confidence interval (CI) is a range of estimates for an unknown parameter. A confidence interval is computed at a designated ''confidence level''; the 95% confidence level is most common, but other levels, such as 9 ...
s, precision planning, and
meta-analysis A meta-analysis is a statistical analysis that combines the results of multiple scientific studies. Meta-analyses can be performed when there are multiple scientific studies addressing the same question, with each individual study reporting me ...
to plan experiments, analyze data and interpret results. It complements hypothesis testing approaches such as null hypothesis significance testing (NHST), by going beyond the question is an effect present or not, and provides information about how large an effect is. Estimation statistics is sometimes referred to as ''the new statistics''. The primary aim of estimation methods is to report an
effect size In statistics, an effect size is a value measuring the strength of the relationship between two variables in a population, or a sample-based estimate of that quantity. It can refer to the value of a statistic calculated from a sample of data, the ...
(a
point estimate In statistics, point estimation involves the use of sample data to calculate a single value (known as a point estimate since it identifies a point in some parameter space) which is to serve as a "best guess" or "best estimate" of an unknown popu ...
) along with its
confidence interval In frequentist statistics, a confidence interval (CI) is a range of estimates for an unknown parameter. A confidence interval is computed at a designated ''confidence level''; the 95% confidence level is most common, but other levels, such as 9 ...
, the latter of which is related to the precision of the estimate. The confidence interval summarizes a range of likely values of the underlying population effect. Proponents of estimation see reporting a ''P'' value as an unhelpful distraction from the important business of reporting an effect size with its confidence intervals, and believe that estimation should replace significance testing for data analysis.


History

Starting in 1929, physicist
Raymond Thayer Birge Raymond Thayer Birge (March 13, 1887 – March 22, 1980) was an American physicist. Career Born in Brooklyn, New York, into an academic scientific family, Birge obtained his doctorate from the University of Wisconsin in 1913. In the same year h ...
published review papers in which he used weighted-averages methods to calculate estimates of physical constants, a procedure that can be seen as the precursor to modern
meta-analysis A meta-analysis is a statistical analysis that combines the results of multiple scientific studies. Meta-analyses can be performed when there are multiple scientific studies addressing the same question, with each individual study reporting me ...
. In the 1960s, estimation statistics was adopted by the non-physical sciences with the development of the standardized effect size by Jacob Cohen. In the 1970s, modern research synthesis was pioneered by
Gene V. Glass Gene V Glass (born June 19, 1940) is an American statistician and researcher working in educational psychology and the social sciences. According to the science writer Morton Hunt, he coined the term "meta-analysis" and illustrated its first us ...
with the first systematic review and
meta-analysis A meta-analysis is a statistical analysis that combines the results of multiple scientific studies. Meta-analyses can be performed when there are multiple scientific studies addressing the same question, with each individual study reporting me ...
for psychotherapy. This pioneering work subsequently influenced the adoption of meta-analyses for medical treatments more generally. In the 1980s and 1990s, estimation methods were extended and refined by biostatisticians including
Larry Hedges Larry Vernon Hedges is a researcher in statistical methods for meta-analysis and evaluation of education policy. He is Professor of Statistics and Education and Social Policy, Institute for Policy Research, Northwestern University. Previously, he ...
, Michael Borenstein,
Doug Altman Douglas Graham Altman FMedSci (12 July 1948 – 3 June 2018) was an English statistician best known for his work on improving the reliability and reporting of medical research and for highly cited papers on statistical methodology. He was profe ...
, Martin Gardner, and many others, with the development of the modern (medical
meta-analysis
Starting in the 1980s, the
systematic review A systematic review is a Literature review, scholarly synthesis of the evidence on a clearly presented topic using critical methods to identify, define and assess research on the topic. A systematic review extracts and interprets data from publ ...
, used in conjunction with meta-analysis, became a technique widely used in medical research. There are over 200,00
citations
to "meta-analysis" in
PubMed PubMed is a free search engine accessing primarily the MEDLINE database of references and abstracts on life sciences and biomedical topics. The United States National Library of Medicine (NLM) at the National Institutes of Health maintain the ...
. In the 1990s, edito
Kenneth Rothman
banned the use of p-values from the journal ''Epidemiology''; compliance was high among authors but this did not substantially change their analytical thinking. In the 2010s, Geoff Cumming published
textbook
dedicated to estimation statistics, along with software in Excel designed to teach effect-size thinking, primarily to psychologists. Also in the 2010s, estimation methods were increasingly adopted in neuroscience. In 2013, the Publication Manual of the American Psychological Association recommended to use estimation in addition to hypothesis testing. Also in 2013, the Uniform Requirements for Manuscripts Submitted to Biomedical Journals document made a similar recommendation: "Avoid relying solely on statistical hypothesis testing, such as P values, which fail to convey important information about effect size." In 2019, over 800 scientists signed an open comment calling for the entire concept of statistical significance to be abandoned. In 2019, the
Society for Neuroscience The Society for Neuroscience (SfN) is a professional society, headquartered in Washington, DC, for basic scientists and physicians around the world whose research is focused on the study of the brain and nervous system. It is especially well kn ...
journal eNeuro instituted a policy recommending the use of estimation graphics as the preferred method for data presentation. And in 2022, the International Society of Physiotherapy Journal Editors recommended the use of estimation methods instead of null hypothesis statistical tests. Despite the widespread adoption of meta-analysis for clinical research, and recommendations by several major publishing institutions, the estimation framework is not routinely used in primary biomedical research.


Methodology

Many significance tests have an estimation counterpart; in almost every case, the test result (or its
p-value In null-hypothesis significance testing, the ''p''-value is the probability of obtaining test results at least as extreme as the result actually observed, under the assumption that the null hypothesis is correct. A very small ''p''-value means ...
) can be simply substituted with the effect size and a precision estimate. For example, instead of using
Student's t-test A ''t''-test is any statistical hypothesis test in which the test statistic follows a Student's ''t''-distribution under the null hypothesis. It is most commonly applied when the test statistic would follow a normal distribution if the value of ...
, the analyst can compare two independent groups by calculating the mean difference and its 95%
confidence interval In frequentist statistics, a confidence interval (CI) is a range of estimates for an unknown parameter. A confidence interval is computed at a designated ''confidence level''; the 95% confidence level is most common, but other levels, such as 9 ...
. Corresponding methods can be used for a
paired t-test A ''t''-test is any statistical hypothesis test in which the test statistic follows a Student's ''t''-distribution under the null hypothesis. It is most commonly applied when the test statistic would follow a normal distribution if the value of a ...
and multiple comparisons. Similarly, for a regression analysis, an analyst would report the
coefficient of determination In statistics, the coefficient of determination, denoted ''R''2 or ''r''2 and pronounced "R squared", is the proportion of the variation in the dependent variable that is predictable from the independent variable(s). It is a statistic used i ...
(R2) and the model equation instead of the model's p-value. However, proponents of estimation statistics warn against reporting only a few numbers. Rather, it is advised to analyze and present data using data visualization. Examples of appropriate visualizations include the
Scatter plot A scatter plot (also called a scatterplot, scatter graph, scatter chart, scattergram, or scatter diagram) is a type of plot or mathematical diagram using Cartesian coordinates to display values for typically two variables for a set of data. ...
for regression, and Gardner-Altman plots for two independent groups. While historical data-group plots (bar charts, box plots, and violin plots) do not display the comparison, estimation plots add a second axis to explicitly visualize the effect size.


Gardner–Altman plot

The Gardner–Altman mean difference plot was first described by
Martin Gardner Martin Gardner (October 21, 1914May 22, 2010) was an American popular mathematics and popular science writer with interests also encompassing scientific skepticism, micromagic, philosophy, religion, and literatureespecially the writings of Lewis ...
and
Doug Altman Douglas Graham Altman FMedSci (12 July 1948 – 3 June 2018) was an English statistician best known for his work on improving the reliability and reporting of medical research and for highly cited papers on statistical methodology. He was profe ...
in 1986; it is a statistical graph designed to display data from two independent groups. There is also a version suitable fo
paired data
The key instructions to make this chart are as follows: (1) display all observed values for both groups side-by-side; (2) place a second axis on the right, shifted to show the mean difference scale; and (3) plot the mean difference with its confidence interval as a marker with error bars. Gardner-Altman plots can be generated wit
DABEST-Python
o
dabestr
alternatively, the analyst can use GUI software like th
Estimation Stats
app.


Cumming plot

For multiple groups
Geoff Cumming
introduced the use of a secondary panel to plot two or more mean differences and their confidence intervals, placed below the observed values panel; this arrangement enable
easy comparison
of mean differences ('deltas') over several data groupings. Cumming plots can be generated with th
ESCI packageDABEST
or th
Estimation Stats app


Other methodologies

In addition to the mean difference, there are numerous other
effect size In statistics, an effect size is a value measuring the strength of the relationship between two variables in a population, or a sample-based estimate of that quantity. It can refer to the value of a statistic calculated from a sample of data, the ...
types, all with relative benefits. Major types include effect sizes in the Cohen's ''d'' class of standardized metrics, and the
coefficient of determination In statistics, the coefficient of determination, denoted ''R''2 or ''r''2 and pronounced "R squared", is the proportion of the variation in the dependent variable that is predictable from the independent variable(s). It is a statistic used i ...
(R2) for
regression analysis In statistical modeling, regression analysis is a set of statistical processes for estimating the relationships between a dependent variable (often called the 'outcome' or 'response' variable, or a 'label' in machine learning parlance) and one ...
. For non-normal distributions, there are a number of mor
robust effect sizes
including Cliff's delta and the Kolmogorov-Smirnov statistic.


Flaws in hypothesis testing

In
hypothesis testing A statistical hypothesis test is a method of statistical inference used to decide whether the data at hand sufficiently support a particular hypothesis. Hypothesis testing allows us to make probabilistic statements about population parameters. ...
, the primary objective of statistical calculations is to obtain a
p-value In null-hypothesis significance testing, the ''p''-value is the probability of obtaining test results at least as extreme as the result actually observed, under the assumption that the null hypothesis is correct. A very small ''p''-value means ...
, the probability of seeing an obtained result, or a more extreme result, when assuming the
null hypothesis In scientific research, the null hypothesis (often denoted ''H''0) is the claim that no difference or relationship exists between two sets of data or variables being analyzed. The null hypothesis is that any experimentally observed difference is d ...
is true. If the p-value is low (usually < 0.05), the statistical practitioner is then encouraged to reject the null hypothesis. Proponents of estimation reject the validity of hypothesis testing for the following reasons, among others: * P-values are easily and commonly misinterpreted. For example, the p-value is often mistakenly thought of as 'the probability that the null hypothesis is true.' * The null hypothesis is always wrong for every set of observations: there is always some effect, even if it is minuscule. * Hypothesis testing produces dichotomous yes-no answers, while discarding important information about magnitude. * Any particular p-value arises through the interaction of the
effect size In statistics, an effect size is a value measuring the strength of the relationship between two variables in a population, or a sample-based estimate of that quantity. It can refer to the value of a statistic calculated from a sample of data, the ...
, the
sample size Sample size determination is the act of choosing the number of observations or Replication (statistics), replicates to include in a statistical sample. The sample size is an important feature of any empirical study in which the goal is to make stat ...
(all things being equal a larger sample size produces a smaller p-value) and sampling error. *At low
power Power most often refers to: * Power (physics), meaning "rate of doing work" ** Engine power, the power put out by an engine ** Electric power * Power (social and political), the ability to influence people or events ** Abusive power Power may a ...
, simulation reveals that sampling error makes p-values extremely volatile.


Benefits of estimation statistics


Advantages of confidence intervals

Confidence intervals behave in a predictable way. By definition, 95% confidence intervals have a 95% chance of covering the underlying population mean (μ). This feature remains constant with increasing sample size; what changes is that the interval becomes smaller. In addition, 95% confidence intervals are also 83% prediction intervals: one (pre experimental) confidence interval has an 83% chance of covering any future experiment's mean. As such, knowing a single experiment's 95% confidence intervals gives the analyst a reasonable range for the population mean. Nevertheless, confidence distributions and posterior distributions provide a whole lot more information than a single point estimate or intervals, that can exacerbate dichotomous thinking according to the interval covering or not covering a "null" value of interest (i.e. the Inductive behavior of Neyman as opposed to that of Fisher).


Evidence-based statistics

Psychological studies of the perception of statistics reveal that reporting interval estimates leaves a more accurate perception of the data than reporting p-values.


Precision planning

The precision of an estimate is formally defined as 1/
variance In probability theory and statistics, variance is the expectation of the squared deviation of a random variable from its population mean or sample mean. Variance is a measure of dispersion, meaning it is a measure of how far a set of numbers ...
, and like power, increases (improves) with increasing sample size. Like
power Power most often refers to: * Power (physics), meaning "rate of doing work" ** Engine power, the power put out by an engine ** Electric power * Power (social and political), the ability to influence people or events ** Abusive power Power may a ...
, a high level of precision is expensive; research grant applications would ideally include precision/cost analyses. Proponents of estimation believe precision planning should replace
power Power most often refers to: * Power (physics), meaning "rate of doing work" ** Engine power, the power put out by an engine ** Electric power * Power (social and political), the ability to influence people or events ** Abusive power Power may a ...
since statistical power itself is conceptually linked to significance testing. Precision planning can be done with th
ESCI web app


See also

*
Effect size In statistics, an effect size is a value measuring the strength of the relationship between two variables in a population, or a sample-based estimate of that quantity. It can refer to the value of a statistic calculated from a sample of data, the ...
*
Cohen's h In statistics, Cohen's ''h'', popularized by Jacob Cohen, is a measure of distance between two proportions or probabilities. Cohen's ''h'' has several related uses: * It can be used to describe the difference between two proportions as "small", ...
*
Interval estimation In statistics, interval estimation is the use of sample data to estimate an '' interval'' of plausible values of a parameter of interest. This is in contrast to point estimation, which gives a single value. The most prevalent forms of interval es ...
*
Meta-analysis A meta-analysis is a statistical analysis that combines the results of multiple scientific studies. Meta-analyses can be performed when there are multiple scientific studies addressing the same question, with each individual study reporting me ...
*
Statistical significance In statistical hypothesis testing, a result has statistical significance when it is very unlikely to have occurred given the null hypothesis (simply by chance alone). More precisely, a study's defined significance level, denoted by \alpha, is the p ...


References

{{Statistics Estimation theory Effect size