In
probability theory
Probability theory is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expressing it through a set o ...
and
statistics
Statistics (from German language, German: ''wikt:Statistik#German, Statistik'', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of ...
, a conditional variance is the
variance
In probability theory and statistics, variance is the expectation of the squared deviation of a random variable from its population mean or sample mean. Variance is a measure of dispersion, meaning it is a measure of how far a set of numbers ...
of a
random variable
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
given the value(s) of one or more other variables.
Particularly in
econometrics
Econometrics is the application of Statistics, statistical methods to economic data in order to give Empirical evidence, empirical content to economic relationships.M. Hashem Pesaran (1987). "Econometrics," ''The New Palgrave: A Dictionary of ...
, the conditional variance is also known as the scedastic function or skedastic function.
Conditional variances are important parts of
autoregressive conditional heteroskedasticity
In econometrics, the autoregressive conditional heteroskedasticity (ARCH) model is a statistical model for time series data that describes the variance of the current error term or innovation as a function of the actual sizes of the previous ti ...
(ARCH) models.
Definition
The conditional variance of a
random variable
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
''Y'' given another random variable ''X'' is
:
The conditional variance tells us how much variance is left if we use
to "predict" ''Y''.
Here, as usual,
stands for the
conditional expectation
In probability theory, the conditional expectation, conditional expected value, or conditional mean of a random variable is its expected value – the value it would take “on average” over an arbitrarily large number of occurrences – give ...
of ''Y'' given ''X'',
which we may recall, is a random variable itself (a function of ''X'', determined up to probability one).
As a result,
itself is a random variable (and is a function of ''X'').
Explanation, relation to least-squares
Recall that variance is the expected squared deviation between a random variable (say, ''Y'') and its expected value.
The expected value can be thought of as a reasonable prediction of the outcomes of the random experiment (in particular, the expected value is the best constant prediction when predictions are assessed by expected squared prediction error). Thus, one interpretation of variance is that it gives the smallest possible expected squared prediction error. If we have the knowledge of another random variable (''X'') that we can use to predict ''Y'', we can potentially use this knowledge to reduce the expected squared error. As it turns out, the best prediction of ''Y'' given ''X'' is the conditional expectation. In particular, for any
measurable,
:
By selecting
, the second, nonnegative term becomes zero, showing the claim.
Here, the second equality used the
law of total expectation
The proposition in probability theory known as the law of total expectation, the law of iterated expectations (LIE), Adam's law, the tower rule, and the smoothing theorem, among other names, states that if X is a random variable whose expected v ...
.
We also see that the expected conditional variance of ''Y'' given ''X'' shows up as the irreducible error of predicting ''Y'' given only the knowledge of ''X''.
Special cases, variations
Conditioning on discrete random variables
When ''X'' takes on countable many values
with positive probability, i.e., it is a
discrete random variable
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
, we can introduce
, the conditional variance of ''Y'' given that ''X=x'' for any ''x'' from ''S'' as follows:
:
where recall that
is the
conditional expectation of ''Z'' given that ''X=x'', which is well-defined for
.
An alternative notation for
is
Note that here
defines a constant for possible values of ''x'', and in particular,
, is ''not'' a random variable.
The connection of this definition to
is as follows:
Let ''S'' be as above and define the function
as
. Then,
almost surely
In probability theory, an event is said to happen almost surely (sometimes abbreviated as a.s.) if it happens with probability 1 (or Lebesgue measure 1). In other words, the set of possible exceptions may be non-empty, but it has probability 0. ...
.
Definition using conditional distributions
The "conditional expectation of ''Y'' given ''X=x''" can also be defined more generally
using the
conditional distribution
In probability theory and statistics, given two jointly distributed random variables X and Y, the conditional probability distribution of Y given X is the probability distribution of Y when X is known to be a particular value; in some cases the ...
of ''Y'' given ''X'' (this exists in this case, as both here ''X'' and ''Y'' are real-valued).
In particular, letting
be the (regular)
conditional distribution
In probability theory and statistics, given two jointly distributed random variables X and Y, the conditional probability distribution of Y given X is the probability distribution of Y when X is known to be a particular value; in some cases the ...
of ''Y'' given ''X'', i.e.,