Conditional Density
   HOME

TheInfoList



OR:

In
probability theory Probability theory is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expressing it through a set o ...
and
statistics Statistics (from German language, German: ''wikt:Statistik#German, Statistik'', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of ...
, given two jointly distributed
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
s X and Y, the conditional probability distribution of Y given X is the
probability distribution In probability theory and statistics, a probability distribution is the mathematical function that gives the probabilities of occurrence of different possible outcomes for an experiment. It is a mathematical description of a random phenomenon i ...
of Y when X is known to be a particular value; in some cases the conditional probabilities may be expressed as functions containing the unspecified value x of X as a parameter. When both X and Y are
categorical variable In statistics, a categorical variable (also called qualitative variable) is a variable that can take on one of a limited, and usually fixed, number of possible values, assigning each individual or other unit of observation to a particular group or ...
s, a
conditional probability table In statistics, the conditional probability table (CPT) is defined for a set of discrete and mutually dependent random variables to display conditional probabilities of a single variable with respect to the others (i.e., the probability of each pos ...
is typically used to represent the conditional probability. The conditional distribution contrasts with the
marginal distribution In probability theory and statistics, the marginal distribution of a subset of a collection of random variables is the probability distribution of the variables contained in the subset. It gives the probabilities of various values of the variables ...
of a random variable, which is its distribution without reference to the value of the other variable. If the conditional distribution of Y given X is a
continuous distribution In probability theory and statistics, a probability distribution is the mathematical Function (mathematics), function that gives the probabilities of occurrence of different possible outcomes for an Experiment (probability theory), experiment. ...
, then its
probability density function In probability theory, a probability density function (PDF), or density of a continuous random variable, is a function whose value at any given sample (or point) in the sample space (the set of possible values taken by the random variable) can ...
is known as the conditional density function. The properties of a conditional distribution, such as the moments, are often referred to by corresponding names such as the
conditional mean In probability theory, the conditional expectation, conditional expected value, or conditional mean of a random variable is its expected value – the value it would take “on average” over an arbitrarily large number of occurrences – given ...
and
conditional variance In probability theory and statistics, a conditional variance is the variance of a random variable given the value(s) of one or more other variables. Particularly in econometrics, the conditional variance is also known as the scedastic function or ...
. More generally, one can refer to the conditional distribution of a subset of a set of more than two variables; this conditional distribution is contingent on the values of all the remaining variables, and if more than one variable is included in the subset then this conditional distribution is the conditional
joint distribution Given two random variables that are defined on the same probability space, the joint probability distribution is the corresponding probability distribution on all possible pairs of outputs. The joint distribution can just as well be considered ...
of the included variables.


Conditional discrete distributions

For
discrete random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
s, the conditional probability mass function of Y given X=x can be written according to its definition as: Due to the occurrence of P(X=x) in the denominator, this is defined only for non-zero (hence strictly positive) P(X=x). The relation with the probability distribution of X given Y is: :P(Y=y \mid X=x) P(X=x) = P(\ \cap \) = P(X=x \mid Y=y)P(Y=y).


Example

Consider the roll of a fair and let X=1 if the number is even (i.e., 2, 4, or 6) and X=0 otherwise. Furthermore, let Y=1 if the number is prime (i.e., 2, 3, or 5) and Y=0 otherwise. Then the unconditional probability that X=1 is 3/6 = 1/2 (since there are six possible rolls of the dice, of which three are even), whereas the probability that X=1 conditional on Y=1 is 1/3 (since there are three possible prime number rolls—2, 3, and 5—of which one is even).


Conditional continuous distributions

Similarly for
continuous random variable In probability theory and statistics, a probability distribution is the mathematical function that gives the probabilities of occurrence of different possible outcomes for an experiment. It is a mathematical description of a random phenomenon i ...
s, the conditional
probability density function In probability theory, a probability density function (PDF), or density of a continuous random variable, is a function whose value at any given sample (or point) in the sample space (the set of possible values taken by the random variable) can ...
of Y given the occurrence of the value x of X can be written as where f_(x,y) gives the joint density of X and Y, while f_X(x) gives the marginal density for X. Also in this case it is necessary that f_X(x)>0. The relation with the probability distribution of X given Y is given by: :f_(y \mid x)f_X(x) = f_(x, y) = f_(x \mid y)f_Y(y). The concept of the conditional distribution of a continuous random variable is not as intuitive as it might seem: Borel's paradox shows that conditional probability density functions need not be invariant under coordinate transformations.


Example

The graph shows a bivariate normal joint density for random variables X and Y. To see the distribution of Y conditional on X=70, one can first visualize the line X=70 in the X,Y
plane Plane(s) most often refers to: * Aero- or airplane, a powered, fixed-wing aircraft * Plane (geometry), a flat, 2-dimensional surface Plane or planes may also refer to: Biology * Plane (tree) or ''Platanus'', wetland native plant * ''Planes' ...
, and then visualize the plane containing that line and perpendicular to the X,Y plane. The intersection of that plane with the joint normal density, once rescaled to give unit area under the intersection, is the relevant conditional density of Y. Y\mid X=70 \ \sim\ \mathcal\left(\mu_1+\frac\rho( 70 - \mu_2),\, (1-\rho^2)\sigma_1^2\right).


Relation to independence

Random variables X, Y are
independent Independent or Independents may refer to: Arts, entertainment, and media Artist groups * Independents (artist group), a group of modernist painters based in the New Hope, Pennsylvania, area of the United States during the early 1930s * Independ ...
if and only if the conditional distribution of Y given X is, for all possible realizations of X, equal to the unconditional distribution of Y. For discrete random variables this means P(Y=y, X=x) = P(Y=y) for all possible y and x with P(X=x)>0. For continuous random variables X and Y, having a joint density function, it means f_Y(y, X=x) = f_Y(y) for all possible y and x with f_X(x)>0.


Properties

Seen as a function of y for given x, P(Y=y, X=x) is a probability mass function and so the sum over all y (or integral if it is a conditional probability density) is 1. Seen as a function of x for given y, it is a
likelihood function The likelihood function (often simply called the likelihood) represents the probability of random variable realizations conditional on particular values of the statistical parameters. Thus, when evaluated on a given sample, the likelihood funct ...
, so that the sum over all x need not be 1. Additionally, a marginal of a joint distribution can be expressed as the expectation of the corresponding conditional distribution. For instance, p_X(x) = E_ \ Y).


Measure-theoretic formulation

Let (\Omega, \mathcal, P) be a probability space, \mathcal \subseteq \mathcal a \sigma-field in \mathcal. Given A\in \mathcal, the Radon-Nikodym theorem implies that there is a \mathcal-measurable random variable P(A\mid\mathcal):\Omega\to \mathbb, called the conditional probability, such that\int_G P(A\mid\mathcal)(\omega) dP(\omega)=P(A\cap G)for every G\in \mathcal, and such a random variable is uniquely defined up to sets of probability zero. A conditional probability is called regular if \operatorname(\cdot\mid\mathcal)(\omega) is a
probability measure In mathematics, a probability measure is a real-valued function defined on a set of events in a probability space that satisfies measure properties such as ''countable additivity''. The difference between a probability measure and the more gener ...
on (\Omega, \mathcal) for all \omega \in \Omega a.e. Special cases: * For the trivial sigma algebra \mathcal G= \, the conditional probability is the constant function \operatorname\!\left( A\mid \ \right) = \operatorname(A). * If A\in \mathcal, then \operatorname(A\mid\mathcal)=1_A, the indicator function (defined below). Let X : \Omega \to E be a (E, \mathcal)-valued random variable. For each B \in \mathcal, define \mu_ (B \, , \, \mathcal) = \mathrm (X^(B) \, , \, \mathcal).For any \omega \in \Omega, the function \mu_(\cdot \, , \mathcal) (\omega) : \mathcal \to \mathbb is called the
conditional probability In probability theory, conditional probability is a measure of the probability of an event occurring, given that another event (by assumption, presumption, assertion or evidence) has already occurred. This particular method relies on event B occur ...
distribution of X given \mathcal. If it is a probability measure on (E, \mathcal), then it is called regular. For a real-valued random variable (with respect to the Borel \sigma-field \mathcal^1 on \mathbb), every conditional probability distribution is regular. Billingsley (1995), p. 439 In this case,E \mid \mathcal= \int_^\infty x \, \mu(d x, \cdot) almost surely.


Relation to conditional expectation

For any event A \in \mathcal, define the
indicator function In mathematics, an indicator function or a characteristic function of a subset of a set is a function that maps elements of the subset to one, and all other elements to zero. That is, if is a subset of some set , one has \mathbf_(x)=1 if x\i ...
: :\mathbf_A (\omega) = \begin 1 \; &\text \omega \in A, \\ 0 \; &\text \omega \notin A, \end which is a random variable. Note that the expectation of this random variable is equal to the probability of ''A'' itself: :\operatorname(\mathbf_A) = \operatorname(A). \; Given a \sigma-field \mathcal \subseteq \mathcal, the conditional probability \operatorname(A\mid\mathcal) is a version of the
conditional expectation In probability theory, the conditional expectation, conditional expected value, or conditional mean of a random variable is its expected value – the value it would take “on average” over an arbitrarily large number of occurrences – give ...
of the indicator function for A: :\operatorname(A\mid\mathcal) = \operatorname(\mathbf_A\mid\mathcal) \; An expectation of a random variable with respect to a regular conditional probability is equal to its conditional expectation.


See also

*
Conditioning (probability) Beliefs depend on the available information. This idea is formalized in probability theory by conditioning. Conditional probabilities, conditional expectations, and conditional probability distributions are treated on three levels: discrete proba ...
*
Conditional probability In probability theory, conditional probability is a measure of the probability of an event occurring, given that another event (by assumption, presumption, assertion or evidence) has already occurred. This particular method relies on event B occur ...
*
Regular conditional probability In probability theory, regular conditional probability is a concept that formalizes the notion of conditioning on the outcome of a random variable. The resulting conditional probability distribution is a parametrized family of probability measures c ...
*
Bayes' theorem In probability theory and statistics, Bayes' theorem (alternatively Bayes' law or Bayes' rule), named after Thomas Bayes, describes the probability of an event, based on prior knowledge of conditions that might be related to the event. For examp ...


References


Citations


Sources

* {{refend Theory of probability distributions
Distribution Distribution may refer to: Mathematics *Distribution (mathematics), generalized functions used to formulate solutions of partial differential equations * Probability distribution, the probability of a particular value or value range of a vari ...