In
probability theory
Probability theory is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expressing it through a set o ...
, there exist several different notions of convergence of random variables. The
convergence
Convergence may refer to:
Arts and media Literature
*''Convergence'' (book series), edited by Ruth Nanda Anshen
* "Convergence" (comics), two separate story lines published by DC Comics:
**A four-part crossover storyline that united the four Weir ...
of
sequence
In mathematics, a sequence is an enumerated collection of objects in which repetitions are allowed and order matters. Like a set, it contains members (also called ''elements'', or ''terms''). The number of elements (possibly infinite) is called ...
s of
random variable
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the p ...
s to some
limit
Limit or Limits may refer to:
Arts and media
* ''Limit'' (manga), a manga by Keiko Suenobu
* ''Limit'' (film), a South Korean film
* Limit (music), a way to characterize harmony
* "Limit" (song), a 2016 single by Luna Sea
* "Limits", a 2019 ...
random variable is an important concept in probability theory, and its applications to
statistics and
stochastic processes. The same concepts are known in more general
mathematics as stochastic convergence and they formalize the idea that a sequence of essentially random or unpredictable events can sometimes be expected to settle down into a behavior that is essentially unchanging when items far enough into the sequence are studied. The different possible notions of convergence relate to how such a behavior can be characterized: two readily understood behaviors are that the sequence eventually takes a constant value, and that values in the sequence continue to change but can be described by an unchanging probability distribution.
Background
"Stochastic convergence" formalizes the idea that a sequence of essentially random or unpredictable events can sometimes be expected to settle into a pattern. The pattern may for instance be
*
Convergence
Convergence may refer to:
Arts and media Literature
*''Convergence'' (book series), edited by Ruth Nanda Anshen
* "Convergence" (comics), two separate story lines published by DC Comics:
**A four-part crossover storyline that united the four Weir ...
in the classical sense to a fixed value, perhaps itself coming from a random event
*An increasing similarity of outcomes to what a purely deterministic function would produce
*An increasing preference towards a certain outcome
*An increasing "aversion" against straying far away from a certain outcome
*That the probability distribution describing the next outcome may grow increasingly similar to a certain distribution
Some less obvious, more theoretical patterns could be
*That the series formed by calculating the
expected value
In probability theory, the expected value (also called expectation, expectancy, mathematical expectation, mean, average, or first moment) is a generalization of the weighted average. Informally, the expected value is the arithmetic mean of a ...
of the outcome's distance from a particular value may converge to 0
*That the variance of the
random variable
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the p ...
describing the next event grows smaller and smaller.
These other types of patterns that may arise are reflected in the different types of stochastic convergence that have been studied.
While the above discussion has related to the convergence of a single series to a limiting value, the notion of the convergence of two series towards each other is also important, but this is easily handled by studying the sequence defined as either the difference or the ratio of the two series.
For example, if the average of ''n''
independent
Independent or Independents may refer to:
Arts, entertainment, and media Artist groups
* Independents (artist group), a group of modernist painters based in the New Hope, Pennsylvania, area of the United States during the early 1930s
* Independe ...
random variables ''Y''
''i'', ''i'' = 1, ..., ''n'', all having the same finite
mean
There are several kinds of mean in mathematics, especially in statistics. Each mean serves to summarize a given group of data, often to better understand the overall value ( magnitude and sign) of a given data set.
For a data set, the '' ari ...
and
variance
In probability theory and statistics, variance is the expectation of the squared deviation of a random variable from its population mean or sample mean. Variance is a measure of dispersion, meaning it is a measure of how far a set of number ...
, is given by
:
then as ''n'' tends to infinity, converges ''in probability'' (see below) to the common
mean
There are several kinds of mean in mathematics, especially in statistics. Each mean serves to summarize a given group of data, often to better understand the overall value ( magnitude and sign) of a given data set.
For a data set, the '' ari ...
, μ, of the random variables ''Y''
''i''. This result is known as the
weak law of large numbers
In probability theory, the law of large numbers (LLN) is a theorem that describes the result of performing the same experiment a large number of times. According to the law, the average of the results obtained from a large number of trials shou ...
. Other forms of convergence are important in other useful theorems, including the
central limit theorem
In probability theory, the central limit theorem (CLT) establishes that, in many situations, when independent random variables are summed up, their properly normalized sum tends toward a normal distribution even if the original variables thems ...
.
Throughout the following, we assume that (''X''
''n'') is a sequence of random variables, and ''X'' is a random variable, and all of them are defined on the same
probability space
In probability theory, a probability space or a probability triple (\Omega, \mathcal, P) is a mathematical construct that provides a formal model of a random process or "experiment". For example, one can define a probability space which models t ...
.
Convergence in distribution
With this mode of convergence, we increasingly expect to see the next outcome in a sequence of random experiments becoming better and better modeled by a given
probability distribution
In probability theory and statistics, a probability distribution is the mathematical function that gives the probabilities of occurrence of different possible outcomes for an experiment. It is a mathematical description of a random phenomeno ...
.
Convergence in distribution is the weakest form of convergence typically discussed, since it is implied by all other types of convergence mentioned in this article. However, convergence in distribution is very frequently used in practice; most often it arises from application of the
central limit theorem
In probability theory, the central limit theorem (CLT) establishes that, in many situations, when independent random variables are summed up, their properly normalized sum tends toward a normal distribution even if the original variables thems ...
.
Definition
A sequence
of real-valued
random variable
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the p ...
s, with
cumulative distribution function
In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x.
Ev ...
s
, is said to converge in distribution, or converge weakly, or converge in law to a random variable with
cumulative distribution function
In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x.
Ev ...
if
:
for every number
at which is
continuous.
The requirement that only the continuity points of should be considered is essential. For example, if are distributed
uniformly
Uniform distribution may refer to:
* Continuous uniform distribution
* Discrete uniform distribution
* Uniform distribution (ecology)
* Equidistributed sequence In mathematics, a sequence (''s''1, ''s''2, ''s''3, ...) of real numbers is said to be ...
on intervals , then this sequence converges in distribution to the
degenerate
Degeneracy, degenerate, or degeneration may refer to:
Arts and entertainment
* ''Degenerate'' (album), a 2010 album by the British band Trigger the Bloodshed
* Degenerate art, a term adopted in the 1920s by the Nazi Party in Germany to descr ...
random variable . Indeed,
for all ''n'' when , and for all when . However, for this limiting random variable , even though for all . Thus the convergence of cdfs fails at the point where is discontinuous.
Convergence in distribution may be denoted as
where
is the law (probability distribution) of . For example, if is standard normal we can write
.
For
random vector
In probability, and statistics, a multivariate random variable or random vector is a list of mathematical variables each of whose value is unknown, either because the value has not yet occurred or because there is imperfect knowledge of its valu ...
s the convergence in distribution is defined similarly. We say that this sequence converges in distribution to a random -vector if
:
for every which is a
continuity set In measure theory, a branch of mathematics, a continuity set of a measure ''μ'' is any Borel set ''B'' such that
: \mu(\partial B) = 0\,,
where \partial B is the (topological) boundary of ''B''. For signed measures, one asks that
: , \mu, ( ...
of .
The definition of convergence in distribution may be extended from random vectors to more general
random element In probability theory, random element is a generalization of the concept of random variable to more complicated spaces than the simple real line. The concept was introduced by who commented that the “development of probability theory and expansio ...
s in arbitrary
metric space
In mathematics, a metric space is a set together with a notion of '' distance'' between its elements, usually called points. The distance is measured by a function called a metric or distance function. Metric spaces are the most general sett ...
s, and even to the “random variables” which are not measurable — a situation which occurs for example in the study of
empirical process
In probability theory, an empirical process is a stochastic process that describes the proportion of objects in a system in a given state.
For a process in a discrete state space a population continuous time Markov chain or Markov population model ...
es. This is the “weak convergence of laws without laws being defined” — except asymptotically.
In this case the term weak convergence is preferable (see
weak convergence of measures), and we say that a sequence of random elements converges weakly to (denoted as ) if
:
for all continuous bounded functions . Here E* denotes the ''outer expectation'', that is the expectation of a “smallest measurable function that dominates ”.
Properties
* Since , the convergence in distribution means that the probability for to be in a given range is approximately equal to the probability that the value of is in that range, provided is
sufficiently large
In the mathematical areas of number theory and analysis, an infinite sequence or a function is said to eventually have a certain property, if it doesn't have the said property across all its ordered instances, but will after some instances have pa ...
.
*In general, convergence in distribution does not imply that the sequence of corresponding
probability density function
In probability theory, a probability density function (PDF), or density of a continuous random variable, is a function whose value at any given sample (or point) in the sample space (the set of possible values taken by the random variable) c ...
s will also converge. As an example one may consider random variables with densities . These random variables converge in distribution to a uniform ''U''(0, 1), whereas their densities do not converge at all.
** However, according to ''Scheffé’s theorem'', convergence of the
probability density function
In probability theory, a probability density function (PDF), or density of a continuous random variable, is a function whose value at any given sample (or point) in the sample space (the set of possible values taken by the random variable) c ...
s implies convergence in distribution.
* The
portmanteau lemma provides several equivalent definitions of convergence in distribution. Although these definitions are less intuitive, they are used to prove a number of statistical theorems. The lemma states that converges in distribution to if and only if any of the following statements are true:
**
for all continuity points of
;
**
for all
bounded
Boundedness or bounded may refer to:
Economics
* Bounded rationality, the idea that human rationality in decision-making is bounded by the available information, the cognitive limitations, and the time available to make the decision
* Bounded e ...
,
continuous functions
(where
denotes the
expected value
In probability theory, the expected value (also called expectation, expectancy, mathematical expectation, mean, average, or first moment) is a generalization of the weighted average. Informally, the expected value is the arithmetic mean of a ...
operator);
**
for all bounded,
Lipschitz functions
;
**
for all nonnegative, continuous functions
;
**
for every
open set
In mathematics, open sets are a generalization of open intervals in the real line.
In a metric space (a set along with a distance defined between any two points), open sets are the sets that, with every point , contain all points that a ...
;
**
for every
closed set
In geometry, topology, and related branches of mathematics, a closed set is a set whose complement is an open set. In a topological space, a closed set can be defined as a set which contains all its limit points. In a complete metric spac ...
;
**
for all
continuity set In measure theory, a branch of mathematics, a continuity set of a measure ''μ'' is any Borel set ''B'' such that
: \mu(\partial B) = 0\,,
where \partial B is the (topological) boundary of ''B''. For signed measures, one asks that
: , \mu, ( ...
s
of random variable
;
**
for every
upper semi-continuous function
bounded above;
**
for every
lower semi-continuous
In mathematical analysis, semicontinuity (or semi-continuity) is a property of extended real-valued functions that is weaker than continuity. An extended real-valued function f is upper (respectively, lower) semicontinuous at a point x_0 if, r ...
function
bounded below.
* The
continuous mapping theorem
In probability theory, the continuous mapping theorem states that continuous functions preserve limits even if their arguments are sequences of random variables. A continuous function, in Heine’s definition, is such a function that maps converg ...
states that for a continuous function , if the sequence converges in distribution to , then converges in distribution to .
** Note however that convergence in distribution of to and to does in general ''not'' imply convergence in distribution of to or of to .
*
Lévy’s continuity theorem: the sequence converges in distribution to if and only if the sequence of corresponding
characteristic function In mathematics, the term "characteristic function" can refer to any of several distinct concepts:
* The indicator function of a subset, that is the function
::\mathbf_A\colon X \to \,
:which for a given subset ''A'' of ''X'', has value 1 at point ...
s
converges pointwise to the characteristic function of .
* Convergence in distribution is
metrizable
In topology and related areas of mathematics, a metrizable space is a topological space that is homeomorphic to a metric space. That is, a topological space (X, \mathcal) is said to be metrizable if there is a metric d : X \times X \to , \inf ...
by the
Lévy–Prokhorov metric.
* A natural link to convergence in distribution is the
Skorokhod's representation theorem.
Convergence in probability
The basic idea behind this type of convergence is that the probability of an “unusual” outcome becomes smaller and smaller as the sequence progresses.
The concept of convergence in probability is used very often in statistics. For example, an estimator is called
consistent
In classical deductive logic, a consistent theory is one that does not lead to a logical contradiction. The lack of contradiction can be defined in either semantic or syntactic terms. The semantic definition states that a theory is consisten ...
if it converges in probability to the quantity being estimated. Convergence in probability is also the type of convergence established by the
weak law of large numbers
In probability theory, the law of large numbers (LLN) is a theorem that describes the result of performing the same experiment a large number of times. According to the law, the average of the results obtained from a large number of trials shou ...
.
Definition
A sequence of random variables converges in probability towards the random variable ''X'' if for all ''ε'' > 0
:
More explicitly, let ''P''
''n''(''ε'') be the probability that ''X''
''n'' is outside the ball of radius ''ε'' centered at ''X''. Then is said to converge in probability to ''X'' if for any and any ''δ'' > 0 there exists a number ''N'' (which may depend on ''ε'' and ''δ'') such that for all ''n'' ≥ ''N'', ''P''
''n''(''ε'') < ''δ'' (the definition of limit).
Notice that for the condition to be satisfied, it is not possible that for each ''n'' the random variables ''X'' and ''X''
''n'' are independent (and thus convergence in probability is a condition on the joint cdf's, as opposed to convergence in distribution, which is a condition on the individual cdf's), unless ''X'' is deterministic like for the weak law of large numbers. At the same time, the case of a deterministic ''X'' cannot, whenever the deterministic value is a discontinuity point (not isolated), be handled by convergence in distribution, where discontinuity points have to be explicitly excluded.
Convergence in probability is denoted by adding the letter ''p'' over an arrow indicating convergence, or using the "plim" probability limit operator:
For random elements on a
separable metric space , convergence in probability is defined similarly by
:
Properties
* Convergence in probability implies convergence in distribution.
">roof/sup>
* In the opposite direction, convergence in distribution implies convergence in probability when the limiting random variable ''X'' is a constant. ">roof/sup>
* Convergence in probability does not imply almost sure convergence. ">roof/sup>
* The continuous mapping theorem
In probability theory, the continuous mapping theorem states that continuous functions preserve limits even if their arguments are sequences of random variables. A continuous function, in Heine’s definition, is such a function that maps converg ...
states that for every continuous function , if , then also
* Convergence in probability defines a topology
In mathematics, topology (from the Greek words , and ) is concerned with the properties of a geometric object that are preserved under continuous deformations, such as stretching, twisting, crumpling, and bending; that is, without closing ho ...
on the space of random variables over a fixed probability space. This topology is metrizable
In topology and related areas of mathematics, a metrizable space is a topological space that is homeomorphic to a metric space. That is, a topological space (X, \mathcal) is said to be metrizable if there is a metric d : X \times X \to , \inf ...
by the '' Ky Fan metric'': or alternately by this metric
Almost sure convergence
This is the type of stochastic convergence that is most similar to pointwise convergence
In mathematics, pointwise convergence is one of various senses in which a sequence of functions can converge to a particular function. It is weaker than uniform convergence, to which it is often compared.
Definition
Suppose that X is a set and ...
known from elementary real analysis
In mathematics, the branch of real analysis studies the behavior of real numbers, sequences and series of real numbers, and real functions. Some particular properties of real-valued sequences and functions that real analysis studies include con ...
.
Definition
To say that the sequence converges almost surely or almost everywhere or with probability 1 or strongly towards ''X'' means that
:
This means that the values of approach the value of ''X'', in the sense (see almost surely
In probability theory, an event is said to happen almost surely (sometimes abbreviated as a.s.) if it happens with probability 1 (or Lebesgue measure 1). In other words, the set of possible exceptions may be non-empty, but it has probability 0 ...
) that events for which does not converge to ''X'' have probability 0. Using the probability space and the concept of the random variable as a function from Ω to R, this is equivalent to the statement
:
Using the notion of the limit superior of a sequence of sets, almost sure convergence can also be defined as follows:
:
Almost sure convergence is often denoted by adding the letters ''a.s.'' over an arrow indicating convergence:
For generic random element In probability theory, random element is a generalization of the concept of random variable to more complicated spaces than the simple real line. The concept was introduced by who commented that the “development of probability theory and expansio ...
s on a metric space
In mathematics, a metric space is a set together with a notion of '' distance'' between its elements, usually called points. The distance is measured by a function called a metric or distance function. Metric spaces are the most general sett ...
, convergence almost surely is defined similarly:
:
Properties
* Almost sure convergence implies convergence in probability (by Fatou's lemma), and hence implies convergence in distribution. It is the notion of convergence used in the strong law of large numbers
In probability theory, the law of large numbers (LLN) is a theorem that describes the result of performing the same experiment a large number of times. According to the law, the average of the results obtained from a large number of trials sho ...
.
* The concept of almost sure convergence does not come from a topology
In mathematics, topology (from the Greek words , and ) is concerned with the properties of a geometric object that are preserved under continuous deformations, such as stretching, twisting, crumpling, and bending; that is, without closing ho ...
on the space of random variables. This means there is no topology on the space of random variables such that the almost surely convergent sequences are exactly the converging sequences with respect to that topology. In particular, there is no metric of almost sure convergence.
Sure convergence or pointwise convergence
To say that the sequence of random variables (''X''''n'') defined over the same probability space
In probability theory, a probability space or a probability triple (\Omega, \mathcal, P) is a mathematical construct that provides a formal model of a random process or "experiment". For example, one can define a probability space which models t ...
(i.e., a random process) converges surely or everywhere or pointwise towards ''X'' means
where Ω is the sample space
In probability theory, the sample space (also called sample description space, possibility space, or outcome space) of an experiment or random trial is the set of all possible outcomes or results of that experiment. A sample space is usually de ...
of the underlying probability space
In probability theory, a probability space or a probability triple (\Omega, \mathcal, P) is a mathematical construct that provides a formal model of a random process or "experiment". For example, one can define a probability space which models t ...
over which the random variables are defined.
This is the notion of pointwise convergence
In mathematics, pointwise convergence is one of various senses in which a sequence of functions can converge to a particular function. It is weaker than uniform convergence, to which it is often compared.
Definition
Suppose that X is a set and ...
of a sequence of functions extended to a sequence of random variables. (Note that random variables themselves are functions).
Sure convergence of a random variable implies all the other kinds of convergence stated above, but there is no payoff in probability theory
Probability theory is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expressing it through a set o ...
by using sure convergence compared to using almost sure convergence. The difference between the two only exists on sets with probability zero. This is why the concept of sure convergence of random variables is very rarely used.
Convergence in mean
Given a real number , we say that the sequence converges in the ''r''-th mean (or in the ''Lr''-norm) towards the random variable ''X'', if the -th absolute moments E(, ''Xn'', ''r '') and E(, ''X'', ''r '') of and ''X'' exist, and
:
where the operator E denotes the expected value
In probability theory, the expected value (also called expectation, expectancy, mathematical expectation, mean, average, or first moment) is a generalization of the weighted average. Informally, the expected value is the arithmetic mean of a ...
. Convergence in -th mean tells us that the expectation of the -th power of the difference between and converges to zero.
This type of convergence is often denoted by adding the letter ''Lr'' over an arrow indicating convergence:
The most important cases of convergence in ''r''-th mean are:
* When converges in ''r''-th mean to ''X'' for ''r'' = 1, we say that converges in mean to ''X''.
* When converges in ''r''-th mean to ''X'' for ''r'' = 2, we say that converges in mean square (or in quadratic mean) to ''X''.
Convergence in the ''r''-th mean, for ''r'' ≥ 1, implies convergence in probability (by Markov's inequality
In probability theory, Markov's inequality gives an upper bound for the probability that a non-negative function of a random variable is greater than or equal to some positive constant. It is named after the Russian mathematician Andrey Marko ...
). Furthermore, if ''r'' > ''s'' ≥ 1, convergence in ''r''-th mean implies convergence in ''s''-th mean. Hence, convergence in mean square implies convergence in mean.
It is also worth noticing that if
then
:
Properties
Provided the probability space is complete
Complete may refer to:
Logic
* Completeness (logic)
* Completeness of a theory, the property of a theory that every formula in the theory's language or its negation is provable
Mathematics
* The completeness of the real numbers, which implies ...
:
* If and , then almost surely
In probability theory, an event is said to happen almost surely (sometimes abbreviated as a.s.) if it happens with probability 1 (or Lebesgue measure 1). In other words, the set of possible exceptions may be non-empty, but it has probability 0 ...
.
* If and , then almost surely.
* If and , then almost surely.
* If and , then (for any real numbers and ) and .
* If and , then (for any real numbers and ) and .
* If and , then (for any real numbers and ).
* None of the above statements are true for convergence in distribution.
The chain of implications between the various notions of convergence are noted in their respective sections. They are, using the arrow notation:
:
These properties, together with a number of other special cases, are summarized in the following list:
* Almost sure convergence implies convergence in probability: ">roof/sup>
*:
* Convergence in probability implies there exists a sub-sequence which almost surely converges:
*:
* Convergence in probability implies convergence in distribution: ">roof/sup>
*:
* Convergence in ''r''-th order mean implies convergence in probability:
*:
* Convergence in ''r''-th order mean implies convergence in lower order mean, assuming that both orders are greater than or equal to one:
*: provided ''r'' ≥ ''s'' ≥ 1.
* If ''X''''n'' converges in distribution to a constant ''c'', then ''X''''n'' converges in probability to ''c'': ">roof/sup>
*: provided ''c'' is a constant.
* If converges in distribution to ''X'' and the difference between ''Xn'' and ''Yn'' converges in probability to zero, then ''Yn'' also converges in distribution to ''X'': ">roof/sup>
*:
* If converges in distribution to ''X'' and ''Yn'' converges in distribution to a constant ''c'', then the joint vector converges in distribution to : ">roof/sup>
*: provided ''c'' is a constant.
*:Note that the condition that converges to a constant is important, if it were to converge to a random variable ''Y'' then we wouldn't be able to conclude that converges to .
* If ''Xn'' converges in probability to ''X'' and ''Yn'' converges in probability to ''Y'', then the joint vector converges in probability to : ">roof/sup>
*:
* If converges in probability to ''X'', and if for all ''n'' and some ''b'', then converges in ''r''th mean to ''X'' for all . In other words, if converges in probability to ''X'' and all random variables are almost surely bounded above and below, then converges to ''X'' also in any ''r''th mean.
* Almost sure representation. Usually, convergence in distribution does not imply convergence almost surely. However, for a given sequence which converges in distribution to ''X''0 it is always possible to find a new probability space (Ω, ''F'', P) and random variables defined on it such that ''Yn'' is equal in distribution to for each , and ''Yn'' converges to ''Y''0 almost surely.
* If for all ''ε'' > 0,
*::
*:then we say that ''converges almost completely'', or ''almost in probability'' towards ''X''. When converges almost completely towards ''X'' then it also converges almost surely to ''X''. In other words, if converges in probability to ''X'' sufficiently quickly (i.e. the above sequence of tail probabilities is summable for all ), then also converges almost surely to ''X''. This is a direct implication from the Borel–Cantelli lemma
In probability theory, the Borel–Cantelli lemma is a theorem about sequences of events. In general, it is a result in measure theory. It is named after Émile Borel and Francesco Paolo Cantelli, who gave statement to the lemma in the first ...
.
* If is a sum of ''n'' real independent random variables:
*::
*:then converges almost surely if and only if converges in probability.
* The dominated convergence theorem
In measure theory, Lebesgue's dominated convergence theorem provides sufficient conditions under which almost everywhere convergence of a sequence of functions implies convergence in the ''L''1 norm. Its power and utility are two of the primary ...
gives sufficient conditions for almost sure convergence to imply ''L''1-convergence:
}
*A necessary and sufficient condition for ''L''1 convergence is and the sequence (''Xn'') is uniformly integrable
In mathematics, uniform integrability is an important concept in real analysis, functional analysis and measure theory, and plays a vital role in the theory of martingales.
Measure-theoretic definition
Uniform integrability is an extension to th ...
.
*If are discrete and independent, then implies that . This is a consequence of the second Borel–Cantelli lemma.
See also
* Proofs of convergence of random variables
This article is supplemental for “Convergence of random variables” and provides proofs for selected results.
Several results will be established using the portmanteau lemma: A sequence converges in distribution to ''X'' if and only if any ...
* Convergence of measures
* Convergence in measure
Convergence in measure is either of two distinct mathematical concepts both of which generalize
the concept of convergence in probability.
Definitions
Let f, f_n\ (n \in \mathbb N): X \to \mathbb R be measurable functions on a measure space (X ...
* Continuous stochastic process: the question of continuity of a stochastic process is essentially a question of convergence, and many of the same concepts and relationships used above apply to the continuity question.
* Asymptotic distribution
In mathematics and statistics, an asymptotic distribution is a probability distribution that is in a sense the "limiting" distribution of a sequence of distributions. One of the main uses of the idea of an asymptotic distribution is in providing ...
* Big O in probability notation The order in probability notation is used in probability theory and statistical theory in direct parallel to the big-O notation that is standard in mathematics. Where the big-O notation deals with the convergence of sequences or sets of ordinary n ...
* Skorokhod's representation theorem
* The Tweedie convergence theorem
* Slutsky's theorem In probability theory, Slutsky’s theorem extends some properties of algebraic operations on convergent sequences of real numbers to sequences of random variables.
The theorem was named after Eugen Slutsky. Slutsky's theorem is also attributed ...
* Continuous mapping theorem
In probability theory, the continuous mapping theorem states that continuous functions preserve limits even if their arguments are sequences of random variables. A continuous function, in Heine’s definition, is such a function that maps converg ...
Notes
References
*
*
*
*
*
*
*
*
*
*
*
*
*
*
* https://www.ma.utexas.edu/users/gordanz/notes/weak.pdf
{{DEFAULTSORT:Convergence Of Random Variables
Stochastic processes
Random variables, Convergence of>X_n, ^r= E X, ^r
Properties
Provided the probability space is complete
Complete may refer to:
Logic
* Completeness (logic)
* Completeness of a theory, the property of a theory that every formula in the theory's language or its negation is provable
Mathematics
* The completeness of the real numbers, which implies ...
:
* If and , then almost surely
In probability theory, an event is said to happen almost surely (sometimes abbreviated as a.s.) if it happens with probability 1 (or Lebesgue measure 1). In other words, the set of possible exceptions may be non-empty, but it has probability 0 ...
.
* If and , then almost surely.
* If and , then almost surely.
* If and , then (for any real numbers and ) and .
* If and , then (for any real numbers and ) and .
* If and , then (for any real numbers and ).
* None of the above statements are true for convergence in distribution.
The chain of implications between the various notions of convergence are noted in their respective sections. They are, using the arrow notation:
:
These properties, together with a number of other special cases, are summarized in the following list:
* Almost sure convergence implies convergence in probability: ">roof/sup>
*:
* Convergence in probability implies there exists a sub-sequence which almost surely converges:
*:
* Convergence in probability implies convergence in distribution: ">roof/sup>
*:
* Convergence in ''r''-th order mean implies convergence in probability:
*:
* Convergence in ''r''-th order mean implies convergence in lower order mean, assuming that both orders are greater than or equal to one:
*: provided ''r'' ≥ ''s'' ≥ 1.
* If ''X''''n'' converges in distribution to a constant ''c'', then ''X''''n'' converges in probability to ''c'': ">roof/sup>
*: provided ''c'' is a constant.
* If converges in distribution to ''X'' and the difference between ''Xn'' and ''Yn'' converges in probability to zero, then ''Yn'' also converges in distribution to ''X'': ">roof/sup>
*:
* If converges in distribution to ''X'' and ''Yn'' converges in distribution to a constant ''c'', then the joint vector converges in distribution to : ">roof/sup>
*: provided ''c'' is a constant.
*:Note that the condition that converges to a constant is important, if it were to converge to a random variable ''Y'' then we wouldn't be able to conclude that converges to .
* If ''Xn'' converges in probability to ''X'' and ''Yn'' converges in probability to ''Y'', then the joint vector converges in probability to : ">roof/sup>
*:
* If converges in probability to ''X'', and if for all ''n'' and some ''b'', then converges in ''r''th mean to ''X'' for all . In other words, if converges in probability to ''X'' and all random variables are almost surely bounded above and below, then converges to ''X'' also in any ''r''th mean.
* Almost sure representation. Usually, convergence in distribution does not imply convergence almost surely. However, for a given sequence which converges in distribution to ''X''0 it is always possible to find a new probability space (Ω, ''F'', P) and random variables defined on it such that ''Yn'' is equal in distribution to for each , and ''Yn'' converges to ''Y''0 almost surely.
* If for all ''ε'' > 0,
*::
*:then we say that ''converges almost completely'', or ''almost in probability'' towards ''X''. When converges almost completely towards ''X'' then it also converges almost surely to ''X''. In other words, if converges in probability to ''X'' sufficiently quickly (i.e. the above sequence of tail probabilities is summable for all ), then also converges almost surely to ''X''. This is a direct implication from the Borel–Cantelli lemma
In probability theory, the Borel–Cantelli lemma is a theorem about sequences of events. In general, it is a result in measure theory. It is named after Émile Borel and Francesco Paolo Cantelli, who gave statement to the lemma in the first ...
.
* If is a sum of ''n'' real independent random variables:
*::
*:then converges almost surely if and only if converges in probability.
* The dominated convergence theorem
In measure theory, Lebesgue's dominated convergence theorem provides sufficient conditions under which almost everywhere convergence of a sequence of functions implies convergence in the ''L''1 norm. Its power and utility are two of the primary ...
gives sufficient conditions for almost sure convergence to imply ''L''1-convergence:
}
*A necessary and sufficient condition for ''L''1 convergence is and the sequence (''Xn'') is uniformly integrable
In mathematics, uniform integrability is an important concept in real analysis, functional analysis and measure theory, and plays a vital role in the theory of martingales.
Measure-theoretic definition
Uniform integrability is an extension to th ...
.
*If are discrete and independent, then implies that . This is a consequence of the second Borel–Cantelli lemma.
See also
* Proofs of convergence of random variables
This article is supplemental for “Convergence of random variables” and provides proofs for selected results.
Several results will be established using the portmanteau lemma: A sequence converges in distribution to ''X'' if and only if any ...
* Convergence of measures
* Convergence in measure
Convergence in measure is either of two distinct mathematical concepts both of which generalize
the concept of convergence in probability.
Definitions
Let f, f_n\ (n \in \mathbb N): X \to \mathbb R be measurable functions on a measure space (X ...
* Continuous stochastic process: the question of continuity of a stochastic process is essentially a question of convergence, and many of the same concepts and relationships used above apply to the continuity question.
* Asymptotic distribution
In mathematics and statistics, an asymptotic distribution is a probability distribution that is in a sense the "limiting" distribution of a sequence of distributions. One of the main uses of the idea of an asymptotic distribution is in providing ...
* Big O in probability notation The order in probability notation is used in probability theory and statistical theory in direct parallel to the big-O notation that is standard in mathematics. Where the big-O notation deals with the convergence of sequences or sets of ordinary n ...
* Skorokhod's representation theorem
* The Tweedie convergence theorem
* Slutsky's theorem In probability theory, Slutsky’s theorem extends some properties of algebraic operations on convergent sequences of real numbers to sequences of random variables.
The theorem was named after Eugen Slutsky. Slutsky's theorem is also attributed ...
* Continuous mapping theorem
In probability theory, the continuous mapping theorem states that continuous functions preserve limits even if their arguments are sequences of random variables. A continuous function, in Heine’s definition, is such a function that maps converg ...
Notes
References
*
*
*
*
*
*
*
*
*
*
*
*
*
*
* https://www.ma.utexas.edu/users/gordanz/notes/weak.pdf
{{DEFAULTSORT:Convergence Of Random Variables
Stochastic processes
Random variables, Convergence of