In
probability theory
Probability theory or probability calculus is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expre ...
, Slutsky's theorem extends some properties of algebraic operations on
convergent sequences of
real number
In mathematics, a real number is a number that can be used to measure a continuous one- dimensional quantity such as a duration or temperature. Here, ''continuous'' means that pairs of values can have arbitrarily small differences. Every re ...
s to sequences of
random variable
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a Mathematics, mathematical formalization of a quantity or object which depends on randomness, random events. The term 'random variable' in its mathema ...
s.
The theorem was named after
Eugen Slutsky. Slutsky's theorem is also attributed to
Harald Cramér
Harald Cramér (; 25 September 1893 – 5 October 1985) was a Swedish mathematician, actuary, and statistician, specializing in mathematical statistics and probabilistic number theory. John Kingman described him as "one of the giants of statis ...
.
Statement
Let
be sequences of scalar/vector/matrix
random element In probability theory, random element is a generalization of the concept of random variable to more complicated spaces than the simple real line. The concept was introduced by who commented that the “development of probability theory and expansio ...
s.
If
converges in distribution to a random element
and
converges in probability to a constant
, then
*
*
*
provided that ''c'' is invertible,
where
denotes
convergence in distribution
In probability theory, there exist several different notions of convergence of sequences of random variables, including ''convergence in probability'', ''convergence in distribution'', and ''almost sure convergence''. The different notions of conve ...
.
Notes:
# The requirement that ''Y
n'' converges to a constant is important — if it were to converge to a non-degenerate random variable, the theorem would be no longer valid. For example, let
and
. The sum
for all values of ''n''. Moreover,
, but
does not converge in distribution to
, where
,
, and
and
are independent.
[See ]
# The theorem remains valid if we replace all convergences in distribution with convergences in probability.
Proof
This theorem follows from the fact that if ''X''
''n'' converges in distribution to ''X'' and ''Y''
''n'' converges in probability to a constant ''c'', then the joint vector (''X''
''n'', ''Y''
''n'') converges in distribution to (''X'', ''c'') (
see here).
Next we apply the
continuous mapping theorem
In probability theory, the continuous mapping theorem states that continuous functions preserve limits even if their arguments are sequences of random variables. A continuous function, in Heine's definition, is such a function that maps converge ...
, recognizing the functions ''g''(''x'',''y'') = ''x'' + ''y'', ''g''(''x'',''y'') = ''xy'', and ''g''(''x'',''y'') = ''x'' ''y''
−1 are continuous (for the last function to be continuous, ''y'' has to be invertible).
See also
*
Convergence of random variables
In probability theory, there exist several different notions of convergence of sequences of random variables, including ''convergence in probability'', ''convergence in distribution'', and ''almost sure convergence''. The different notions of conve ...
References
Further reading
*
*
*
{{DEFAULTSORT:Slutsky's Theorem
Asymptotic theory (statistics)
Theorems in probability theory
Theorems in statistics