Modes Of Variation
   HOME

TheInfoList



OR:

In statistics, modes of variation are a continuously indexed set of vectors or functions that are centered at a mean and are used to depict the variation in a population or sample. Typically, variation patterns in the data can be decomposed in descending order of
eigenvalues In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted b ...
with the directions represented by the corresponding eigenvectors or eigenfunctions. Modes of variation provide a visualization of this decomposition and an efficient description of variation around the mean. Both in
principal component analysis Principal component analysis (PCA) is a popular technique for analyzing large datasets containing a high number of dimensions/features per observation, increasing the interpretability of data while preserving the maximum amount of information, and ...
(PCA) and in
functional principal component analysis Functional principal component analysis (FPCA) is a statistical method for investigating the dominant modes of variation of functional data. Using this method, a random function is represented in the eigenbasis, which is an orthonormal basis of ...
(FPCA), modes of variation play an important role in visualizing and describing the variation in the data contributed by each eigencomponent. In real-world applications, the eigencomponents and associated modes of variation aid to interpret complex data, especially in exploratory data analysis (EDA).


Formulation

Modes of variation are a natural extension of PCA and FPCA.


Modes of variation in PCA

If a random vector \mathbf=(X_1, X_2, \cdots, X_p)^T has the mean vector \boldsymbol_p, and the covariance matrix \mathbf_ with
eigenvalues In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted b ...
\lambda_1\geq \lambda_2\geq \cdots \geq \lambda_p\geq0 and corresponding
orthonormal In linear algebra, two vectors in an inner product space are orthonormal if they are orthogonal (or perpendicular along a line) unit vectors. A set of vectors form an orthonormal set if all vectors in the set are mutually orthogonal and all of un ...
eigenvectors \mathbf_1, \mathbf_2, \cdots,\mathbf_p, by eigendecomposition of a real symmetric matrix, the covariance matrix \mathbf can be decomposed as :\mathbf=\mathbf\mathbf\mathbf^T, where \mathbf is an orthogonal matrix whose columns are the eigenvectors of \mathbf, and \mathbf is a diagonal matrix whose entries are the eigenvalues of \mathbf. By the Karhunen–Loève expansion for random vectors, one can express the centered random vector in the eigenbasis :\mathbf-\boldsymbol=\sum_^p\xi_k\mathbf_k, where \xi_k=\mathbf_k^T(\mathbf-\boldsymbol) is the principal component associated with the k-th eigenvector \mathbf_k, with the properties :\operatorname(\xi_k)=0, \operatorname(\xi_k)=\lambda_k, and \operatorname(\xi_k\xi_l)=0\ \text\ l\neq k. Then the k-th mode of variation of \mathbf is the set of vectors, indexed by \alpha, :\mathbf_=\boldsymbol\pm \alpha\sqrt\mathbf_k, \alpha\in A, A where A is typically selected as 2\ \text\ 3.


Modes of variation in FPCA

For a square-integrable
random function In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a Indexed family, family of random variables. Stochastic processes are widely used as mathematical models of systems and phen ...
X(t), t \in \mathcal\subset R^p, where typically p=1 and \mathcal is an interval, denote the mean function by \mu(t) = \operatorname(X(t)) , and the covariance function by : G(s, t) = \operatorname(X(s), X(t)) = \sum_^\infty \lambda_k \varphi_k(s) \varphi_k(t), where \lambda_1\geq \lambda_2\geq \cdots \geq 0 are the eigenvalues and \ are the
orthonormal In linear algebra, two vectors in an inner product space are orthonormal if they are orthogonal (or perpendicular along a line) unit vectors. A set of vectors form an orthonormal set if all vectors in the set are mutually orthogonal and all of un ...
eigenfunctions of the linear
Hilbert–Schmidt operator In mathematics, a Hilbert–Schmidt operator, named after David Hilbert and Erhard Schmidt, is a bounded operator A \colon H \to H that acts on a Hilbert space H and has finite Hilbert–Schmidt norm \, A\, ^2_ \ \stackrel\ \sum_ \, Ae_i\, ^2_ ...
: G: L^2(\mathcal) \rightarrow L^2(\mathcal),\, G(f) = \int_\mathcal G(s, t) f(s) ds. By the Karhunen–Loève theorem, one can express the centered function in the eigenbasis, : X(t) - \mu(t) = \sum_^\infty \xi_k \varphi_k(t), where : \xi_k = \int_\mathcal (X(t) - \mu(t)) \varphi_k(t) dt is the k-th principal component with the properties : \operatorname(\xi_k) = 0, \operatorname(\xi_k) = \lambda_k, and \operatorname(\xi_k \xi_l) = 0 \text l \ne k. Then the k-th mode of variation of X(t) is the set of functions, indexed by \alpha, :m_(t)=\mu(t)\pm \alpha\sqrt\varphi_k(t),\ t\in \mathcal,\ \alpha\in A, A/math> that are viewed simultaneously over the range of \alpha, usually for A=2\ \text\ 3.


Estimation

The formulation above is derived from properties of the population. Estimation is needed in real-world applications. The key idea is to estimate mean and covariance.


Modes of variation in PCA

Suppose the data \mathbf_1, \mathbf_2, \cdots, \mathbf_n represent n independent drawings from some p-dimensional population \mathbf with mean vector \boldsymbol and covariance matrix \mathbf. These data yield the sample mean vector \overline\mathbf, and the sample covariance matrix \mathbf with eigenvalue-eigenvector pairs (\hat_1, \hat_1), (\hat_2, \hat_2), \cdots, (\hat_p, \hat_p). Then the k-th mode of variation of \mathbf can be estimated by : \hat_=\overline\pm \alpha\sqrt\hat_k, \alpha\in A, A


Modes of variation in FPCA

Consider n realizations X_1(t), X_2(t), \cdots, X_n(t) of a square-integrable
random function In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a Indexed family, family of random variables. Stochastic processes are widely used as mathematical models of systems and phen ...
X(t), t \in \mathcal with the mean function \mu(t) = \operatorname(X(t)) and the covariance function G(s, t) = \operatorname(X(s), X(t)) .
Functional principal component analysis Functional principal component analysis (FPCA) is a statistical method for investigating the dominant modes of variation of functional data. Using this method, a random function is represented in the eigenbasis, which is an orthonormal basis of ...
provides methods for the estimation of \mu(t) and G(s, t) in detail, often involving point wise estimate and
interpolation In the mathematical field of numerical analysis, interpolation is a type of estimation, a method of constructing (finding) new data points based on the range of a discrete set of known data points. In engineering and science, one often has a n ...
. Substituting estimates for the unknown quantities, the k-th mode of variation of X(t) can be estimated by : \hat_(t)=\hat(t)\pm \alpha\sqrt\hat_k(t), t\in \mathcal, \alpha\in A, A


Applications

Modes of variation are useful to visualize and describe the variation patterns in the data sorted by the eigenvalues. In real-world applications, modes of variation associated with eigencomponents allow to interpret complex data, such as the evolution of function traits and other infinite-dimensional data. To illustrate how modes of variation work in practice, two examples are shown in the graphs to the right, which display the first two modes of variation. The solid curve represents the sample mean function. The dashed, dot-dashed, and dotted curves correspond to modes of variation with \alpha=\pm1, \pm2, and \pm3, respectively. The first graph displays the first two modes of variation of female mortality data from 41 countries in 2003. The object of interest is log
hazard function Failure rate is the frequency with which an engineered system or component fails, expressed in failures per unit of time. It is usually denoted by the Greek letter λ (lambda) and is often used in reliability engineering. The failure rate of a ...
between ages 0 and 100 years. The first mode of variation suggests that the variation of female mortality is smaller for ages around 0 or 100, and larger for ages around 25. An appropriate and intuitive interpretation is that mortality around 25 is driven by accidental death, while around 0 or 100, mortality is related to congenital disease or natural death. Compared to female mortality data, modes of variation of male mortality data shows higher mortality after around age 20, possibly related to the fact that life expectancy for women is higher than that for men.


References

{{reflist Dimension reduction Functional analysis Matrix decompositions