HOME

TheInfoList



OR:

In
multilinear algebra Multilinear algebra is a subfield of mathematics that extends the methods of linear algebra. Just as linear algebra is built on the concept of a vector and develops the theory of vector spaces, multilinear algebra builds on the concepts of ''p' ...
, the tensor rank decomposition or the rank-R decomposition of a tensor is the decomposition of a tensor in terms of a sum of minimum R rank-1 tensors. This is an open problem. Canonical polyadic decomposition (CPD) is a variant of the rank decomposition which computes the best fitting K rank-1 terms for a user specified K. The CP decomposition has found some applications in
linguistics Linguistics is the scientific study of human language. It is called a scientific study because it entails a comprehensive, systematic, objective, and precise analysis of all aspects of language, particularly its nature and structure. Linguis ...
and
chemometrics Chemometrics is the science of extracting information from chemical systems by data-driven means. Chemometrics is inherently interdisciplinary, using methods frequently employed in core data-analytic disciplines such as multivariate statistics, a ...
. The CP rank was introduced by
Frank Lauren Hitchcock Frank Lauren Hitchcock (March 6, 1875 – May 31, 1957) was an American mathematician and physicist known for his formulation of the transportation problem in 1941. Academic life Frank did his preparatory study at Phillips Andover Academy. He en ...
in 1927 and later rediscovered several times, notably in psychometrics. The CP decomposition is referred to as CANDECOMP, PARAFAC, or CANDECOMP/PARAFAC (CP). Another popular generalization of the matrix SVD known as the
higher-order singular value decomposition In multilinear algebra, the higher-order singular value decomposition (HOSVD) of a tensor is a specific orthogonal Tucker decomposition. It may be regarded as one generalization of the matrix singular value decomposition. It has applications in co ...
computes orthonormal mode matrices and has found applications in
econometrics Econometrics is the application of Statistics, statistical methods to economic data in order to give Empirical evidence, empirical content to economic relationships.M. Hashem Pesaran (1987). "Econometrics," ''The New Palgrave: A Dictionary of ...
,
signal processing Signal processing is an electrical engineering subfield that focuses on analyzing, modifying and synthesizing ''signals'', such as audio signal processing, sound, image processing, images, and scientific measurements. Signal processing techniq ...
,
computer vision Computer vision is an interdisciplinary scientific field that deals with how computers can gain high-level understanding from digital images or videos. From the perspective of engineering, it seeks to understand and automate tasks that the hum ...
,
computer graphics Computer graphics deals with generating images with the aid of computers. Today, computer graphics is a core technology in digital photography, film, video games, cell phone and computer displays, and many specialized applications. A great de ...
,
psychometrics Psychometrics is a field of study within psychology concerned with the theory and technique of measurement. Psychometrics generally refers to specialized fields within psychology and education devoted to testing, measurement, assessment, and ...
.


Notation

A scalar variable is denoted by lower case italic letters, a and an upper bound scalar is denoted by an upper case italic letter, A . Indices are denoted by a combination of lowercase and upper case italic letters, 1 \le i \le I. Multiple indices that one might encounter when referring to the multiple modes of a tensor are conveniently denoted by 1\le i_m \le I_m where 1\le m \le M. A vector is denoted by a lower case bold Times Roman, \mathbf a and a matrix is denoted by bold upper case letters \mathbf A. A higher order tensor is denoted by calligraphic letters,\mathcal A. An element of an M-order tensor \mathcal A \in \mathbb C^ is denoted by a_ or \mathcal A_.


Definition

A data tensor \in ^ is a collection of multivariate observations organized into a -way array where =+1. Every tensor may be represented with a suitably large R as a linear combination of r rank-1 tensors: : \mathcal = \sum_^ \lambda_r \mathbf_ \otimes\mathbf_ \otimes \mathbf_ \dots \otimes \mathbf_\otimes \cdots \otimes \mathbf_, where \lambda_r \in and \mathbf_ \in ^ where 1\le m\le M. When the number of terms R is minimal in the above expression, then R is called the rank of the tensor, and the decomposition is often referred to as a ''(tensor) rank decomposition'', ''minimal CP decomposition'', or ''Canonical Polyadic Decomposition (CPD)''. If the number of terms is not minimal, then the above decomposition is often referred to as ''CANDECOMP/PARAFAC'', ''Polyadic decomposition'.


Tensor rank

Contrary to the case of matrices, computing the rank of a tensor is
NP-hard In computational complexity theory, NP-hardness ( non-deterministic polynomial-time hardness) is the defining property of a class of problems that are informally "at least as hard as the hardest problems in NP". A simple example of an NP-hard pr ...
. The only notable well-understood case consists of tensors in F^ \otimes F^ \otimes F^2, whose rank can be obtained from the Kronecker
Weierstrass Karl Theodor Wilhelm Weierstrass (german: link=no, Weierstraß ; 31 October 1815 – 19 February 1897) was a German mathematician often cited as the "father of modern mathematical analysis, analysis". Despite leaving university without a degree, ...
normal form of the linear
matrix pencil In linear algebra, if A_0, A_1,\dots,A_\ell are n\times n complex matrices for some nonnegative integer \ell, and A_\ell \ne 0 (the zero matrix), then the matrix pencil of degree \ell is the matrix-valued function defined on the complex numbers L(\ ...
that the tensor represents. A simple polynomial-time algorithm exists for certifying that a tensor is of rank 1, namely the
higher-order singular value decomposition In multilinear algebra, the higher-order singular value decomposition (HOSVD) of a tensor is a specific orthogonal Tucker decomposition. It may be regarded as one generalization of the matrix singular value decomposition. It has applications in co ...
. The rank of the tensor of zeros is zero by convention. The rank of a tensor \mathbf_1 \otimes \cdots \otimes \mathbf_M is one, provided that \mathbf_m \in F^\setminus\ .


Field dependence

The rank of a tensor depends on the field over which the tensor is decomposed. It is known that some real tensors may admit a complex decomposition whose rank is strictly less than the rank of a real decomposition of the same tensor. As an example, consider the following real tensor : \mathcal = \mathbf_1 \otimes \mathbf_2 \otimes \mathbf_3 + \mathbf_1 \otimes \mathbf_2 \otimes \mathbf_3 - \mathbf_1 \otimes \mathbf_2 \otimes \mathbf_3 + \mathbf_1 \otimes \mathbf_2 \otimes \mathbf_3, where \mathbf_i, \mathbf_j \in \mathbb^2. The rank of this tensor over the reals is known to be 3, while its complex rank is only 2 because it is the sum of a complex rank-1 tensor with its
complex conjugate In mathematics, the complex conjugate of a complex number is the number with an equal real part and an imaginary part equal in magnitude but opposite in sign. That is, (if a and b are real, then) the complex conjugate of a + bi is equal to a - ...
, namely : \mathcal = \frac( \bar_1 \otimes \mathbf_2 \otimes \bar_3 + \mathbf_1 \otimes \bar_2 \otimes \mathbf_3), where \mathbf_k=\mathbf_k + i\mathbf_k. In contrast, the rank of real matrices will never decrease under a
field extension In mathematics, particularly in algebra, a field extension is a pair of fields E\subseteq F, such that the operations of ''E'' are those of ''F'' restricted to ''E''. In this case, ''F'' is an extension field of ''E'' and ''E'' is a subfield of ...
to \mathbb: real matrix rank and complex matrix rank coincide for real matrices.


Generic rank

The generic rank r(I_1,\ldots,I_M) is defined as the least rank r such that the closure in the
Zariski topology In algebraic geometry and commutative algebra, the Zariski topology is a topology which is primarily defined by its closed sets. It is very different from topologies which are commonly used in the real or complex analysis; in particular, it is n ...
of the set of tensors of rank at most r is the entire space F^ \otimes \cdots \otimes F^. In the case of complex tensors, tensors of rank at most r(I_1,\ldots,I_M) form a
dense set In topology and related areas of mathematics, a subset ''A'' of a topological space ''X'' is said to be dense in ''X'' if every point of ''X'' either belongs to ''A'' or else is arbitrarily "close" to a member of ''A'' — for instance, the r ...
S: every tensor in the aforementioned space is either of rank less than the generic rank, or it is the limit in the
Euclidean topology In mathematics, and especially general topology, the Euclidean topology is the natural topology induced on n-dimensional Euclidean space \R^n by the Euclidean distance, Euclidean metric. Definition The Euclidean norm on \R^n is the non-negative f ...
of a sequence of tensors from S. In the case of real tensors, the set of tensors of rank at most r(I_1,\ldots,I_M) only forms an open set of positive measure in the Euclidean topology. There may exist Euclidean-open sets of tensors of rank strictly higher than the generic rank. All ranks appearing on open sets in the Euclidean topology are called typical ranks. The smallest typical rank is called the generic rank; this definition applies to both complex and real tensors. The generic rank of tensor spaces was initially studied in 1983 by
Volker Strassen Volker Strassen (born April 29, 1936) is a German mathematician, a professor emeritus in the department of mathematics and statistics at the University of Konstanz. For important contributions to the analysis of algorithms he has received many aw ...
. As an illustration of the above concepts, it is known that both 2 and 3 are typical ranks of \mathbb^2 \otimes \mathbb^2 \otimes \mathbb^2 while the generic rank of \mathbb^2 \otimes \mathbb^2 \otimes \mathbb^2 is 2. Practically, this means that a randomly sampled real tensor (from a continuous probability measure on the space of tensors) of size 2 \times 2 \times 2 will be a rank-1 tensor with probability zero, a rank-2 tensor with positive probability, and rank-3 with positive probability. On the other hand, a randomly sampled complex tensor of the same size will be a rank-1 tensor with probability zero, a rank-2 tensor with probability one, and a rank-3 tensor with probability zero. It is even known that the generic rank-3 real tensor in \mathbb^2 \otimes \mathbb^2 \otimes \mathbb^2 will be of complex rank equal to 2. The generic rank of tensor spaces depends on the distinction between balanced and unbalanced tensor spaces. A tensor space F^ \otimes \cdots \otimes F^, where I_1 \ge I_2 \ge \cdots \ge I_M, is called unbalanced whenever : I_1 > 1 + \prod_^M I_m - \sum_^M (I_m-1), and it is called balanced otherwise.


Unbalanced tensor spaces

When the first factor is very large with respect to the other factors in the tensor product, then the tensor space essentially behaves as a matrix space. The generic rank of tensors living in an unbalanced tensor spaces is known to equal : r(I_1,\ldots,I_M) = \min\left\
almost everywhere In measure theory (a branch of mathematical analysis), a property holds almost everywhere if, in a technical sense, the set for which the property holds takes up nearly all possibilities. The notion of "almost everywhere" is a companion notion to ...
. More precisely, the rank of every tensor in an unbalanced tensor space F^ \setminus Z, where Z is some indeterminate closed set in the Zariski topology, equals the above value.


Balanced tensor spaces

The expected generic rank of tensors living in a balanced tensor space is equal to : r_E(I_1,\ldots,I_M) = \left\lceil \frac \right\rceil
almost everywhere In measure theory (a branch of mathematical analysis), a property holds almost everywhere if, in a technical sense, the set for which the property holds takes up nearly all possibilities. The notion of "almost everywhere" is a companion notion to ...
for complex tensors and on a Euclidean-open set for real tensors, where : \Pi = \prod_^ I_m \quad\text\quad \Sigma = \sum_^ (I_m - 1). More precisely, the rank of every tensor in \mathbb^ \setminus Z, where Z is some indeterminate closed set in the
Zariski topology In algebraic geometry and commutative algebra, the Zariski topology is a topology which is primarily defined by its closed sets. It is very different from topologies which are commonly used in the real or complex analysis; in particular, it is n ...
, is expected to equal the above value. For real tensors, r_E(I_1,\ldots,I_M) is the least rank that is expected to occur on a set of positive Euclidean measure. The value r_E(I_1,\ldots,I_M) is often referred to as the expected generic rank of the tensor space F^ because it is only conjecturally correct. It is known that the true generic rank always satisfies : r(I_1, \ldots, I_M) \ge r_E(I_1, \ldots, I_M). The Abo–Ottaviani–Peterson conjecture states that equality is expected, i.e., r(I_1,\ldots,I_M) = r_E(I_1,\ldots,I_M), with the following exceptional cases: * F^ \text m = 1, 2, \ldots * F^ \text m = 2,3, \ldots In each of these exceptional cases, the generic rank is known to be r(I_1,\ldots,I_m,\ldots,I_M) = r_E(I_1,\ldots,I_M)+1. Note that while the set of tensors of rank 3 in F^ is defective (13 and not the expected 14), the generic rank in that space is still the expected one, 4. Similarly, the set of tensors of rank 5 in F^ is defective (44 and not the expected 45), but the generic rank in that space is still the expected 6. The AOP conjecture has been proved completely in a number of special cases. Lickteig showed already in 1985 that r(n,n,n) = r_E(n,n,n), provided that n \ne 3. In 2011, a major breakthrough was established by Catalisano, Geramita, and Gimigliano who proved that the expected dimension of the set of rank s tensors of format 2\times 2\times \cdots \times 2 is the expected one except for rank 3 tensors in the 4 factor case, yet the expected rank in that case is still 4. As a consequence, r(2,2,\ldots,2) = r_E(2,2,\ldots,2) for all binary tensors.


Maximum rank

The maximum rank that can be admitted by any of the tensors in a tensor space is unknown in general; even a conjecture about this maximum rank is missing. Presently, the best general upper bound states that the maximum rank r_\mbox(I_1,\ldots,I_M) of F^ \otimes \cdots \otimes F^, where I_1 \ge I_2 \ge \cdots \ge I_M, satisfies : r_\mbox(I_1,\ldots,I_M) \le \min\left\, where r(I_1,\ldots,I_M) is the (least) ''generic rank'' of F^ \otimes \cdots \otimes F^. It is well-known that the foregoing inequality may be strict. For instance, the generic rank of tensors in \mathbb^ is two, so that the above bound yields r_\mbox(2,2,2) \le 4, while it is known that the maximum rank equals 3.


Border rank

A rank-s tensor \mathcal is called a border tensor if there exists a sequence of tensors of rank at most r < s whose limit is \mathcal. If r is the least value for which such a convergent sequence exists, then it is called the border rank of \mathcal. For order-2 tensors, i.e., matrices, rank and border rank ''always'' coincide, however, for tensors of order \ge3 they may differ. Border tensors were first studied in the context of fast ''approximate''
matrix multiplication algorithm Because matrix multiplication is such a central operation in many numerical algorithms, much work has been invested in making matrix multiplication algorithms efficient. Applications of matrix multiplication in computational problems are found in m ...
s by Bini, Lotti, and Romani in 1980. A classic example of a border tensor is the rank-3 tensor : \mathcal = \mathbf \otimes \mathbf \otimes \mathbf + \mathbf \otimes \mathbf \otimes \mathbf + \mathbf \otimes \mathbf \otimes \mathbf, \quad \text \, \mathbf\, = \, \mathbf\, = 1 \text \langle \mathbf, \mathbf\rangle \ne 1. It can be approximated arbitrarily well by the following sequence of rank-2 tensors : \begin \mathcal_m &= m (\mathbf + \frac \mathbf) \otimes (\mathbf + \frac \mathbf) \otimes (\mathbf + \frac \mathbf) - m \mathbf\otimes\mathbf\otimes\mathbf \\ &= \mathbf \otimes \mathbf \otimes \mathbf + \mathbf \otimes \mathbf \otimes \mathbf + \mathbf \otimes \mathbf \otimes \mathbf + \frac (\mathbf\otimes\mathbf\otimes\mathbf + \mathbf\otimes\mathbf\otimes\mathbf + \mathbf\otimes\mathbf\otimes\mathbf) + \frac \mathbf\otimes\mathbf\otimes\mathbf \end as m \to \infty. Therefore, its border rank is 2, which is strictly less than its rank. When the two vectors are orthogonal, this example is also known as a
W state The W state is an entangled quantum state of three qubits which in the bra-ket notation has the following shape : , \mathrm\rangle = \frac(, 001\rangle + , 010\rangle + , 100\rangle) and which is remarkable for representing a specific type of ...
.


Properties


Identifiability

It follows from the definition of a pure tensor that \mathcal = \mathbf_1 \otimes \mathbf_2 \otimes \cdots \otimes \mathbf_M = \mathbf_1 \otimes \mathbf_2 \otimes \cdots \otimes \mathbf_M if and only if there exist \lambda_k such that \lambda_1 \lambda_2 \cdots \lambda_M = 1 and \mathbf_m = \lambda_m \mathbf_m for all ''m''. For this reason, the parameters \_^M of a rank-1 tensor \mathcal are called identifiable or essentially unique. A rank-r tensor \mathcal \in F^ \otimes F^ \otimes \cdots \otimes F^ is called identifiable if every of its tensor rank decompositions is the sum of the same set of r distinct tensors \ where the \mathcal_i's are of rank 1. An identifiable rank-r thus has only one essentially unique decomposition \mathcal = \sum_^r \mathcal_i,and all r! tensor rank decompositions of \mathcal can be obtained by permuting the order of the summands. Observe that in a tensor rank decomposition all the \mathcal_i's are distinct, for otherwise the rank of \mathcal would be at most r-1.


Generic identifiability

Order-2 tensors in F^ \otimes F^ \simeq F^, i.e., matrices, are not identifiable for r > 1. This follows essentially from the observation \mathcal = \sum_^r \mathbf_i \otimes \mathbf_i = \sum_^r \mathbf_i \mathbf_i^T = A B^T = (A X^) (B X^T)^T = \sum_^r \mathbf_i \mathbf_i^T = \sum_^r \mathbf_i \otimes \mathbf_i,where X \in \mathrm_(F) is an invertible r \times r matrix, A = mathbf_i^r , B = mathbf_i^r, A X^ = mathbf_i^r and B X^T = mathbf_i^r. It can be shown that for every X \in \mathrm_n(F)\setminus Z, where Z is a closed set in the Zariski topology, the decomposition on the right-hand side is a sum of a different set of rank-1 tensors than the decomposition on the left-hand side, entailing that order-2 tensors of rank r>1 are generically not identifiable. The situation changes completely for higher-order tensors in F^ \otimes F^ \otimes \cdots \otimes F^ with M > 2 and all I_m \ge 2. For simplicity in notation, assume without loss of generality that the factors are ordered such that I_1 \ge I_2 \ge \cdots \ge I_M \ge 2. Let S_r \subset F^ \otimes \cdots F^\otimes \cdots \otimes F^denote the set of tensors of rank bounded by r. Then, the following statement was proved to be correct using a
computer-assisted proof A computer-assisted proof is a mathematical proof that has been at least partially generated by computer. Most computer-aided proofs to date have been implementations of large proofs-by-exhaustion of a mathematical theorem. The idea is to use a ...
for all spaces of dimension \Pi < 15000 , and it is conjectured to be valid in general: There exists a closed set Z_r in the Zariski topology such that ''every tensor'' \mathcal \in S_r\setminus Z_r''is identifiable'' (S_r is called generically identifiable in this case), unless either one of the following exceptional cases holds: # The rank is too large: r > r_E(I_1, I_2, \ldots, I_M); # The space is identifiability-unbalanced, i.e., I_1 > \prod_^M i_m - \sum_^M (I_m - 1), and the rank is too large: r \ge \prod_^M I_m - \sum_^M (I_m-1); # The space is the defective case F^4 \otimes F^4 \otimes F^3 and the rank is r=5; # The space is the defective case F^n \otimes F^n \otimes F^2 \otimes F^2, where n \ge 2, and the rank is r = 2n-1; # The space is F^4 \otimes F^4 \otimes F^4 and the rank is r=6; # The space is F^6 \otimes F^6 \otimes F^3 and the rank is r = 8; or # The space is F^2 \otimes F^2 \otimes F^2 \otimes F^2 \otimes F^2 and the rank is r=5. # The space is perfect, i.e., r_E(I_1,I_2,\ldots,I_M) = \frac is an integer, and the rank is r = r_E(I_1,I_2,\ldots,I_M). In these exceptional cases, the generic (and also minimum) number of ''complex'' decompositions is * proved to be \infty in the first 4 cases; * proved to be two in case 5; * expected to be six in case 6; * proved to be two in case 7; and * expected to be at least two in case 8 with exception of the two identifiable cases F^5 \otimes F^4 \otimes F^3 and F^3 \otimes F^2 \otimes F^2 \otimes F^2 . In summary, the generic tensor of order M > 2 and rank r < \frac that is not identifiability-unbalanced is expected to be identifiable (modulo the exceptional cases in small spaces).


Ill-posedness of the standard approximation problem

The rank approximation problem asks for the rank-r decomposition closest (in the usual Euclidean topology) to some rank-s tensor \mathcal, where r < s. That is, one seeks to solve : \min_ \, \mathcal - \sum_^ \mathbf_i^1 \otimes \mathbf_i^2 \otimes \cdots \otimes \mathbf_i^M \, _F, where \, \cdot\, _F is the
Frobenius norm In mathematics, a matrix norm is a vector norm in a vector space whose elements (vectors) are matrices (of given dimensions). Preliminaries Given a field K of either real or complex numbers, let K^ be the -vector space of matrices with m rows ...
. It was shown in a 2008 paper by de Silva and Lim that the above standard approximation problem may be ''ill-posed''. A solution to aforementioned problem may sometimes not exist because the set over which one optimizes is not closed. As such, a minimizer may not exist, even though an infimum would exist. In particular, it is known that certain so-called ''border tensors'' may be approximated arbitrarily well by a sequence of tensor of rank at most r, even though the limit of the sequence converges to a tensor of rank strictly higher than r. The rank-3 tensor : \mathcal = \mathbf \otimes \mathbf \otimes \mathbf + \mathbf \otimes \mathbf \otimes \mathbf + \mathbf \otimes \mathbf \otimes \mathbf, \quad \text \, \mathbf\, = \, \mathbf\, = 1 \text \langle \mathbf, \mathbf\rangle \ne 1 can be approximated arbitrarily well by the following sequence of rank-2 tensors : \mathcal_n = n (\mathbf + \frac \mathbf) \otimes (\mathbf + \frac \mathbf) \otimes (\mathbf + \frac \mathbf) - n \mathbf\otimes\mathbf\otimes\mathbf as n \to \infty. This example neatly illustrates the general principle that a sequence of rank-r tensors that converges to a tensor of strictly higher rank needs to admit at least two individual rank-1 terms whose norms become unbounded. Stated formally, whenever a sequence : \mathcal_n = \sum_^r \mathbf^1_ \otimes \mathbf^2_ \otimes \cdots \otimes \mathbf^M_ has the property that \mathcal_n \to \mathcal (in the Euclidean topology) as n\to\infty, then there should exist at least 1 \le i \ne j \le r such that : \, \mathbf^1_ \otimes \mathbf^2_ \otimes \cdots \otimes \mathbf^M_ \, _F \to \infty \text \, \mathbf^1_ \otimes \mathbf^2_ \otimes \cdots \otimes \mathbf^M_ \, _F \to \infty as n\to\infty. This phenomenon is often encountered when attempting to approximate a tensor using numerical optimization algorithms. It is sometimes called the problem of ''diverging components''. It was, in addition, shown that a random low-rank tensor over the reals may not admit a rank-2 approximation with positive probability, leading to the understanding that the ill-posedness problem is an important consideration when employing the tensor rank decomposition. A common partial solution to the ill-posedness problem consists of imposing an additional inequality constraint that bounds the norm of the individual rank-1 terms by some constant. Other constraints that result in a closed set, and, thus, well-posed optimization problem, include imposing positivity or a bounded
inner product In mathematics, an inner product space (or, rarely, a Hausdorff space, Hausdorff pre-Hilbert space) is a real vector space or a complex vector space with an operation (mathematics), operation called an inner product. The inner product of two ve ...
strictly less than unity between the rank-1 terms appearing in the sought decomposition.


Calculating the CPD

Alternating algorithms: *
alternating least squares Alternating may refer to: Mathematics * Alternating algebra, an algebra in which odd-grade elements square to zero * Alternating form, a function formula in algebra * Alternating group, the group of even permutations of a finite set * Alternati ...
(ALS) * alternating slice-wise diagonalisation (ASD) Direct algorithms: * pencil-based algorithms *moment-based algorithms General optimization algorithms: *
simultaneous diagonalization Simultaneity may refer to: * Relativity of simultaneity, a concept in special relativity. * Simultaneity (music), more than one complete musical texture occurring at the same time, rather than in succession * Simultaneity, a concept in Endogene ...
(SD) *
simultaneous generalized Schur decomposition Simultaneity may refer to: * Relativity of simultaneity, a concept in special relativity. * Simultaneity (music), more than one complete musical texture occurring at the same time, rather than in succession * Simultaneity, a concept in Endogeneit ...
(SGSD) * Levenberg–Marquardt (LM) * nonlinear conjugate gradient (NCG) * limited memory BFGS (L-BFGS) General polynomial system solving algorithms: * homotopy continuation


Applications

In machine learning, the CP-decomposition is the central ingredient in learning probabilistic latent variables models via the technique of moment-matching. For example, consider the multi-view model which is a probabilistic latent variable model. In this model, the generation of samples are posited as follows: there exists a hidden random variable that is not observed directly, given which, there are several
conditionally independent In probability theory, conditional independence describes situations wherein an observation is irrelevant or redundant when evaluating the certainty of a hypothesis. Conditional independence is usually formulated in terms of conditional probabil ...
random variables known as the different "views" of the hidden variable. For simplicity, assume there are three symmetrical views x of a k-state categorical hidden variable h. Then the empirical third moment of this latent variable model can be written as: T=\sum_^Pr(h=i) E h=i. In applications such as
topic modeling In statistics and natural language processing, a topic model is a type of statistical model for discovering the abstract "topics" that occur in a collection of documents. Topic modeling is a frequently used text-mining tool for discovery of hidden ...
, this can be interpreted as the co-occurrence of words in a document. Then the eigenvalues of this empirical moment tensor can be interpreted as the probability of choosing a specific topic and each column of the factor matrix E h=k/math> corresponds to probabilities of words in the vocabulary in the corresponding topic.


See also

*
Latent class analysis In statistics, a latent class model (LCM) relates a set of observed (usually discrete) multivariate variables to a set of latent variables. It is a type of latent variable model. It is called a latent class model because the latent variable is disc ...
*
Multilinear subspace learning Multilinear subspace learning is an approach to dimensionality reduction.M. A. O. Vasilescu, D. Terzopoulos (2003"Multilinear Subspace Analysis of Image Ensembles" "Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVP ...
*
Singular value decomposition In linear algebra, the singular value decomposition (SVD) is a factorization of a real or complex matrix. It generalizes the eigendecomposition of a square normal matrix with an orthonormal eigenbasis to any \ m \times n\ matrix. It is related ...
*
Tucker decomposition In mathematics, Tucker decomposition decomposes a tensor into a set of matrices and one small core tensor. It is named after Ledyard R. Tucker although it goes back to Hitchcock in 1927. Initially described as a three-mode extension of factor ana ...
*
Higher-order singular value decomposition In multilinear algebra, the higher-order singular value decomposition (HOSVD) of a tensor is a specific orthogonal Tucker decomposition. It may be regarded as one generalization of the matrix singular value decomposition. It has applications in co ...
*
Tensor decomposition In multilinear algebra, a tensor decomposition is any scheme for expressing a "data tensor" (M-way array) as a sequence of elementary operations acting on other, often simpler tensors. Many tensor decompositions generalize some matrix decompositi ...


References


Further reading

* *


External links


PARAFAC TutorialFactoMineR
(free exploratory multivariate data analysis software linked to R) {{Tensors Multilinear algebra *