Generalized Singular Value Decomposition
   HOME

TheInfoList



OR:

In
linear algebra Linear algebra is the branch of mathematics concerning linear equations such as: :a_1x_1+\cdots +a_nx_n=b, linear maps such as: :(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n, and their representations in vector spaces and through matrices. ...
, the generalized singular value decomposition (GSVD) is the name of two different techniques based on the singular value decomposition (SVD). The two versions differ because one version decomposes two matrices (somewhat like the higher-order or tensor SVD) and the other version uses a set of constraints imposed on the left and right singular vectors of a single-matrix SVD.


First version: two-matrix decomposition

The generalized singular value decomposition (GSVD) is a
matrix decomposition In the mathematical discipline of linear algebra, a matrix decomposition or matrix factorization is a factorization of a matrix into a product of matrices. There are many different matrix decompositions; each finds use among a particular class of ...
on a pair of matrices which generalizes the
singular value decomposition In linear algebra, the singular value decomposition (SVD) is a factorization of a real or complex matrix. It generalizes the eigendecomposition of a square normal matrix with an orthonormal eigenbasis to any \ m \times n\ matrix. It is related ...
. It was introduced by Van Loan in 1976 and later developed by Paige and
Saunders Saunders is a surname of English and Scottish patronymic origin derived from Sander, a mediaeval form of Alexander.See also: Sander (name) People * Ab Saunders (1851–1883), American cowboy and gunman * Al Saunders (born 1947), American footb ...
, which is the version described here. In contrast to the SVD, the GSVD decomposes simultaneously a pair of matrices with the same number of columns. The SVD and the GSVD, as well as some other possible generalizations of the SVD, are extensively used in the study of the
conditioning Conditioning may refer to: Science, computing, and technology * Air conditioning, the removal of heat from indoor air for thermal comfort ** Automobile air conditioning, air conditioning in a vehicle ** Ice storage air conditioning, air condition ...
and
regularization Regularization may refer to: * Regularization (linguistics) * Regularization (mathematics) * Regularization (physics) * Regularization (solid modeling) * Regularization Law, an Israeli law intended to retroactively legalize settlements See also ...
of linear systems with respect to quadratic
semi-norm In mathematics, particularly in functional analysis, a seminorm is a vector space norm that need not be positive definite. Seminorms are intimately connected with convex sets: every seminorm is the Minkowski functional of some absorbing disk and ...
s. In the following, let \mathbb = \mathbb, or \mathbb = \mathbb.


Definition

The generalized singular value decomposition of matrices A_1 \in \mathbb^ and A_2 \in \mathbb^ is \begin A_1 & = U_1\Sigma_1 W^* D, 0_DQ^*, \\ A_2 & = U_2\Sigma_2 W^* D, 0_DQ^*, \end where * U_1 \in \mathbb^ is
unitary Unitary may refer to: Mathematics * Unitary divisor * Unitary element * Unitary group * Unitary matrix * Unitary morphism * Unitary operator * Unitary transformation * Unitary representation * Unitarity (physics) * ''E''-unitary inverse semigroup ...
, * U_2 \in \mathbb^ is unitary, * Q \in \mathbb^ is unitary, * W \in \mathbb^ is unitary, * D \in \mathbb^ is real diagonal with positive diagonal, and contains the non-zero singular values of C = \begin A_1 \\ A_2 \end in decreasing order, * 0_D = 0 \in \mathbb^ , * \Sigma_1 = \lceil I_A, S_1, 0_A \rfloor \in \mathbb^ is real non-negative block-diagonal, where S_1 = \lceil \alpha_, \dots, \alpha_ \rfloor with 1 > \alpha_ \ge \cdots \ge \alpha_ > 0, I_A = I_r, and 0_A = 0 \in \mathbb^ , * \Sigma_2 = \lceil 0_B, S_2, I_B \rfloor \in \mathbb^ is real non-negative block-diagonal, where S_2 = \lceil \beta_, \dots, \beta_ \rfloor with 0 < \beta_ \le \cdots \le \beta_ < 1, I_B = I_, and 0_B = 0 \in \mathbb^ , * \Sigma_1^* \Sigma_1 = \lceil\alpha_1^2, \dots, \alpha_k^2\rfloor, * \Sigma_2^* \Sigma_2 = \lceil\beta_1^2, \dots, \beta_k^2\rfloor, * \Sigma_1^* \Sigma_1 + \Sigma_2^* \Sigma_2 = I_k, * k = \textrm(C). We denote \alpha_1 = \cdots = \alpha_r = 1, \alpha_ = \cdots = \alpha_k = 0, \beta_1 = \cdots = \beta_r = 0, and \beta_ = \cdots = \beta_k = 1. While \Sigma_1 is diagonal, \Sigma_2 is not always diagonal, because of the leading rectangular zero matrix; instead \Sigma_2 is "bottom-right-diagonal".


Variations

There are many variations of the GSVD. These variations are related to the fact that it is always possible to multiply Q^* from the left by E E^* = I where E \in \mathbb^ is an arbitrary unitary matrix. We denote * X = ( ^* D, 0_DQ^*)^* * X^* = , R\hat^* , where R \in \mathbb^ is upper-triangular and invertible, and \hat \in \mathbb^ is unitary. Such matrices exist by RQ-decomposition. * Y = W^* D. Then Y is invertible. Here are some variations of the GSVD: *
MATLAB MATLAB (an abbreviation of "MATrix LABoratory") is a proprietary multi-paradigm programming language and numeric computing environment developed by MathWorks. MATLAB allows matrix manipulations, plotting of functions and data, implementation ...
(gsvd): \begin A_1 & = U_1 \Sigma_1 X^*, \\ A_2 & = U_2 \Sigma_2 X^*. \end *
LAPACK LAPACK ("Linear Algebra Package") is a standard software library for numerical linear algebra. It provides routines for solving systems of linear equations and linear least squares, eigenvalue problems, and singular value decomposition. It also ...
(LA_GGSVD): \begin A_1 & = U_1 \Sigma_1 , R\hat^*, \\ A_2 & = U_2 \Sigma_2 , R\hat^*. \end * Simplified: \begin A_1 & = U_1\Sigma_1 Y, 0_DQ^*, \\ A_2 & = U_2\Sigma_2 Y, 0_DQ^*. \end


Generalized singular values

A ''generalized singular value'' of A_1 and A_2 is a pair (a, b) \in \mathbb^2 such that \begin \lim_ \det(b^2 A_1^* A_1 - a^2 A_2^* A_2 + \delta I_n) / \det(\delta I_) & = 0, \\ a^2 + b^2 & = 1, \\ a, b & \geq 0. \end We have * A_i A_j^* = U_i \Sigma_i Y Y^* \Sigma_j^* U_j^* * A_i^* A_j = Q \begin Y^* \Sigma_i^* \Sigma_j Y & 0 \\ 0 & 0 \end Q^* = Q_1 Y^* \Sigma_i^* \Sigma_j Y Q_1^* By these properties we can show that the generalized singular values are exactly the pairs (\alpha_i, \beta_i). We have \begin & \det(b^2 A_1^* A_1 - a^2 A_2^* A_2 + \delta I_n) \\ = & \det(b^2 A_1^* A_1 - a^2 A_2^* A_2 + \delta Q Q^*) \\ = & \det\left(Q \begin Y^* (b^2 \Sigma_1^* \Sigma_1 - a^2 \Sigma_2^* \Sigma_2) Y + \delta I_k & 0 \\ 0 & \delta I_ \end Q^*\right) \\ = & \det(\delta I_) \det(Y^* (b^2 \Sigma_1^* \Sigma_1 - a^2 \Sigma_2^* \Sigma_2) Y + \delta I_k). \end Therefore : \begin & \lim_ \det(b^2 A_1^* A_1 - a^2 A_2^* A_2 + \delta I_n) / \det(\delta I_) \\ = & \lim_ \det(Y^* (b^2 \Sigma_1^* \Sigma_1 - a^2 \Sigma_2^* \Sigma_2) Y + \delta I_k) \\ = & \det(Y^* (b^2 \Sigma_1^* \Sigma_1 - a^2 \Sigma_2^* \Sigma_2) Y) \\ = & , \det(Y), ^2 \prod_^k (b^2 \alpha_i^2 - a^2 \beta_i^2). \end This expression is zero exactly when a = \alpha_i and b = \beta_i for some i. In, the generalized singular values are claimed to be those which solve \det(b^2 A_1^* A_1 - a^2 A_2^* A_2) = 0. However, this claim only holds when k = n, since otherwise the determinant is zero for every pair (a, b) \in \mathbb^2; this can be seen by substituting \delta = 0 above.


Generalized inverse

Define E^+ = E^ for any invertible matrix E \in \mathbb^ , 0^+ = 0^* for any zero matrix 0 \in \mathbb^, and \left\lceil E_1, E_2 \right\rfloor^+ = \left\lceil E_1^+, E_2^+ \right\rfloor for any block-diagonal matrix. Then defineA_i^+ = Q \begin Y^ \\ 0 \end \Sigma_i^+ U_i^*It can be shown that A_i^+ as defined here is a
generalized inverse In mathematics, and in particular, algebra, a generalized inverse (or, g-inverse) of an element ''x'' is an element ''y'' that has some properties of an inverse element but not necessarily all of them. The purpose of constructing a generalized inv ...
of A_i; in particular a \-inverse of A_i. Since it does not in general satisfy (A_i^+ A_i)^* = A_i^+ A_i, this is not the
Moore–Penrose inverse In mathematics, and in particular linear algebra, the Moore–Penrose inverse of a matrix is the most widely known generalization of the inverse matrix. It was independently described by E. H. Moore in 1920, Arne Bjerhammar in 1951, and Roger Pe ...
; otherwise we could derive (AB)^+ = B^+ A^+ for any choice of matrices, which only holds for certain class of matrices. Suppose Q = \beginQ_1 & Q_2\end , where Q_1 \in \mathbb^ and Q_2 \in \mathbb^. This generalized inverse has the following properties: * \Sigma_1^+ = \lceil I_A, S_1^, 0_A^T \rfloor * \Sigma_2^+ = \lceil 0^T_B, S_2^, I_B \rfloor * \Sigma_1 \Sigma_1^+ = \lceil I, I, 0 \rfloor * \Sigma_2 \Sigma_2^+ = \lceil 0, I, I \rfloor * \Sigma_1 \Sigma_2^+ = \lceil 0, S_1 S_2^, 0 \rfloor * \Sigma_1^+ \Sigma_2 = \lceil 0, S_1^ S_2, 0 \rfloor * A_i A_j^+ = U_i \Sigma_i \Sigma_j^+ U_j^* * A_i^+ A_j = Q \begin Y^ \Sigma_i^+ \Sigma_j Y & 0 \\ 0 & 0 \end Q^* = Q_1 Y^ \Sigma_i^+ \Sigma_j Y Q_1^*


Quotient SVD

A ''generalized singular ratio'' of A_1 and A_2 is \sigma_i=\alpha_i \beta_i^+. By the above properties, A_1 A_2^+ = U_1 \Sigma_1 \Sigma_2^+ U_2^*. Note that \Sigma_1 \Sigma_2^+ = \lceil 0, S_1 S_2^, 0 \rfloor is diagonal, and that, ignoring the leading zeros, contains the singular ratios in decreasing order. If A_2 is invertible, then \Sigma_1 \Sigma_2^+ has no leading zeros, and the generalized singular ratios are the singular values, and U_1 and U_2 are the matrices of singular vectors, of the matrix A_1 A_2^+ = A_1 A_2^. In fact, computing the SVD of A_1 A_2^ is one of the motivations for the GSVD, as "forming AB^ and finding its SVD can lead to unnecessary and large numerical errors when B is ill-conditioned for solution of equations". Hence the sometimes used name "quotient SVD", although this is not the only reason for using GSVD. If A_2 is not invertible, then U_1 \Sigma_1 \Sigma_2^+ U_2^*is still the SVD of A_1 A_2^+ if we relax the requirement of having the singular values in decreasing order. Alternatively, a decreasing order SVD can be found by moving the leading zeros to the back: U_1 \Sigma_1 \Sigma_2^+ U_2^* = (U_1 P_1) P_1^* \Sigma_1 \Sigma_2^+ P_2 (P_2^* U_2^*), where P_1 and P_2 are appropriate permutation matrices. Since rank equals the number of non-zero singular values, \mathrm(A_1 A_2^+)=s.


Construction

Let * C = P \lceil D, 0 \rfloor Q^* be the SVD of C = \begin A_1 \\ A_2 \end, where P \in \mathbb^ is unitary, and Q and D are as described, * P = _1, P_2/math>, where P_1 \in \mathbb^ and P_2 \in \mathbb^, * P_1 = \begin P_ \\ P_ \end, where P_ \in \mathbb^ and P_ \in \mathbb^, * P_ = U_1 \Sigma_1 W^* by the SVD of P_, where U_1, \Sigma_1 and W are as described, * P_ W = U_2 \Sigma_2 by a decomposition similar to a QR-decomposition, where U_2 and \Sigma_2 are as described. Then\begin C & = P \lceil D, 0 \rfloor Q^* \\ & = _1 D, 0Q^* \\ & = \begin U_1 \Sigma_1 W^* D & 0 \\ U_2 \Sigma_2 W^* D & 0 \end Q^* \\ & = \begin U_1 \Sigma_1 ^* D, 0Q^* \\ U_2 \Sigma_2 ^* D, 0Q^* \end . \endWe also have\begin U_1^* & 0 \\ 0 & U_2^* \end P_1 W = \begin \Sigma_1 \\ \Sigma_2 \end.Therefore\Sigma_1^* \Sigma_1 + \Sigma_2^* \Sigma_2 = \begin \Sigma_1 \\ \Sigma_2 \end^* \begin \Sigma_1 \\ \Sigma_2 \end = W^* P_1^* \begin U_1 & 0 \\ 0 & U_2 \end \begin U_1^* & 0 \\ 0 & U_2^* \end P_1 W = I.Since P_1 has orthonormal columns, , , P_1, , _2 \leq 1. Therefore, , \Sigma_1, , _2 = , , U_1^* P_1 W, , _2 = , , P_1, , _2 \leq 1.We also have for each x \in \mathbb^k such that , , x, , _2 = 1 that, , P_ x, , _2^2 \leq , , P_ x, , _2^2 + , , P_ x, , _2^2 = , , P_ x, , _2^2 \leq 1.Therefore , , P_, , _2 \leq 1, and, , \Sigma_2, , _2 = , , U_2^* P_ W , , _2 = , , P_, , _2 \leq 1.


Applications

The GSVD, formulated as a comparative spectral decomposition, has been successfully applied to signal processing and data science, e.g., in genomic signal processing. These applications inspired several additional comparative spectral decompositions, i.e., the higher-order GSVD (HO GSVD) and the tensor GSVD.


Second version: weighted single-matrix decomposition

The weighted version of the generalized singular value decomposition (GSVD) is a constrained
matrix decomposition In the mathematical discipline of linear algebra, a matrix decomposition or matrix factorization is a factorization of a matrix into a product of matrices. There are many different matrix decompositions; each finds use among a particular class of ...
with constraints imposed on the left and right singular vectors of the
singular value decomposition In linear algebra, the singular value decomposition (SVD) is a factorization of a real or complex matrix. It generalizes the eigendecomposition of a square normal matrix with an orthonormal eigenbasis to any \ m \times n\ matrix. It is related ...
. This form of the ''GSVD'' is an extension of the ''SVD'' as such. Given the ''SVD'' of an ''m×n'' real or complex matrix ''M'' :M = U\Sigma V^* \, where :U^* W_u U = V^* W_v V = I. Where ''I'' is the
identity matrix In linear algebra, the identity matrix of size n is the n\times n square matrix with ones on the main diagonal and zeros elsewhere. Terminology and notation The identity matrix is often denoted by I_n, or simply by I if the size is immaterial o ...
and where U and V are orthonormal given their constraints (W_u and W_v). Additionally, W_u and W_v are positive definite matrices (often diagonal matrices of weights). This form of the ''GSVD'' is the core of certain techniques, such as generalized principal component analysis and
Correspondence analysis Correspondence analysis (CA) is a multivariate statistical technique proposed by Herman Otto Hartley (Hirschfeld) and later developed by Jean-Paul Benzécri. It is conceptually similar to principal component analysis, but applies to categorical rat ...
. The weighted form of the ''GSVD'' is called as such because, with the correct selection of weights, it ''generalizes'' many techniques (such as
multidimensional scaling Multidimensional scaling (MDS) is a means of visualizing the level of similarity of individual cases of a dataset. MDS is used to translate "information about the pairwise 'distances' among a set of n objects or individuals" into a configurati ...
and
linear discriminant analysis Linear discriminant analysis (LDA), normal discriminant analysis (NDA), or discriminant function analysis is a generalization of Fisher's linear discriminant, a method used in statistics and other fields, to find a linear combination of features ...
)


References


Further reading

* *
LAPACK LAPACK ("Linear Algebra Package") is a standard software library for numerical linear algebra. It provides routines for solving systems of linear equations and linear least squares, eigenvalue problems, and singular value decomposition. It also ...
manua

{{refend Linear algebra Singular value decomposition