Sparse approximation (also known as sparse representation) theory deals with
sparse solutions for
systems of linear equations
In mathematics, a system of linear equations (or linear system) is a collection of one or more linear equations involving the same variables.
For example,
:\begin
3x+2y-z=1\\
2x-2y+4z=-2\\
-x+\fracy-z=0
\end
is a system of three equations in th ...
. Techniques for finding these solutions and exploiting them in applications have found wide use in
image processing
An image is a visual representation of something. It can be two-dimensional, three-dimensional, or somehow otherwise feed into the visual system to convey information. An image can be an artifact, such as a photograph or other two-dimensiona ...
,
signal processing
Signal processing is an electrical engineering subfield that focuses on analyzing, modifying and synthesizing ''signals'', such as audio signal processing, sound, image processing, images, and scientific measurements. Signal processing techniq ...
,
machine learning
Machine learning (ML) is a field of inquiry devoted to understanding and building methods that 'learn', that is, methods that leverage data to improve performance on some set of tasks. It is seen as a part of artificial intelligence.
Machine ...
,
medical imaging
Medical imaging is the technique and process of imaging the interior of a body for clinical analysis and medical intervention, as well as visual representation of the function of some organs or tissues (physiology). Medical imaging seeks to rev ...
, and more.
Sparse decomposition
Noiseless observations
Consider a
linear system of equations
In mathematics, a system of linear equations (or linear system) is a collection of one or more linear equations involving the same variables.
For example,
:\begin
3x+2y-z=1\\
2x-2y+4z=-2\\
-x+\fracy-z=0
\end
is a system of three equations in ...
, where
is an
underdetermined matrix
Matrix most commonly refers to:
* ''The Matrix'' (franchise), an American media franchise
** ''The Matrix'', a 1999 science-fiction action film
** "The Matrix", a fictional setting, a virtual reality environment, within ''The Matrix'' (franchis ...
and
. The matrix
(typically assumed to be full-rank) is referred to as the dictionary, and
is a signal of interest. The core sparse representation problem is defined as the quest for the sparsest possible representation
satisfying
. Due to the underdetermined nature of
, this linear system admits in general infinitely many possible solutions, and among these we seek the one with the fewest non-zeros. Put formally, we solve
:
where
is the
pseudo-norm, which counts the
number
A number is a mathematical object used to count, measure, and label. The original examples are the natural numbers 1, 2, 3, 4, and so forth. Numbers can be represented in language with number words. More universally, individual numbers c ...
of non-zero components of
. This problem is known to be NP-hard with a reduction to NP-complete subset selection problems in
combinatorial optimization
Combinatorial optimization is a subfield of mathematical optimization that consists of finding an optimal object from a finite set of objects, where the set of feasible solutions is discrete or can be reduced to a discrete set. Typical combi ...
.
Sparsity of
implies that only a few (
) components in it are non-zero. The underlying motivation for such a sparse decomposition is the desire to provide the simplest possible explanation of
as a linear combination of as few as possible columns from
, also referred to as atoms. As such, the signal
can be viewed as a molecule composed of a few fundamental elements taken from
.
While the above posed problem is indeed NP-Hard, its solution can often be found using approximation algorithms. One such option is a convex
relaxation of the problem, obtained by using the
-norm instead of
, where
simply sums the absolute values of the entries in
. This is known as the
basis pursuit
Basis pursuit is the mathematical optimization problem of the form
: \min_x \, x\, _1 \quad \text \quad y = Ax,
where ''x'' is a ''N''-dimensional solution vector (signal), ''y'' is a ''M''-dimensional vector of observations (measurements), ''A' ...
(BP) algorithm, which can be handled using any
linear programming
Linear programming (LP), also called linear optimization, is a method to achieve the best outcome (such as maximum profit or lowest cost) in a mathematical model whose requirements are represented by linear function#As a polynomial function, li ...
solver. An alternative approximation method is a greedy technique, such as the
matching pursuit
Matching pursuit (MP) is a sparse approximation algorithm which finds the "best matching" projections of multidimensional data onto the span of an over-complete (i.e., redundant) dictionary D. The basic idea is to approximately represent a signal ...
(MP), which finds the location of the non-zeros one at a time.
Surprisingly, under mild conditions on
(using the
spark (mathematics)
In mathematics, more specifically in linear algebra, the spark of a m \times n matrix A is the smallest integer k such that there exists a set of k columns in A which are linearly dependent. If all the columns are linearly independent, \mathrm(A) ...
, the
mutual coherence or the
restricted isometry property In linear algebra, the restricted isometry property (RIP) characterizes matrices which are nearly orthonormal, at least when operating on sparse vectors. The concept was introduced by Emmanuel Candès and Terence TaoE. J. Candes and T. Tao, "Decodin ...
) and the level of sparsity in the solution,
, the sparse representation problem can be shown to have a unique solution, and BP and MP are guaranteed to find it perfectly.
Noisy observations
Often the observed signal
is noisy. By relaxing the equality constraint and imposing an
-norm on the data-fitting term, the sparse decomposition problem becomes
:
or put in a Lagrangian form,
:
where
is replacing the
.
Just as in the noiseless case, these two problems are NP-Hard in general, but can be approximated using pursuit algorithms. More specifically, changing the
to an
-norm, we obtain
:
which is known as the
basis pursuit denoising
In applied mathematics and statistics, basis pursuit denoising (BPDN) refers to a mathematical optimization problem of the form
: \min_x \left(\frac \, y - Ax\, ^2_2 + \lambda \, x\, _1\right),
where \lambda is a parameter that controls the trade ...
. Similarly,
matching pursuit
Matching pursuit (MP) is a sparse approximation algorithm which finds the "best matching" projections of multidimensional data onto the span of an over-complete (i.e., redundant) dictionary D. The basic idea is to approximately represent a signal ...
can be used for approximating the solution of the above problems, finding the locations of the non-zeros one at a time until the error threshold is met. Here as well, theoretical guarantees suggest that BP and MP lead to nearly optimal solutions depending on the properties of
and the cardinality of the solution
.
Another interesting theoretical result refers to the case in which
is a
unitary matrix
In linear algebra, a complex square matrix is unitary if its conjugate transpose is also its inverse, that is, if
U^* U = UU^* = UU^ = I,
where is the identity matrix.
In physics, especially in quantum mechanics, the conjugate transpose is ...
. Under this assumption, the problems posed above (with either
or
) admit closed-form solutions in the form of non-linear shrinkage.
Variations
There are several variations to the basic sparse approximation problem.
Structured sparsity: In the original version of the problem, any of the atoms in the dictionary can be picked. In the structured (block) sparsity model, instead of picking atoms individually, groups of them are to be picked. These groups can be overlapping and of varying size. The objective is to represent
such that it is sparse while forcing this block-structure.
Collaborative (joint) sparse coding: The original version of the problem is defined for a single signal
. In the collaborative (joint) sparse coding model, a set of signals is available, each believed to emerge from (nearly) the same set of atoms from
. In this case, the pursuit task aims to recover a set of sparse representations that best describe the data while forcing them to share the same (or close-by) support.
Other structures: More broadly, the sparse approximation problem can be cast while forcing a specific desired structure on the pattern of non-zero locations in
. Two cases of interest that have been extensively studied are tree-based structure, and more generally, a Boltzmann distributed support.
Algorithms
As already mentioned above, there are various approximation (also referred to as pursuit) algorithms that have been developed for addressing the sparse representation problem:
:
We mention below a few of these main methods.
*
Matching pursuit
Matching pursuit (MP) is a sparse approximation algorithm which finds the "best matching" projections of multidimensional data onto the span of an over-complete (i.e., redundant) dictionary D. The basic idea is to approximately represent a signal ...
is a greedy iterative algorithm for approximately solving the above problem. It works by gradually finding the locations of the non-zeros in
one at a time. The core idea is to find in each step the column (atom) in
that best correlates with the current residual (initialized to
), and then updating this residual to take the new atom and its coefficient into account. Matching pursuit might pick the same atom multiple times.
* Orthogonal matching pursuit is very similar to matching pursuit, with one major difference: in each of the algorithm's step, all the non-zero coefficients are updated by a
least squares
The method of least squares is a standard approach in regression analysis to approximate the solution of overdetermined systems (sets of equations in which there are more equations than unknowns) by minimizing the sum of the squares of the res ...
. As a consequence, the residual is orthogonal to the already chosen atoms, and thus an atom cannot be picked more than once.
* Stage-wise greedy methods: Improved variations over the above are algorithms that operate greedily while adding two critical features: (i) the ability to add groups of non-zeros at a time (instead of one non-zero per round); and (ii) including a pruning step in each round in which several of the atoms are discarded from the support. Representatives of this approach are the Subspace-Pursuit algorithm and the CoSaMP.
*
Basis pursuit
Basis pursuit is the mathematical optimization problem of the form
: \min_x \, x\, _1 \quad \text \quad y = Ax,
where ''x'' is a ''N''-dimensional solution vector (signal), ''y'' is a ''M''-dimensional vector of observations (measurements), ''A' ...
solves a convex relaxed version of the problem by replacing the
by an
-norm. Note that this only defines a new objective, while leaving open the question of the algorithm to use for getting the desired solution. Commonly considered such algorithms are the
IRLS,
LARS
Lars is a common male name in Scandinavian countries.
Origin
''Lars'' means "from the city of Laurentum". Lars is derived from the Latin name Laurentius, which means "from Laurentum" or "crowned with laurel".
A homonymous Etruscan name was bo ...
, and iterative soft-shrinkage methods.
* There are several other methods for solving sparse decomposition problems: homotopy method,
coordinate descent Coordinate descent is an optimization algorithm that successively minimizes along coordinate directions to find the minimum of a function. At each iteration, the algorithm determines a coordinate or coordinate block via a coordinate selection rule, ...
, iterative hard-thresholding, first order
proximal methods, which are related to the above-mentioned iterative soft-shrinkage algorithms, and Dantzig selector.
Applications
Sparse approximation ideas and algorithms have been extensively used in
signal processing
Signal processing is an electrical engineering subfield that focuses on analyzing, modifying and synthesizing ''signals'', such as audio signal processing, sound, image processing, images, and scientific measurements. Signal processing techniq ...
,
image processing
An image is a visual representation of something. It can be two-dimensional, three-dimensional, or somehow otherwise feed into the visual system to convey information. An image can be an artifact, such as a photograph or other two-dimensiona ...
,
machine learning
Machine learning (ML) is a field of inquiry devoted to understanding and building methods that 'learn', that is, methods that leverage data to improve performance on some set of tasks. It is seen as a part of artificial intelligence.
Machine ...
,
medical imaging
Medical imaging is the technique and process of imaging the interior of a body for clinical analysis and medical intervention, as well as visual representation of the function of some organs or tissues (physiology). Medical imaging seeks to rev ...
,
array processing
Array processing is a wide area of research in the field of signal processing that extends from the simplest form of 1 dimensional line arrays to 2 and 3 dimensional array geometries. Array structure can be defined as a set of sensors that are sp ...
,
data mining, and more. In most of these applications, the unknown signal of interest is modeled as a sparse combination of a few atoms from a given dictionary, and this is used as the
regularization
Regularization may refer to:
* Regularization (linguistics)
* Regularization (mathematics)
* Regularization (physics)
In physics, especially quantum field theory, regularization is a method of modifying observables which have singularities in ...
of the problem. These problems are typically accompanied by a
dictionary learning
Sparse coding is a representation learning method which aims at finding a sparse representation of the input data (also known as sparse coding) in the form of a linear combination of basic elements as well as those basic elements themselves. These ...
mechanism that aims to fit
to best match the model to the given data. The use of sparsity-inspired models has led to state-of-the-art results in a wide set of applications.
Recent work suggests that there is a tight connection between sparse representation modeling and deep-learning.
See also
*
Compressed sensing
Compressed sensing (also known as compressive sensing, compressive sampling, or sparse sampling) is a signal processing technique for efficiently acquiring and reconstructing a Signal (electronics), signal, by finding solutions to Underdetermined ...
*
Sparse dictionary learning
Sparse coding is a representation learning method which aims at finding a sparse representation of the input data (also known as sparse coding) in the form of a linear combination of basic elements as well as those basic elements themselves. Thes ...
*
K-SVD
In applied mathematics, K-SVD is a dictionary learning algorithm for creating a dictionary for sparse representations, via a singular value decomposition approach. K-SVD is a generalization of the k-means clustering method, and it works by iterati ...
*
Lasso (statistics)
In statistics and machine learning, lasso (least absolute shrinkage and selection operator; also Lasso or LASSO) is a regression analysis method that performs both variable selection and regularization in order to enhance the prediction accuracy ...
*
Regularization (mathematics)
In mathematics, statistics, finance, computer science, particularly in machine learning and inverse problems, regularization is a process that changes the result answer to be "simpler". It is often used to obtain results for ill-posed problems o ...
and
inverse problems
An inverse problem in science is the process of calculating from a set of observations the causal factors that produced them: for example, calculating an image in X-ray computed tomography, source reconstruction in acoustics, or calculating the ...
References
{{Numerical linear algebra
Numerical linear algebra