Linear prediction
   HOME

TheInfoList



OR:

Linear prediction is a mathematical operation where future values of a
discrete-time In mathematical dynamics, discrete time and continuous time are two alternative frameworks within which variables that evolve over time are modeled. Discrete time Discrete time views values of variables as occurring at distinct, separate "po ...
signal In signal processing, a signal is a function that conveys information about a phenomenon. Any quantity that can vary over space or time can be used as a signal to share messages between observers. The '' IEEE Transactions on Signal Processing' ...
are estimated as a
linear function In mathematics, the term linear function refers to two distinct but related notions: * In calculus and related areas, a linear function is a function whose graph is a straight line, that is, a polynomial function of degree zero or one. For dist ...
of previous samples. In
digital signal processing Digital signal processing (DSP) is the use of digital processing, such as by computers or more specialized digital signal processors, to perform a wide variety of signal processing operations. The digital signals processed in this manner are ...
, linear prediction is often called
linear predictive coding Linear predictive coding (LPC) is a method used mostly in audio signal processing and speech processing for representing the spectral envelope of a digital signal of speech in compressed form, using the information of a linear predictive mod ...
(LPC) and can thus be viewed as a subset of
filter theory Filter design is the process of designing a signal processing filter that satisfies a set of requirements, some of which may be conflicting. The purpose is to find a realization of the filter that meets each of the requirements to a sufficient ...
. In system analysis, a subfield of
mathematics Mathematics is an area of knowledge that includes the topics of numbers, formulas and related structures, shapes and the spaces in which they are contained, and quantities and their changes. These topics are represented in modern mathematics ...
, linear prediction can be viewed as a part of
mathematical model A mathematical model is a description of a system using mathematical concepts and language. The process of developing a mathematical model is termed mathematical modeling. Mathematical models are used in the natural sciences (such as physics, ...
ling or
optimization Mathematical optimization (alternatively spelled ''optimisation'') or mathematical programming is the selection of a best element, with regard to some criterion, from some set of available alternatives. It is generally divided into two subfi ...
.


The prediction model

The most common representation is :\widehat(n) = \sum_^p a_i x(n-i)\, where \widehat(n) is the predicted signal value, x(n-i) the previous observed values, with p \leq n , and a_i the predictor coefficients. The error generated by this estimate is :e(n) = x(n) - \widehat(n)\, where x(n) is the true signal value. These equations are valid for all types of (one-dimensional) linear prediction. The differences are found in the way the predictor coefficients a_i are chosen. For multi-dimensional signals the error metric is often defined as :e(n) = \, x(n) - \widehat(n)\, \, where \, \cdot\, is a suitable chosen vector norm. Predictions such as \widehat(n) are routinely used within
Kalman filter For statistics and control theory, Kalman filtering, also known as linear quadratic estimation (LQE), is an algorithm that uses a series of measurements observed over time, including statistical noise and other inaccuracies, and produces estima ...
s and smoothers to estimate current and past signal values, respectively.


Estimating the parameters

The most common choice in optimization of parameters a_i is the
root mean square In mathematics and its applications, the root mean square of a set of numbers x_i (abbreviated as RMS, or rms and denoted in formulas as either x_\mathrm or \mathrm_x) is defined as the square root of the mean square (the arithmetic mean of the ...
criterion which is also called the
autocorrelation Autocorrelation, sometimes known as serial correlation in the discrete time case, is the correlation of a signal with a delayed copy of itself as a function of delay. Informally, it is the similarity between observations of a random variable ...
criterion. In this method we minimize the expected value of the squared error E ^2(n)/math>, which yields the equation :\sum_^p a_i R(j-i) = R(j), for 1 ≤ ''j'' ≤ ''p'', where ''R'' is the
autocorrelation Autocorrelation, sometimes known as serial correlation in the discrete time case, is the correlation of a signal with a delayed copy of itself as a function of delay. Informally, it is the similarity between observations of a random variable ...
of signal ''x''''n'', defined as :\ R(i) = E\\,, and ''E'' is the
expected value In probability theory, the expected value (also called expectation, expectancy, mathematical expectation, mean, average, or first moment) is a generalization of the weighted average. Informally, the expected value is the arithmetic mean of a ...
. In the multi-dimensional case this corresponds to minimizing the L2 norm. The above equations are called the
normal equations In statistics, ordinary least squares (OLS) is a type of linear least squares method for choosing the unknown statistical parameter, parameters in a linear regression model (with fixed level-one effects of a linear function of a set of explanatory ...
or Yule-Walker equations. In matrix form the equations can be equivalently written as :\mathbf = \mathbf where the autocorrelation matrix \mathbf is a symmetric, p \times p
Toeplitz matrix In linear algebra, a Toeplitz matrix or diagonal-constant matrix, named after Otto Toeplitz, is a matrix in which each descending diagonal from left to right is constant. For instance, the following matrix is a Toeplitz matrix: :\qquad\begin a & b ...
with elements r_ = R(i-j), 0 \leq i, j

, the vector \mathbf is the autocorrelation vector r_j = R(j), 0, and \mathbf = _1, a_2, \,\cdots\, , a_, a_p/math>, the parameter vector. Another, more general, approach is to minimize the sum of squares of the errors defined in the form :e(n) = x(n) - \widehat(n) = x(n) - \sum_^p a_i x(n-i) = - \sum_^p a_i x(n-i) where the optimisation problem searching over all a_i must now be constrained with a_0=-1. On the other hand, if the mean square prediction error is constrained to be unity and the prediction error equation is included on top of the normal equations, the augmented set of equations is obtained as :\ \mathbf = , 0, ... , 0 where the index i ranges from 0 to p, and \mathbf is a (p+1)\times(p+1) matrix. Specification of the parameters of the linear predictor is a wide topic and a large number of other approaches have been proposed. In fact, the autocorrelation method is the most common and it is used, for example, for

speech coding Speech coding is an application of data compression of digital audio signals containing speech. Speech coding uses speech-specific parameter estimation using audio signal processing techniques to model the speech signal, combined with generic d ...
in the
GSM The Global System for Mobile Communications (GSM) is a standard developed by the European Telecommunications Standards Institute (ETSI) to describe the protocols for second-generation ( 2G) digital cellular networks used by mobile devices such ...
standard. Solution of the matrix equation \mathbf = \mathbf is computationally a relatively expensive process. The
Gaussian elimination In mathematics, Gaussian elimination, also known as row reduction, is an algorithm for solving systems of linear equations. It consists of a sequence of operations performed on the corresponding matrix of coefficients. This method can also be used ...
for matrix inversion is probably the oldest solution but this approach does not efficiently use the symmetry of \mathbf. A faster algorithm is the
Levinson recursion Levinson recursion or Levinson–Durbin recursion is a procedure in linear algebra to recursively calculate the solution to an equation involving a Toeplitz matrix. The algorithm runs in time, which is a strong improvement over Gauss–Jordan eli ...
proposed by Norman Levinson in 1947, which recursively calculates the solution. In particular, the autocorrelation equations above may be more efficiently solved by the Durbin algorithm. In 1986, Philippe Delsarte and Y.V. Genin proposed an improvement to this algorithm called the split Levinson recursion, which requires about half the number of multiplications and divisions.Delsarte, P. and Genin, Y. V. (1986), ''The split Levinson algorithm'', ''IEEE Transactions on Acoustics, Speech, and Signal Processing'', v. ASSP-34(3), pp. 470–478 It uses a special symmetrical property of parameter vectors on subsequent recursion levels. That is, calculations for the optimal predictor containing p terms make use of similar calculations for the optimal predictor containing p-1 terms. Another way of identifying model parameters is to iteratively calculate state estimates using
Kalman filter For statistics and control theory, Kalman filtering, also known as linear quadratic estimation (LQE), is an algorithm that uses a series of measurements observed over time, including statistical noise and other inaccuracies, and produces estima ...
s and obtaining
maximum likelihood In statistics, maximum likelihood estimation (MLE) is a method of estimating the parameters of an assumed probability distribution, given some observed data. This is achieved by maximizing a likelihood function so that, under the assumed stat ...
estimates within expectation–maximization algorithms. For equally-spaced values, a polynomial interpolation is a linear combination of the known values. If the discrete time signal is estimated to obey a polynomial of degree p-1, then the predictor coefficients a_i are given by the corresponding row of the triangle of binomial transform coefficients. This estimate might be suitable for a slowly varying signal with low noise. The predictions for the first few values of p are : \begin p=1 & : & \widehat(n) = 1x(n-1)\\ p=2 & : & \widehat(n) = 2x(n-1) - 1x(n-2) \\ p=3 & : & \widehat(n) = 3x(n-1) - 3x(n-2) + 1x(n-3)\\ p=4 & : & \widehat(n) = 4x(n-1) - 6x(n-2) + 4x(n-3) - 1x(n-4)\\ \end


See also

*
Autoregressive model In statistics, econometrics and signal processing, an autoregressive (AR) model is a representation of a type of random process; as such, it is used to describe certain time-varying processes in nature, economics, etc. The autoregressive model spe ...
*
Linear predictive analysis Linear predictive analysis is a simple form of first-order extrapolation: if it has been changing at this rate then it will probably continue to change at approximately the same rate, at least in the short term. This is equivalent to fitting a ...
*
Minimum mean square error In statistics and signal processing, a minimum mean square error (MMSE) estimator is an estimation method which minimizes the mean square error (MSE), which is a common measure of estimator quality, of the fitted values of a dependent variable. In ...
*
Prediction interval In statistical inference, specifically predictive inference, a prediction interval is an estimate of an interval in which a future observation will fall, with a certain probability, given what has already been observed. Prediction intervals are ...
* Rasta filtering


References


Further reading

* * * *


External links


PLP and RASTA (and MFCC, and inversion) in Matlab
{{DEFAULTSORT:Linear prediction Signal estimation Statistical forecasting