In
statistical signal processing, the goal of spectral density estimation (SDE) or simply spectral estimation is to
estimate the
spectral density (also known as the
power spectral density) of a signal from a sequence of time samples of the signal. Intuitively speaking, the spectral density characterizes the
frequency content of the signal. One purpose of estimating the spectral density is to detect any
periodicities in the data, by observing peaks at the frequencies corresponding to these periodicities.
Some SDE techniques assume that a signal is composed of a limited (usually small) number of generating frequencies plus noise and seek to find the location and intensity of the generated frequencies. Others make no assumption on the number of components and seek to estimate the whole generating spectrum.
Overview
Spectrum analysis, also referred to as
frequency domain analysis or spectral density estimation, is the technical process of decomposing a complex signal into simpler parts. As described above, many physical processes are best described as a sum of many individual frequency components. Any process that quantifies the various amounts (e.g. amplitudes, powers, intensities) versus frequency (or
phase) can be called spectrum analysis.
Spectrum analysis can be performed on the entire signal. Alternatively, a signal can be broken into short segments (sometimes called ''frames''), and spectrum analysis may be applied to these individual segments.
Periodic functions (such as
) are particularly well-suited for this sub-division. General mathematical techniques for analyzing non-periodic functions fall into the category of
Fourier analysis
In mathematics, Fourier analysis () is the study of the way general functions may be represented or approximated by sums of simpler trigonometric functions. Fourier analysis grew from the study of Fourier series, and is named after Josep ...
.
The
Fourier transform
A Fourier transform (FT) is a mathematical transform that decomposes functions into frequency components, which are represented by the output of the transform as a function of frequency. Most commonly functions of time or space are transformed, ...
of a function produces a frequency spectrum which contains all of the information about the original signal, but in a different form. This means that the original function can be completely reconstructed (''synthesized'') by an
inverse Fourier transform. For perfect reconstruction, the spectrum analyzer must preserve both the
amplitude and
phase of each frequency component. These two pieces of information can be represented as a 2-dimensional vector, as a
complex number, or as magnitude (amplitude) and phase in
polar coordinates (i.e., as a
phasor
In physics and engineering, a phasor (a portmanteau of phase vector) is a complex number representing a sinusoidal function whose amplitude (''A''), angular frequency (''ω''), and initial phase (''θ'') are time-invariant. It is related to ...
). A common technique in signal processing is to consider the squared amplitude, or
power; in this case the resulting plot is referred to as a
power spectrum.
Because of reversibility, the Fourier transform is called a ''representation'' of the function, in terms of frequency instead of time; thus, it is a
frequency domain representation. Linear operations that could be performed in the time domain have counterparts that can often be performed more easily in the frequency domain. Frequency analysis also simplifies the understanding and interpretation of the effects of various time-domain operations, both linear and non-linear. For instance, only
non-linear
In mathematics and science, a nonlinear system is a system in which the change of the output is not proportional to the change of the input. Nonlinear problems are of interest to engineers, biologists, physicists, mathematicians, and many other ...
or
time-variant operations can create new frequencies in the frequency spectrum.
In practice, nearly all software and electronic devices that generate frequency spectra utilize a
discrete Fourier transform (DFT), which operates on
samples of the signal, and which provides a mathematical approximation to the full integral solution. The DFT is almost invariably implemented by an efficient algorithm called ''
fast Fourier transform
A fast Fourier transform (FFT) is an algorithm that computes the discrete Fourier transform (DFT) of a sequence, or its inverse (IDFT). Fourier analysis converts a signal from its original domain (often time or space) to a representation in th ...
'' (FFT). The array of squared-magnitude components of a DFT is a type of power spectrum called
periodogram, which is widely used for examining the frequency characteristics of noise-free functions such as
filter impulse responses and
window functions
A window is an opening in a wall, door, roof, or vehicle that allows the exchange of light and may also allow the passage of sound and sometimes air. Modern windows are usually glazed or covered in some other transparent or translucent mater ...
. But the periodogram does not provide processing-gain when applied to noiselike signals or even sinusoids at low signal-to-noise ratios. In other words, the variance of its spectral estimate at a given frequency does not decrease as the number of samples used in the computation increases. This can be mitigated by averaging over time (
Welch's method) or over frequency (
smoothing). Welch's method is widely used for spectral density estimation (SDE). However, periodogram-based techniques introduce small biases that are unacceptable in some applications. So other alternatives are presented in the next section.
Techniques
Many other techniques for spectral estimation have been developed to mitigate the disadvantages of the basic periodogram. These techniques can generally be divided into ''
non-parametric,'' ''
parametric,'' and more recently
semi-parametric (also called sparse) methods.
The non-parametric approaches explicitly estimate the
covariance or the spectrum of the process without assuming that the process has any particular structure. Some of the most common estimators in use for basic applications (e.g.
Welch's method) are non-parametric estimators closely related to the periodogram. By contrast, the parametric approaches assume that the underlying
stationary stochastic process has a certain structure that can be described using a small number of parameters (for example, using an
auto-regressive or moving average model). In these approaches, the task is to estimate the parameters of the model that describes the stochastic process. When using the semi-parametric methods, the underlying process is modeled using a non-parametric framework, with the additional assumption that the number of non-zero components of the model is small (i.e., the model is sparse). Similar approaches may also be used for missing data recovery as well as
signal reconstruction.
Following is a partial list of non-parametric spectral density estimation techniques:
*
Periodogram, the
modulus squared of the discrete Fourier transform
**
Lomb–Scargle periodogram, for which data need not be equally spaced
*
Bartlett's method In time series analysis, Bartlett's method (also known as the method of averaged periodograms), is used for estimating power spectra. It provides a way to reduce the variance of the periodogram in exchange for a reduction of resolution, compared to ...
is the average of the periodograms taken of multiple segments of the signal to reduce variance of the spectral density estimate
*
Welch's method a windowed version of Bartlett's method that uses overlapping segments
*
Multitaper
In signal processing, multitaper is a spectral density estimation technique developed by David J. Thomson. It can estimate the power spectrum ''S'X'' of a stationary ergodic finite-variance random process ''X'', given a finite contiguous real ...
is a periodogram-based method that uses multiple tapers, or windows, to form independent estimates of the spectral density to reduce variance of the spectral density estimate
*
Least-squares spectral analysis, based on
least squares
The method of least squares is a standard approach in regression analysis to approximate the solution of overdetermined systems (sets of equations in which there are more equations than unknowns) by minimizing the sum of the squares of the res ...
fitting to known frequencies
*
Non-uniform discrete Fourier transform is used when the signal samples are
unevenly spaced in time
*
Singular spectrum analysis is a nonparametric method that uses a
singular value decomposition of the
covariance matrix
In probability theory and statistics, a covariance matrix (also known as auto-covariance matrix, dispersion matrix, variance matrix, or variance–covariance matrix) is a square matrix giving the covariance between each pair of elements of ...
to estimate the spectral density
*
Short-time Fourier transform
*
Critical filter is a nonparametric method based on
information field theory that can deal with noise, incomplete data, and instrumental response functions
Below is a partial list of parametric techniques:
*
Autoregressive model (AR) estimation, which assumes that the ''n''th sample is correlated with the previous ''p'' samples.
*
Moving-average model (MA) estimation, which assumes that the ''n''th sample is correlated with noise terms in the previous ''p'' samples.
*
Autoregressive moving average (ARMA) estimation, which generalizes the AR and MA models.
*
MUltiple SIgnal Classification (MUSIC) is a popular
superresolution
Super-resolution imaging (SR) is a class of techniques that enhance (increase) the resolution of an imaging system. In optical SR the diffraction limit of systems is transcended, while in geometrical SR the resolution of digital imaging sensors ...
method.
*
Maximum entropy spectral estimation is an ''all-poles'' method useful for SDE when singular spectral features, such as sharp peaks, are expected.
And finally some examples of semi-parametric techniques:
* SParse Iterative Covariance-based Estimation (SPICE) estimation,
and the more generalized
-SPICE.
*Iterative Adaptive Approach (IAA) estimation.
*
Lasso, similar to
Least-squares spectral analysis but with a sparsity enforcing penalty.
Parametric estimation
In parametric spectral estimation, one assumes that the signal is modeled by a
stationary process which has a spectral density function (SDF)
that is a function of the frequency
and
parameters
.
The estimation problem then becomes one of estimating these parameters.
The most common form of parametric SDF estimate uses as a model an
autoregressive model of order
. A signal sequence
obeying a zero mean
process satisfies the equation
:
where the
are fixed coefficients and
is a white noise process with zero mean and ''innovation variance''
. The SDF for this process is
:
with
the sampling time interval and
the
Nyquist frequency.
There are a number of approaches to estimating the parameters
of the
process and thus the spectral density:
* The ''
Yule-Walker estimators'' are found by recursively solving the Yule-Walker equations for an
process
* The ''Burg estimators'' are found by treating the Yule-Walker equations as a form of ordinary least squares problem. The Burg estimators are generally considered superior to the Yule-Walker estimators. Burg associated these with
maximum entropy spectral estimation.
[Burg, J.P. (1967) "Maximum Entropy Spectral Analysis", ''Proceedings of the 37th Meeting of the Society of Exploration Geophysicists'', Oklahoma City, Oklahoma.]
* The ''forward-backward least-squares estimators'' treat the
process as a regression problem and solves that problem using forward-backward method. They are competitive with the Burg estimators.
* The ''maximum likelihood estimators'' estimate the parameters using a
maximum likelihood approach. This involves a nonlinear optimization and is more complex than the first three.
Alternative parametric methods include fitting to a
moving average model (MA) and to a full
autoregressive moving average model (ARMA).
Frequency estimation
Frequency estimation is the process of
estimating the
frequency, amplitude, and phase-shift of a
signal in the presence of
noise given assumptions about the number of the components.
[Hayes, Monson H., ''Statistical Digital Signal Processing and Modeling'', John Wiley & Sons, Inc., 1996. .] This contrasts with the general methods above, which do not make prior assumptions about the components.
Single tone
If one only wants to estimate the single loudest frequency, one can use a
pitch detection algorithm. If the dominant frequency changes over time, then the problem becomes the estimation of the
instantaneous frequency as defined in the
time–frequency representation. Methods for instantaneous frequency estimation include those based on the
Wigner-Ville distribution
The Wigner quasiprobability distribution (also called the Wigner function or the Wigner–Ville distribution, after Eugene Wigner and Jean-André Ville) is a quasiprobability distribution. It was introduced by Eugene Wigner in 1932 to study quan ...
and higher order
ambiguity function In pulsed radar and sonar signal processing, an ambiguity function is a two-dimensional function of propagation delay \tau and Doppler frequency f, \chi(\tau,f). It represents the distortion of a returned pulse due to the receiver matched filter ( ...
s.
If one wants to know ''all'' the (possibly complex) frequency components of a received signal (including transmitted signal and noise), one uses a multiple-tone approach.
Multiple tones
A typical model for a signal
consists of a sum of
complex exponentials in the presence of
white noise,
:
.
The power spectral density of
is composed of
impulse functions in addition to the spectral density function due to noise.
The most common methods for frequency estimation involve identifying the noise
subspace to extract these components. These methods are based on
eigen decomposition of the
autocorrelation matrix into a signal subspace and a noise subspace. After these subspaces are identified, a frequency estimation function is used to find the component frequencies from the noise subspace. The most popular methods of noise subspace based frequency estimation are
Pisarenko's method, the
multiple signal classification (MUSIC) method, the eigenvector method, and the minimum norm method.
;
Pisarenko's method:
;
MUSIC:
,
; Eigenvector method:
; Minimum norm method:
Example calculation
Suppose
, from
to
is a time series (discrete time) with zero mean. Suppose that it is a sum of a finite number of periodic components (all frequencies are positive):
:
The variance of
is, for a zero-mean function as above, given by
:
If these data were samples taken from an electrical signal, this would be its average power (power is energy per unit time, so it is analogous to variance if energy is analogous to the amplitude squared).
Now, for simplicity, suppose the signal extends infinitely in time, so we pass to the limit as
If the average power is bounded, which is almost always the case in reality, then the following limit exists and is the variance of the data.
:
Again, for simplicity, we will pass to continuous time, and assume that the signal extends infinitely in time in both directions. Then these two formulas become
:
and
:
The root mean square of
is
, so the variance of
is
Hence, the contribution to the average power of
coming from the component with frequency
is
All these contributions add up to the average power of
Then the power as a function of frequency is
and its statistical
cumulative distribution function
In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x.
Ev ...
will be
:
is a
step function, monotonically non-decreasing. Its jumps occur at the frequencies of the
periodic components of
, and the value of each jump is the power or variance of that component.
The variance is the covariance of the data with itself. If we now consider the same data but with a lag of
, we can take the
covariance of
with
, and define this to be the
autocorrelation function of the signal (or data)
:
:
If it exists, it is an even function of
If the average power is bounded, then
exists everywhere, is finite, and is bounded by
which is the average power or variance of the data.
It can be shown that
can be decomposed into periodic components with the same periods as
:
:
This is in fact the spectral decomposition of
over the different frequencies, and is related to the distribution of power of
over the frequencies: the amplitude of a frequency component of
is its contribution to the average power of the signal.
The power spectrum of this example is not continuous, and therefore does not have a derivative, and therefore this signal does not have a power spectral density function. In general, the power spectrum will usually be the sum of two parts: a line spectrum such as in this example, which is not continuous and does not have a density function, and a residue, which is absolutely continuous and does have a density function.
See also
*
Multidimensional spectral estimation
Multidimension spectral estimation is a generalization of spectral estimation, normally formulated for one-dimensional signals, to multidimensional signals or multivariate data, such as wave vectors.
Motivation
Multidimensional spectral estimatio ...
*
Periodogram
*
SigSpec
SigSpec (acronym of SIGnificance SPECtrum) is a statistical technique to provide the reliability of periodicities in a measured (noisy and not necessarily equidistant) time series. It relies on the amplitude spectrum obtained by the Discrete Fourie ...
*
Spectrogram
*
Time–frequency analysis
*
Time–frequency representation
*
Whittle likelihood
*
Spectral power distribution
In radiometry, photometry, and color science, a spectral power distribution (SPD) measurement describes the power per unit area per unit wavelength of an illumination (radiant exitance). More generally, the term ''spectral power distribution'' ...
References
Further reading
*
*
*
*
{{Statistics, analysis
Statistical signal processing
Signal estimation
Frequency-domain analysis