Nyquist–Shannon Sampling Theorem
   HOME

TheInfoList



OR:

The Nyquist–Shannon sampling theorem is a theorem in the field of
signal processing Signal processing is an electrical engineering subfield that focuses on analyzing, modifying and synthesizing ''signals'', such as audio signal processing, sound, image processing, images, and scientific measurements. Signal processing techniq ...
which serves as a fundamental bridge between continuous-time signals and discrete-time signals. It establishes a sufficient condition for a
sample rate In signal processing, sampling is the reduction of a continuous-time signal to a discrete-time signal. A common example is the conversion of a sound wave to a sequence of "samples". A sample is a value of the signal at a point in time and/or spa ...
that permits a discrete sequence of ''samples'' to capture all the information from a continuous-time signal of finite
bandwidth Bandwidth commonly refers to: * Bandwidth (signal processing) or ''analog bandwidth'', ''frequency bandwidth'', or ''radio bandwidth'', a measure of the width of a frequency range * Bandwidth (computing), the rate of data transfer, bit rate or thr ...
. Strictly speaking, the theorem only applies to a class of
mathematical function In mathematics, a function from a set to a set assigns to each element of exactly one element of .; the words map, mapping, transformation, correspondence, and operator are often used synonymously. The set is called the domain of the functi ...
s having a
Fourier transform A Fourier transform (FT) is a mathematical transform that decomposes functions into frequency components, which are represented by the output of the transform as a function of frequency. Most commonly functions of time or space are transformed, ...
that is zero outside of a finite region of frequencies. Intuitively we expect that when one reduces a continuous function to a discrete sequence and interpolates back to a continuous function, the fidelity of the result depends on the density (or
sample rate In signal processing, sampling is the reduction of a continuous-time signal to a discrete-time signal. A common example is the conversion of a sound wave to a sequence of "samples". A sample is a value of the signal at a point in time and/or spa ...
) of the original samples. The sampling theorem introduces the concept of a sample rate that is sufficient for perfect fidelity for the class of functions that are
band-limited Bandlimiting is the limiting of a signal's frequency domain representation or spectral density to zero above a certain finite frequency. A band-limited signal is one whose Fourier transform or spectral density has bounded support. A bandli ...
to a given bandwidth, such that no actual information is lost in the sampling process. It expresses the sufficient sample rate in terms of the bandwidth for the class of functions. The theorem also leads to a formula for perfectly reconstructing the original continuous-time function from the samples. Perfect reconstruction may still be possible when the sample-rate criterion is not satisfied, provided other constraints on the signal are known (see below and
compressed sensing Compressed sensing (also known as compressive sensing, compressive sampling, or sparse sampling) is a signal processing technique for efficiently acquiring and reconstructing a signal, by finding solutions to underdetermined linear systems. This ...
). In some cases (when the sample-rate criterion is not satisfied), utilizing additional constraints allows for approximate reconstructions. The fidelity of these reconstructions can be verified and quantified utilizing
Bochner's theorem In mathematics, Bochner's theorem (named for Salomon Bochner) characterizes the Fourier transform of a positive finite Borel measure on the real line. More generally in harmonic analysis, Bochner's theorem asserts that under Fourier transform a c ...
. The name ''Nyquist–Shannon sampling theorem'' honours
Harry Nyquist Harry Nyquist (, ; February 7, 1889 – April 4, 1976) was a Swedish-American physicist and electronic engineer who made important contributions to communication theory. Personal life Nyquist was born in the village Nilsby of the parish Stora Ki ...
and
Claude Shannon Claude Elwood Shannon (April 30, 1916 – February 24, 2001) was an American people, American mathematician, electrical engineering, electrical engineer, and cryptography, cryptographer known as a "father of information theory". As a 21-year-o ...
, but the theorem was also previously discovered by
E. T. Whittaker Sir Edmund Taylor Whittaker (24 October 1873 – 24 March 1956) was a British mathematician, physicist, and historian of science. Whittaker was a leading mathematical scholar of the early 20th-century who contributed widely to applied mathema ...
(published in 1915) and Shannon cited Whittaker's paper in his work. The theorem is thus also known by the names ''Whittaker–Shannon sampling theorem'', ''Whittaker–Shannon'', and ''Whittaker–Nyquist–Shannon'', and may also be referred to as the ''cardinal theorem of interpolation''.


Introduction

Sampling is a process of converting a signal (for example, a function of continuous time or space) into a sequence of values (a function of discrete time or space). Shannon's version of the theorem states:Reprint as classic paper in: ''Proc. IEEE'', Vol. 86, No. 2, (Feb 1998)
A sufficient sample-rate is therefore anything larger than 2B samples per second. Equivalently, for a given sample rate f_s, perfect reconstruction is guaranteed possible for a bandlimit B < f_s/2. When the bandlimit is too high (or there is no bandlimit), the reconstruction exhibits imperfections known as
aliasing In signal processing and related disciplines, aliasing is an effect that causes different signals to become indistinguishable (or ''aliases'' of one another) when sampled. It also often refers to the distortion or artifact that results when a ...
. Modern statements of the theorem are sometimes careful to explicitly state that x(t) must contain no sinusoidal component at exactly frequency B, or that B must be strictly less than ½ the sample rate. The threshold 2B is called the
Nyquist rate In signal processing, the Nyquist rate, named after Harry Nyquist, is a value (in units of samples per second or hertz, Hz) equal to twice the highest frequency ( bandwidth) of a given function or signal. When the function is digitized at a hi ...
and is an attribute of the continuous-time input x(t) to be sampled. The sample rate must exceed the Nyquist rate for the samples to suffice to represent x(t).  The threshold f_s/2 is called the
Nyquist frequency In signal processing, the Nyquist frequency (or folding frequency), named after Harry Nyquist, is a characteristic of a sampler, which converts a continuous function or signal into a discrete sequence. In units of cycles per second ( Hz), it ...
and is an attribute of the sampling equipment. All meaningful frequency components of the properly sampled x(t) exist below the Nyquist frequency. The condition described by these inequalities is called the ''Nyquist criterion'', or sometimes the ''Raabe condition''. The theorem is also applicable to functions of other domains, such as space, in the case of a digitized image. The only change, in the case of other domains, is the units of measure attributed to t, f_s, and B. The symbol T \triangleq 1/f_s is customarily used to represent the interval between samples and is called the ''sample period'' or ''sampling interval''. The samples of function x(t) are commonly denoted by x \triangleq x(nT) (alternatively x_n in older signal processing literature), for all integer values of n.  Another convenient definition is x \triangleq T\cdot x(nT), which preserves the energy of the signal as T varies. A mathematically ideal way to interpolate the sequence involves the use of sinc functions. Each sample in the sequence is replaced by a sinc function, centered on the time axis at the original location of the sample nT, with the amplitude of the sinc function scaled to the sample value, x Subsequently, the sinc functions are summed into a continuous function. A mathematically equivalent method uses the Dira comb and proceeds by convolving one sinc function with a series of
Dirac delta In mathematics, the Dirac delta distribution ( distribution), also known as the unit impulse, is a generalized function or distribution (mathematics), distribution over the real numbers, whose value is zero everywhere except at zero, and who ...
pulses, weighted by the sample values. Neither method is numerically practical. Instead, some type of approximation of the sinc functions, finite in length, is used. The imperfections attributable to the approximation are known as ''interpolation error''. Practical
digital-to-analog converter In electronics, a digital-to-analog converter (DAC, D/A, D2A, or D-to-A) is a system that converts a digital signal into an analog signal. An analog-to-digital converter (ADC) performs the reverse function. There are several DAC archit ...
s produce neither scaled and delayed sinc functions, nor ideal Dirac pulses. Instead they produce a
piecewise-constant In mathematics, a function (mathematics), function on the real numbers is called a step function if it can be written as a finite set, finite linear combination of indicator functions of interval (mathematics), intervals. Informally speaking, a s ...
sequence of scaled and delayed rectangular pulses (the
zero-order hold The zero-order hold (ZOH) is a mathematical model of the practical signal reconstruction done by a conventional digital-to-analog converter (DAC). That is, it describes the effect of converting a discrete-time signal to a continuous-time sign ...
), usually followed by a
lowpass filter A low-pass filter is a filter that passes signals with a frequency lower than a selected cutoff frequency and attenuates signals with frequencies higher than the cutoff frequency. The exact frequency response of the filter depends on the filter ...
(called an "anti-imaging filter") to remove spurious high-frequency replicas (images) of the original baseband signal.


Aliasing

When x(t) is a function with a
Fourier transform A Fourier transform (FT) is a mathematical transform that decomposes functions into frequency components, which are represented by the output of the transform as a function of frequency. Most commonly functions of time or space are transformed, ...
X(f): :X(f)\ \triangleq\ \int_^ x(t) \ e^ \ t, the
Poisson summation formula In mathematics, the Poisson summation formula is an equation that relates the Fourier series coefficients of the periodic summation of a function to values of the function's continuous Fourier transform. Consequently, the periodic summation of ...
indicates that the samples, x(nT), of x(t) are sufficient to create a
periodic summation In signal processing, any periodic function s_P(t) with period ''P'' can be represented by a summation of an infinite number of instances of an aperiodic function s(t), that are offset by integer multiples of ''P''. This representation is called p ...
of X(f). The result is: :X_s(f)\ \triangleq \sum_^ X\left(f - k f_s\right) = \sum_^ T\cdot x(nT)\ e^ () which is a periodic function and its equivalent representation as a
Fourier series A Fourier series () is a summation of harmonically related sinusoidal functions, also known as components or harmonics. The result of the summation is a periodic function whose functional form is determined by the choices of cycle length (or ''p ...
, whose coefficients are T\cdot x(nT). This function is also known as the
discrete-time Fourier transform In mathematics, the discrete-time Fourier transform (DTFT) is a form of Fourier analysis that is applicable to a sequence of values. The DTFT is often used to analyze samples of a continuous function. The term ''discrete-time'' refers to the ...
(DTFT) of the sample sequence. As depicted, copies of X(f) are shifted by multiples of the sampling rate f_s and combined by addition. For a band-limited function  (X(f) = 0, \text , f, \ge B)  and sufficiently large f_s, it is possible for the copies to remain distinct from each other. But if the Nyquist criterion is not satisfied, adjacent copies overlap, and it is not possible in general to discern an unambiguous X(f). Any frequency component above f_s/2 is indistinguishable from a lower-frequency component, called an ''alias'', associated with one of the copies. In such cases, the customary interpolation techniques produce the alias, rather than the original component. When the sample-rate is pre-determined by other considerations (such as an industry standard), x(t) is usually filtered to reduce its high frequencies to acceptable levels before it is sampled. The type of filter required is a
lowpass filter A low-pass filter is a filter that passes signals with a frequency lower than a selected cutoff frequency and attenuates signals with frequencies higher than the cutoff frequency. The exact frequency response of the filter depends on the filter ...
, and in this application it is called an
anti-aliasing filter An anti-aliasing filter (AAF) is a filter used before a signal sampler to restrict the bandwidth of a signal to satisfy the Nyquist–Shannon sampling theorem over the band of interest. Since the theorem states that unambiguous reconstruct ...
.


Derivation as a special case of Poisson summation

When there is no overlap of the copies (also known as "images") of X(f), the k=0 term of can be recovered by the product: :X(f) = H(f) \cdot X_s(f),      where: :H(f)\ \triangleq\ \begin1 & , f, < B \\ 0 & , f, > f_s - B. \end The sampling theorem is proved since X(f) uniquely determines x(t). All that remains is to derive the formula for reconstruction. H(f) need not be precisely defined in the region ,\ f_s-B/math> because X_s(f) is zero in that region. However, the worst case is when B=f_s/2, the Nyquist frequency. A function that is sufficient for that and all less severe cases is: :H(f) = \mathrm \left(\frac \right) = \begin1 & , f, < \frac \\ 0 & , f, > \frac, \end where rect(•) is the
rectangular function The rectangular function (also known as the rectangle function, rect function, Pi function, Heaviside Pi function, gate function, unit pulse, or the normalized boxcar function) is defined as \operatorname(t) = \Pi(t) = \left\{\begin{array}{rl ...
.  Therefore: :X(f) = \mathrm \left(\frac \right) \cdot X_s(f) ::: = \mathrm(Tf)\cdot \sum_^ T\cdot x(nT)\ e^      (from  , above). ::: = \sum_^ x(nT)\cdot \underbrace_.      The inverse transform of both sides produces the
Whittaker–Shannon interpolation formula The Whittaker–Shannon interpolation formula or sinc interpolation is a method to construct a continuous-time bandlimited function from a sequence of real numbers. The formula dates back to the works of E. Borel in 1898, and E. T. Whittaker in ...
: :x(t) = \sum_^ x(nT)\cdot \mathrm \left( \frac\right), which shows how the samples, x(nT), can be combined to reconstruct x(t). * Larger-than-necessary values of ''fs'' (smaller values of ''T''), called ''oversampling'', have no effect on the outcome of the reconstruction and have the benefit of leaving room for a ''transition band'' in which ''H''(''f'') is free to take intermediate values.
Undersampling In signal processing, undersampling or bandpass sampling is a technique where one samples a bandpass-filtered signal at a sample rate below its Nyquist rate (twice the upper cutoff frequency), but is still able to reconstruct the signal. When ...
, which causes aliasing, is not in general a reversible operation. * Theoretically, the interpolation formula can be implemented as a
low-pass filter A low-pass filter is a filter that passes signals with a frequency lower than a selected cutoff frequency and attenuates signals with frequencies higher than the cutoff frequency. The exact frequency response of the filter depends on the filter des ...
, whose impulse response is sinc(''t''/''T'') and whose input is \textstyle\sum_^ x(nT)\cdot \delta(t - nT), which is a
Dirac comb In mathematics, a Dirac comb (also known as shah function, impulse train or sampling function) is a periodic function with the formula \operatorname_(t) \ := \sum_^ \delta(t - k T) for some given period T. Here ''t'' is a real variable and th ...
function modulated by the signal samples. Practical
digital-to-analog converter In electronics, a digital-to-analog converter (DAC, D/A, D2A, or D-to-A) is a system that converts a digital signal into an analog signal. An analog-to-digital converter (ADC) performs the reverse function. There are several DAC archit ...
s (DAC) implement an approximation like the
zero-order hold The zero-order hold (ZOH) is a mathematical model of the practical signal reconstruction done by a conventional digital-to-analog converter (DAC). That is, it describes the effect of converting a discrete-time signal to a continuous-time sign ...
. In that case, oversampling can reduce the approximation error.


Shannon's original proof

Poisson shows that the Fourier series in produces the periodic summation of X(f), regardless of f_s and B. Shannon, however, only derives the series coefficients for the case f_s=2B. Virtually quoting Shannon's original paper: :Let X(\omega) be the spectrum of x(t).  Then ::x(t) = \int_^ X(\omega) e^\;\omega = \int_^ X(\omega) e^\;\omega, :because X(\omega) is assumed to be zero outside the band \left, \tfrac\ < B.  If we let t = \tfrac, where n is any positive or negative integer, we obtain: :On the left are values of x(t) at the sampling points. The integral on the right will be recognized as essentially the ''n''th coefficient in a Fourier-series expansion of the function X(\omega), taking the interval -B to B as a fundamental period. This means that the values of the samples x(n/2B) determine the Fourier coefficients in the series expansion of X(\omega).  Thus they determine X(\omega), since X(\omega) is zero for frequencies greater than ''B'', and for lower frequencies X(\omega) is determined if its Fourier coefficients are determined. But X(\omega) determines the original function x(t) completely, since a function is determined if its spectrum is known. Therefore the original samples determine the function x(t) completely. Shannon's proof of the theorem is complete at that point, but he goes on to discuss reconstruction via sinc functions, what we now call the
Whittaker–Shannon interpolation formula The Whittaker–Shannon interpolation formula or sinc interpolation is a method to construct a continuous-time bandlimited function from a sequence of real numbers. The formula dates back to the works of E. Borel in 1898, and E. T. Whittaker in ...
as discussed above. He does not derive or prove the properties of the sinc function, but these would have been familiar to engineers reading his works at the time, since the Fourier pair relationship between rect (the rectangular function) and sinc was well known. :Let x_n be the ''n''th sample. Then the function x(t) is represented by: ::x(t) = \sum_^x_n. As in the other proof, the existence of the Fourier transform of the original signal is assumed, so the proof does not say whether the sampling theorem extends to bandlimited stationary random processes.


Notes


Application to multivariable signals and images

The sampling theorem is usually formulated for functions of a single variable. Consequently, the theorem is directly applicable to time-dependent signals and is normally formulated in that context. However, the sampling theorem can be extended in a straightforward way to functions of arbitrarily many variables. Grayscale images, for example, are often represented as two-dimensional arrays (or matrices) of real numbers representing the relative intensities of
pixel In digital imaging, a pixel (abbreviated px), pel, or picture element is the smallest addressable element in a raster image, or the smallest point in an all points addressable display device. In most digital display devices, pixels are the smal ...
s (picture elements) located at the intersections of row and column sample locations. As a result, images require two independent variables, or indices, to specify each pixel uniquely—one for the row, and one for the column. Color images typically consist of a composite of three separate grayscale images, one to represent each of the three primary colors—red, green, and blue, or ''RGB'' for short. Other colorspaces using 3-vectors for colors include HSV, CIELAB, XYZ, etc. Some colorspaces such as cyan, magenta, yellow, and black (CMYK) may represent color by four dimensions. All of these are treated as
vector-valued function A vector-valued function, also referred to as a vector function, is a mathematical function of one or more variables whose range is a set of multidimensional vectors or infinite-dimensional vectors. The input of a vector-valued function could ...
s over a two-dimensional sampled domain. Similar to one-dimensional discrete-time signals, images can also suffer from aliasing if the sampling resolution, or pixel density, is inadequate. For example, a digital photograph of a striped shirt with high frequencies (in other words, the distance between the stripes is small), can cause aliasing of the shirt when it is sampled by the camera's
image sensor An image sensor or imager is a sensor that detects and conveys information used to make an image. It does so by converting the variable attenuation of light waves (as they pass through or reflect off objects) into signals, small bursts of c ...
. The aliasing appears as a
moiré pattern In mathematics, physics, and art, moiré patterns ( , , ) or moiré fringes are large-scale interference patterns that can be produced when an opaque ruled pattern with transparent gaps is overlaid on another similar pattern. For the moiré ...
. The "solution" to higher sampling in the spatial domain for this case would be to move closer to the shirt, use a higher resolution sensor, or to optically blur the image before acquiring it with the sensor using an optical low-pass filter. Another example is shown to the right in the brick patterns. The top image shows the effects when the sampling theorem's condition is not satisfied. When software rescales an image (the same process that creates the thumbnail shown in the lower image) it, in effect, runs the image through a
low-pass filter A low-pass filter is a filter that passes signals with a frequency lower than a selected cutoff frequency and attenuates signals with frequencies higher than the cutoff frequency. The exact frequency response of the filter depends on the filter des ...
first and then downsamples the image to result in a smaller image that does not exhibit the
moiré pattern In mathematics, physics, and art, moiré patterns ( , , ) or moiré fringes are large-scale interference patterns that can be produced when an opaque ruled pattern with transparent gaps is overlaid on another similar pattern. For the moiré ...
. The top image is what happens when the image is downsampled without low-pass filtering: aliasing results. The sampling theorem applies to camera systems, where the scene and lens constitute an analog spatial signal source, and the image sensor is a spatial sampling device. Each of these components is characterized by a
modulation transfer function The optical transfer function (OTF) of an optical system such as a camera, microscope, human eye, or image projector, projector specifies how different spatial frequencies are captured or transmitted. It is used by optical engineers to describe h ...
(MTF), representing the precise resolution (spatial bandwidth) available in that component. Effects of aliasing or blurring can occur when the lens MTF and sensor MTF are mismatched. When the optical image which is sampled by the sensor device contains higher spatial frequencies than the sensor, the under sampling acts as a low-pass filter to reduce or eliminate aliasing. When the area of the sampling spot (the size of the pixel sensor) is not large enough to provide sufficient
spatial anti-aliasing In digital signal processing, spatial anti-aliasing is a technique for minimizing the distortion artifacts ( aliasing) when representing a high-resolution image at a lower resolution. Anti-aliasing is used in digital photography, computer graphi ...
, a separate anti-aliasing filter (optical low-pass filter) may be included in a camera system to reduce the MTF of the optical image. Instead of requiring an optical filter, the
graphics processing unit A graphics processing unit (GPU) is a specialized electronic circuit designed to manipulate and alter memory to accelerate the creation of images in a frame buffer intended for output to a display device. GPUs are used in embedded systems, mobi ...
of
smartphone A smartphone is a portable computer device that combines mobile telephone and computing functions into one unit. They are distinguished from feature phones by their stronger hardware capabilities and extensive mobile operating systems, whic ...
cameras performs
digital signal processing Digital signal processing (DSP) is the use of digital processing, such as by computers or more specialized digital signal processors, to perform a wide variety of signal processing operations. The digital signals processed in this manner are ...
to remove aliasing with a digital filter. Digital filters also apply sharpening to amplify the contrast from the lens at high spatial frequencies, which otherwise falls off rapidly at diffraction limits. The sampling theorem also applies to post-processing digital images, such as to up or down sampling. Effects of aliasing, blurring, and sharpening may be adjusted with digital filtering implemented in software, which necessarily follows the theoretical principles.


Critical frequency

To illustrate the necessity of f_s>2B, consider the family of sinusoids generated by different values of \theta in this formula: :x(t) = \frac\ = \ \cos(2 \pi B t) - \sin(2 \pi B t)\tan(\theta ), \quad -\pi/2 < \theta < \pi/2. With f_s=2B or equivalently T=1/2B, the samples are given by: :x(nT) = \cos(\pi n) - \underbrace_\tan(\theta ) = (-1)^n . That sort of ambiguity is the reason for the ''strict'' inequality of the sampling theorem's condition.


Sampling of non-baseband signals

As discussed by Shannon: :
A similar result is true if the band does not start at zero frequency but at some higher value, and can be proved by a linear translation (corresponding physically to single-sideband modulation) of the zero-frequency case. In this case the elementary pulse is obtained from sin(''x'')/''x'' by single-side-band modulation.
That is, a sufficient no-loss condition for sampling
signal In signal processing, a signal is a function that conveys information about a phenomenon. Any quantity that can vary over space or time can be used as a signal to share messages between observers. The '' IEEE Transactions on Signal Processing' ...
s that do not have
baseband In telecommunications and signal processing, baseband is the range of frequencies occupied by a signal that has not been modulated to higher frequencies. Baseband signals typically originate from transducers, converting some other variable into a ...
components exists that involves the ''width'' of the non-zero frequency interval as opposed to its highest frequency component. See ''
Sampling (signal processing) In signal processing, sampling is the reduction of a continuous-time signal to a discrete-time signal. A common example is the conversion of a sound wave to a sequence of "samples". A sample is a value of the signal at a point in time and/or sp ...
'' for more details and examples. For example, in order to sample the
FM radio FM broadcasting is a method of radio broadcasting using frequency modulation (FM). Invented in 1933 by American engineer Edwin Armstrong, wide-band FM is used worldwide to provide high fidelity sound over broadcast radio. FM broadcasting is cap ...
signals in the frequency range of 100–102 
MHz The hertz (symbol: Hz) is the unit of frequency in the International System of Units (SI), equivalent to one event (or cycle) per second. The hertz is an SI derived unit whose expression in terms of SI base units is s−1, meaning that one he ...
, it is not necessary to sample at 204 MHz (twice the upper frequency), but rather it is sufficient to sample at 4 MHz (twice the width of the frequency interval). A bandpass condition is that ''X''(''f'') = 0, for all nonnegative ''f'' outside the open band of frequencies: \left(\frac2 f_\mathrm, \frac2 f_\mathrm\right), for some nonnegative integer ''N''. This formulation includes the normal baseband condition as the case ''N''=0. The corresponding interpolation function is the impulse response of an ideal brick-wall
bandpass filter A band-pass filter or bandpass filter (BPF) is a device that passes frequencies within a certain range and rejects ( attenuates) frequencies outside that range. Description In electronics and signal processing, a filter is usually a two-p ...
(as opposed to the ideal brick-wall
lowpass filter A low-pass filter is a filter that passes signals with a frequency lower than a selected cutoff frequency and attenuates signals with frequencies higher than the cutoff frequency. The exact frequency response of the filter depends on the filter ...
used above) with cutoffs at the upper and lower edges of the specified band, which is the difference between a pair of lowpass impulse responses: (N+1)\,\operatorname \left(\fracT\right) - N\,\operatorname\left( \fracT \right). Other generalizations, for example to signals occupying multiple non-contiguous bands, are possible as well. Even the most generalized form of the sampling theorem does not have a provably true converse. That is, one cannot conclude that information is necessarily lost just because the conditions of the sampling theorem are not satisfied; from an engineering perspective, however, it is generally safe to assume that if the sampling theorem is not satisfied then information will most likely be lost.


Nonuniform sampling

The sampling theory of Shannon can be generalized for the case of
nonuniform sampling Nonuniform sampling is a branch of sampling theory involving results related to the Nyquist–Shannon sampling theorem. Nonuniform sampling is based on Lagrange interpolation and the relationship between itself and the (uniform) sampling theorem. N ...
, that is, samples not taken equally spaced in time. The Shannon sampling theory for non-uniform sampling states that a band-limited signal can be perfectly reconstructed from its samples if the average sampling rate satisfies the Nyquist condition. Therefore, although uniformly spaced samples may result in easier reconstruction algorithms, it is not a necessary condition for perfect reconstruction. The general theory for non-baseband and nonuniform samples was developed in 1967 by
Henry Landau Henry Jacob Landau is an American mathematician known for his contributions to information theory, including the theory of bandlimited functions and on moment issues. Landau attended the Bronx High School of Science. He received an A.B. (1953), ...
. He proved that the average sampling rate (uniform or otherwise) must be twice the ''occupied'' bandwidth of the signal, assuming it is ''a priori'' known what portion of the spectrum was occupied. In the late 1990s, this work was partially extended to cover signals whose amount of occupied bandwidth was known, but the actual occupied portion of the spectrum was unknown. In the 2000s, a complete theory was developed (see the section Sampling below the Nyquist rate under additional restrictions below) using
compressed sensing Compressed sensing (also known as compressive sensing, compressive sampling, or sparse sampling) is a signal processing technique for efficiently acquiring and reconstructing a signal, by finding solutions to underdetermined linear systems. This ...
. In particular, the theory, using signal processing language, is described in this 2009 paper. They show, among other things, that if the frequency locations are unknown, then it is necessary to sample at least at twice the Nyquist criteria; in other words, you must pay at least a factor of 2 for not knowing the location of the
spectrum A spectrum (plural ''spectra'' or ''spectrums'') is a condition that is not limited to a specific set of values but can vary, without gaps, across a continuum. The word was first used scientifically in optics to describe the rainbow of colors i ...
. Note that minimum sampling requirements do not necessarily guarantee
stability Stability may refer to: Mathematics *Stability theory, the study of the stability of solutions to differential equations and dynamical systems ** Asymptotic stability ** Linear stability ** Lyapunov stability ** Orbital stability ** Structural sta ...
.


Sampling below the Nyquist rate under additional restrictions

The Nyquist–Shannon sampling theorem provides a sufficient condition for the sampling and reconstruction of a band-limited signal. When reconstruction is done via the
Whittaker–Shannon interpolation formula The Whittaker–Shannon interpolation formula or sinc interpolation is a method to construct a continuous-time bandlimited function from a sequence of real numbers. The formula dates back to the works of E. Borel in 1898, and E. T. Whittaker in ...
, the Nyquist criterion is also a necessary condition to avoid aliasing, in the sense that if samples are taken at a slower rate than twice the band limit, then there are some signals that will not be correctly reconstructed. However, if further restrictions are imposed on the signal, then the Nyquist criterion may no longer be a
necessary condition In logic and mathematics, necessity and sufficiency are terms used to describe a conditional or implicational relationship between two statements. For example, in the conditional statement: "If then ", is necessary for , because the truth of ...
. A non-trivial example of exploiting extra assumptions about the signal is given by the recent field of
compressed sensing Compressed sensing (also known as compressive sensing, compressive sampling, or sparse sampling) is a signal processing technique for efficiently acquiring and reconstructing a signal, by finding solutions to underdetermined linear systems. This ...
, which allows for full reconstruction with a sub-Nyquist sampling rate. Specifically, this applies to signals that are sparse (or compressible) in some domain. As an example, compressed sensing deals with signals that may have a low over-all bandwidth (say, the ''effective'' bandwidth ''EB''), but the frequency locations are unknown, rather than all together in a single band, so that the passband technique does not apply. In other words, the frequency spectrum is sparse. Traditionally, the necessary sampling rate is thus 2''B''. Using compressed sensing techniques, the signal could be perfectly reconstructed if it is sampled at a rate slightly lower than 2''EB''. With this approach, reconstruction is no longer given by a formula, but instead by the solution to a linear optimization program. Another example where sub-Nyquist sampling is optimal arises under the additional constraint that the samples are quantized in an optimal manner, as in a combined system of sampling and optimal
lossy compression In information technology, lossy compression or irreversible compression is the class of data compression methods that uses inexact approximations and partial data discarding to represent the content. These techniques are used to reduce data size ...
. This setting is relevant in cases where the joint effect of sampling and quantization is to be considered, and can provide a lower bound for the minimal reconstruction error that can be attained in sampling and quantizing a
random signal In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables. Stochastic processes are widely used as mathematical models of systems and phenomena that appe ...
. For stationary Gaussian random signals, this lower bound is usually attained at a sub-Nyquist sampling rate, indicating that sub-Nyquist sampling is optimal for this signal model under optimal quantization.


Historical background

The sampling theorem was implied by the work of
Harry Nyquist Harry Nyquist (, ; February 7, 1889 – April 4, 1976) was a Swedish-American physicist and electronic engineer who made important contributions to communication theory. Personal life Nyquist was born in the village Nilsby of the parish Stora Ki ...
in 1928, in which he showed that up to 2''B'' independent pulse samples could be sent through a system of bandwidth ''B''; but he did not explicitly consider the problem of sampling and reconstruction of continuous signals. About the same time,
Karl Küpfmüller Karl Küpfmüller (6 October 1897 – 26 December 1977) was a German electrical engineer, who was prolific in the areas of communications technology, measurement and control engineering, acoustics, communication theory, and theoretical electro-te ...
showed a similar result and discussed the sinc-function impulse response of a band-limiting filter, via its integral, the step-response
sine integral In mathematics, trigonometric integrals are a family of integrals involving trigonometric functions. Sine integral The different sine integral definitions are \operatorname(x) = \int_0^x\frac\,dt \operatorname(x) = -\int_x^\infty\frac ...
; this bandlimiting and reconstruction filter that is so central to the sampling theorem is sometimes referred to as a ''Küpfmüller filter'' (but seldom so in English). The sampling theorem, essentially a dual of Nyquist's result, was proved by Claude E. Shannon. V. A. Kotelnikov published similar results in 1933, as did the mathematician
E. T. Whittaker Sir Edmund Taylor Whittaker (24 October 1873 – 24 March 1956) was a British mathematician, physicist, and historian of science. Whittaker was a leading mathematical scholar of the early 20th-century who contributed widely to applied mathema ...
in 1915, J. M. Whittaker in 1935, and Gabor in 1946 ("Theory of communication"). In 1999, the
Eduard Rhein Foundation The Eduard Rhein Foundation was founded in 1976 in Hamburg (Germany) by Eduard Rhein. The goal of the foundation is to promote scientific research, learning, arts, and culture. This is done in particular by granting awards for outstanding achievemen ...
awarded Kotelnikov their Basic Research Award "for the first theoretically exact formulation of the sampling theorem". In 1948 and 1949, Claude E. Shannon published – 16 years after Vladimir Kotelnikov – the two revolutionary articles in which he founded the information theory. In Shannon 1948 the sampling theorem is formulated as “Theorem 13”: Let ''f''(''t'') contain no frequencies over W. Then :f(t) = \sum_^\infty X_n \frac, where X_n = f\left(\frac n \right). It was not until these articles were published that the theorem known as “Shannon’s sampling theorem” became common property among communication engineers, although Shannon himself writes that this is a fact which is common knowledge in the communication art. A few lines further on, however, he adds: "but in spite of its evident importance, tseems not to have appeared explicitly in the literature of communication theory".


Other discoverers

Others who have independently discovered or played roles in the development of the sampling theorem have been discussed in several historical articles, for example, by Jerri and by Lüke. For example, Lüke points out that H. Raabe, an assistant to Küpfmüller, proved the theorem in his 1939 Ph.D. dissertation; the term ''Raabe condition'' came to be associated with the criterion for unambiguous representation (sampling rate greater than twice the bandwidth). Meijering mentions several other discoverers and names in a paragraph and pair of footnotes:
As pointed out by Higgins 35 the sampling theorem should really be considered in two parts, as done above: the first stating the fact that a bandlimited function is completely determined by its samples, the second describing how to reconstruct the function using its samples. Both parts of the sampling theorem were given in a somewhat different form by J. M. Whittaker 50, 351, 353and before him also by Ogura 41, 242 They were probably not aware of the fact that the first part of the theorem had been stated as early as 1897 by Borel 527 As we have seen, Borel also used around that time what became known as the cardinal series. However, he appears not to have made the link 35 In later years it became known that the sampling theorem had been presented before Shannon to the Russian communication community by Kotel'nikov 73 In more implicit, verbal form, it had also been described in the German literature by Raabe 57 Several authors 3, 205have mentioned that Someya 96introduced the theorem in the Japanese literature parallel to Shannon. In the English literature, Weston 47introduced it independently of Shannon around the same time.28 27 Several authors, following Black 6 have claimed that this first part of the sampling theorem was stated even earlier by Cauchy, in a paper 1published in 1841. However, the paper of Cauchy does not contain such a statement, as has been pointed out by Higgins 35 28 As a consequence of the discovery of the several independent introductions of the sampling theorem, people started to refer to the theorem by including the names of the aforementioned authors, resulting in such catchphrases as “the Whittaker–Kotel’nikov–Shannon (WKS) sampling theorem" 55or even "the Whittaker–Kotel'nikov–Raabe–Shannon–Someya sampling theorem" 3 To avoid confusion, perhaps the best thing to do is to refer to it as the sampling theorem, "rather than trying to find a title that does justice to all claimants" 36


Why Nyquist?

Exactly how, when, or why
Harry Nyquist Harry Nyquist (, ; February 7, 1889 – April 4, 1976) was a Swedish-American physicist and electronic engineer who made important contributions to communication theory. Personal life Nyquist was born in the village Nilsby of the parish Stora Ki ...
had his name attached to the sampling theorem remains obscure. The term ''Nyquist Sampling Theorem'' (capitalized thus) appeared as early as 1959 in a book from his former employer,
Bell Labs Nokia Bell Labs, originally named Bell Telephone Laboratories (1925–1984), then AT&T Bell Laboratories (1984–1996) and Bell Labs Innovations (1996–2007), is an American industrial research and scientific development company owned by mult ...
, and appeared again in 1963, and not capitalized in 1965. It had been called the ''Shannon Sampling Theorem'' as early as 1954, but also just ''the sampling theorem'' by several other books in the early 1950s. In 1958, Blackman and Tukey cited Nyquist's 1928 article as a reference for ''the sampling theorem of information theory'', even though that article does not treat sampling and reconstruction of continuous signals as others did. Their glossary of terms includes these entries: ; Sampling theorem (of information theory): Nyquist's result that equi-spaced data, with two or more points per cycle of highest frequency, allows reconstruction of band-limited functions. (See ''Cardinal theorem''.) ; Cardinal theorem (of interpolation theory): A precise statement of the conditions under which values given at a doubly infinite set of equally spaced points can be interpolated to yield a continuous band-limited function with the aid of the function \frac. Exactly what "Nyquist's result" they are referring to remains mysterious. When Shannon stated and proved the sampling theorem in his 1949 article, according to Meijering, "he referred to the critical sampling interval T = \frac 1 as the ''Nyquist interval'' corresponding to the band ''W'', in recognition of Nyquist’s discovery of the fundamental importance of this interval in connection with telegraphy". This explains Nyquist's name on the critical interval, but not on the theorem. Similarly, Nyquist's name was attached to ''
Nyquist rate In signal processing, the Nyquist rate, named after Harry Nyquist, is a value (in units of samples per second or hertz, Hz) equal to twice the highest frequency ( bandwidth) of a given function or signal. When the function is digitized at a hi ...
'' in 1953 by Harold S. Black: According to the
OED The ''Oxford English Dictionary'' (''OED'') is the first and foundational historical dictionary of the English language, published by Oxford University Press (OUP). It traces the historical development of the English language, providing a co ...
, this may be the origin of the term ''Nyquist rate''. In Black's usage, it is not a sampling rate, but a signaling rate.


See also

* 44,100 Hz, a customary rate used to sample audible frequencies is based on the limits of human hearing and the sampling theorem *
Balian–Low theorem In mathematics, the Balian–Low theorem in Fourier analysis is named for Roger Balian and Francis E. Low. The theorem states that there is no well-localized window function (or Gabor atom) ''g'' either in time or frequency for an exact Gabor fram ...
, a similar theoretical lower bound on sampling rates, but which applies to time–frequency transforms *
Cheung–Marks theorem In information theory, the Cheung–Marks theorem, named after K. F. Cheung and Robert J. Marks II, specifies conditions where restoration of a signal by the sampling theorem can become ill-posed. It offers conditions whereby "reconstruction e ...
, which specifies conditions where restoration of a signal by the sampling theorem can become ill-posed *
Shannon–Hartley theorem In information theory, the Shannon–Hartley theorem tells the maximum rate at which information can be transmitted over a communications channel of a specified bandwidth in the presence of noise. It is an application of the noisy-channel coding ...
*
Nyquist ISI criterion In communications, the Nyquist ISI criterion describes the conditions which, when satisfied by a communication channel (including responses of transmit and receive filters), result in no intersymbol interference or ISI. It provides a method for con ...
*
Reconstruction from zero crossings The problem of reconstruction from zero crossings can be stated as: given the zero crossings of a continuous signal, is it possible to Signal reconstruction, reconstruct the signal (to within a constant factor)? Worded differently, what are the con ...
*
Zero-order hold The zero-order hold (ZOH) is a mathematical model of the practical signal reconstruction done by a conventional digital-to-analog converter (DAC). That is, it describes the effect of converting a discrete-time signal to a continuous-time sign ...
*
Dirac comb In mathematics, a Dirac comb (also known as shah function, impulse train or sampling function) is a periodic function with the formula \operatorname_(t) \ := \sum_^ \delta(t - k T) for some given period T. Here ''t'' is a real variable and th ...


Notes


References


Further reading

* * and (10): pp.&nbs
178–182
* * * * *


External links



Interactive simulation of the effects of inadequate sampling
Interactive presentation of the sampling and reconstruction in a web-demo
Institute of Telecommunications, University of Stuttgart


Sampling Theory For Digital Audio

Journal devoted to Sampling Theory
* * {{DEFAULTSORT:Nyquist-Shannon sampling theorem Digital signal processing Information theory Theorems in Fourier analysis Articles containing proofs Mathematical theorems in theoretical computer science Claude Shannon Telecommunication theory