In
information theory
Information theory is the scientific study of the quantification, storage, and communication of information. The field was originally established by the works of Harry Nyquist and Ralph Hartley, in the 1920s, and Claude Shannon in the 1940s. ...
, the noisy-channel coding theorem (sometimes Shannon's theorem or Shannon's limit), establishes that for any given degree of noise contamination of a communication channel, it is possible to communicate discrete data (digital
information
Information is an abstract concept that refers to that which has the power to inform. At the most fundamental level information pertains to the interpretation of that which may be sensed. Any natural process that is not completely random, ...
) nearly error-free up to a computable maximum rate through the channel. This result was presented by
Claude Shannon
Claude Elwood Shannon (April 30, 1916 – February 24, 2001) was an American mathematician, electrical engineer, and cryptographer known as a "father of information theory".
As a 21-year-old master's degree student at the Massachusetts In ...
in 1948 and was based in part on earlier work and ideas of
Harry Nyquist and
Ralph Hartley.
The Shannon limit or Shannon capacity of a communication channel refers to the maximum
rate
Rate or rates may refer to:
Finance
* Rates (tax), a type of taxation system in the United Kingdom used to fund local government
* Exchange rate, rate at which one currency will be exchanged for another
Mathematics and science
* Rate (mathema ...
of error-free data that can theoretically be transferred over the channel if the link is subject to random data transmission errors, for a particular noise level. It was first described by Shannon (1948), and shortly after published in a book by Shannon and
Warren Weaver
Warren Weaver (July 17, 1894 – November 24, 1978) was an American scientist, mathematician, and science administrator. He is widely recognized as one of the pioneers of machine translation and as an important figure in creating support for scien ...
entitled ''
The Mathematical Theory of Communication'' (1949). This founded the modern discipline of
information theory
Information theory is the scientific study of the quantification, storage, and communication of information. The field was originally established by the works of Harry Nyquist and Ralph Hartley, in the 1920s, and Claude Shannon in the 1940s. ...
.
Overview
Stated by
Claude Shannon
Claude Elwood Shannon (April 30, 1916 – February 24, 2001) was an American mathematician, electrical engineer, and cryptographer known as a "father of information theory".
As a 21-year-old master's degree student at the Massachusetts In ...
in 1948, the theorem describes the maximum possible efficiency of
error-correcting methods versus levels of noise interference and data corruption. Shannon's theorem has wide-ranging applications in both communications and
data storage
Data storage is the recording (storing) of information (data) in a storage medium. Handwriting, phonographic recording, magnetic tape, and optical discs are all examples of storage media. Biological molecules such as RNA and DNA are cons ...
. This theorem is of foundational importance to the modern field of
information theory
Information theory is the scientific study of the quantification, storage, and communication of information. The field was originally established by the works of Harry Nyquist and Ralph Hartley, in the 1920s, and Claude Shannon in the 1940s. ...
. Shannon only gave an outline of the proof. The first rigorous proof for the discrete case is due to
Amiel Feinstein Amiel is both a surname and a given name. Notable people with the name include:
Surname:
* Barbara Amiel (born 1940), writer and wife of Conrad Black
* Gausbert Amiel (fl. 13th century), troubadour
* Henri-Frédéric Amiel (1821–1881), Swiss phi ...
in 1954.
The Shannon theorem states that given a noisy channel with
channel capacity ''C'' and information transmitted at a rate ''R'', then if
there exist
codes that allow the
probability of error
In statistics, the term "error" arises in two ways. Firstly, it arises in the context of decision making, where the probability of error may be considered as being the probability of making a wrong decision and which would have a different value f ...
at the receiver to be made arbitrarily small. This means that, theoretically, it is possible to transmit information nearly without error at any rate below a limiting rate, ''C''.
The converse is also important. If
, an arbitrarily small probability of error is not achievable. All codes will have a probability of error greater than a certain positive minimal level, and this level increases as the rate increases. So, information cannot be guaranteed to be transmitted reliably across a channel at rates beyond the channel capacity. The theorem does not address the rare situation in which rate and capacity are equal.
The channel capacity
can be calculated from the physical properties of a channel; for a band-limited channel with Gaussian noise, using the
Shannon–Hartley theorem.
Simple schemes such as "send the message 3 times and use a best 2 out of 3 voting scheme if the copies differ" are inefficient error-correction methods, unable to asymptotically guarantee that a block of data can be communicated free of error. Advanced techniques such as
Reed–Solomon codes and, more recently,
low-density parity-check (LDPC) codes and
turbo code
In information theory, turbo codes (originally in French ''Turbocodes'') are a class of high-performance forward error correction (FEC) codes developed around 1990–91, but first published in 1993. They were the first practical codes to closely ...
s, come much closer to reaching the theoretical Shannon limit, but at a cost of high computational complexity. Using these highly efficient codes and with the computing power in today's
digital signal processors, it is now possible to reach very close to the Shannon limit. In fact, it was shown that LDPC codes can reach within 0.0045 dB of the Shannon limit (for binary
Additive white Gaussian noise
Additive white Gaussian noise (AWGN) is a basic noise model used in information theory to mimic the effect of many random processes that occur in nature. The modifiers denote specific characteristics:
* ''Additive'' because it is added to any nois ...
(AWGN) channels, with very long block lengths).
Mathematical statement
The basic mathematical model for a communication system is the following:
: