HOME

TheInfoList



OR:

Channel capacity, in
electrical engineering Electrical engineering is an engineering discipline concerned with the study, design, and application of equipment, devices, and systems that use electricity, electronics, and electromagnetism. It emerged as an identifiable occupation in the l ...
,
computer science Computer science is the study of computation, information, and automation. Computer science spans Theoretical computer science, theoretical disciplines (such as algorithms, theory of computation, and information theory) to Applied science, ...
, and
information theory Information theory is the mathematical study of the quantification (science), quantification, Data storage, storage, and telecommunications, communication of information. The field was established and formalized by Claude Shannon in the 1940s, ...
, is the theoretical maximum rate at which
information Information is an Abstraction, abstract concept that refers to something which has the power Communication, to inform. At the most fundamental level, it pertains to the Interpretation (philosophy), interpretation (perhaps Interpretation (log ...
can be reliably transmitted over a
communication channel A communication channel refers either to a physical transmission medium such as a wire, or to a logical connection over a multiplexed medium such as a radio channel in telecommunications and computer networking. A channel is used for infor ...
. Following the terms of the
noisy-channel coding theorem In information theory, the noisy-channel coding theorem (sometimes Shannon's theorem or Shannon's limit), establishes that for any given degree of noise contamination of a communication channel, it is possible (in theory) to communicate discrete ...
, the channel capacity of a given channel is the highest information rate (in units of
information Information is an Abstraction, abstract concept that refers to something which has the power Communication, to inform. At the most fundamental level, it pertains to the Interpretation (philosophy), interpretation (perhaps Interpretation (log ...
per unit time) that can be achieved with arbitrarily small error probability.
Information theory Information theory is the mathematical study of the quantification (science), quantification, Data storage, storage, and telecommunications, communication of information. The field was established and formalized by Claude Shannon in the 1940s, ...
, developed by Claude E. Shannon in 1948, defines the notion of channel capacity and provides a mathematical model by which it may be computed. The key result states that the capacity of the channel, as defined above, is given by the maximum of the
mutual information In probability theory and information theory, the mutual information (MI) of two random variables is a measure of the mutual Statistical dependence, dependence between the two variables. More specifically, it quantifies the "Information conten ...
between the input and output of the channel, where the maximization is with respect to the input distribution. The notion of channel capacity has been central to the development of modern wireline and wireless communication systems, with the advent of novel error correction coding mechanisms that have resulted in achieving performance very close to the limits promised by channel capacity.


Formal definition

The basic mathematical model for a communication system is the following: :\xrightarrow
text Text may refer to: Written word * Text (literary theory) In literary theory, a text is any object that can be "read", whether this object is a work of literature, a street sign, an arrangement of buildings on a city block, or styles of clothi ...
\begin\hline \text \\ f_n \\ \hline\end \xrightarrow mathrm\begin\hline \text \\ p(y, x) \\ \hline\end \xrightarrow mathrm \begin\hline \text \\ g_n \\ \hline\end \xrightarrow mathrm where: * W is the message to be transmitted; * X is the channel input symbol (X^n is a sequence of n symbols) taken in an
alphabet An alphabet is a standard set of letter (alphabet), letters written to represent particular sounds in a spoken language. Specifically, letters largely correspond to phonemes as the smallest sound segments that can distinguish one word from a ...
\mathcal; * Y is the channel output symbol (Y^n is a sequence of n symbols) taken in an alphabet \mathcal; * \hat is the estimate of the transmitted message; * f_n is the encoding function for a block of length n; * p(y, x) = p_(y, x) is the noisy channel, which is modeled by a
conditional probability distribution In probability theory and statistics, the conditional probability distribution is a probability distribution that describes the probability of an outcome given the occurrence of a particular event. Given two jointly distributed random variables X ...
; and, * g_n is the decoding function for a block of length n. Let X and Y be modeled as random variables. Furthermore, let p_(y, x) be the
conditional probability distribution In probability theory and statistics, the conditional probability distribution is a probability distribution that describes the probability of an outcome given the occurrence of a particular event. Given two jointly distributed random variables X ...
function of Y given X, which is an inherent fixed property of the communication channel. Then the choice of the
marginal distribution In probability theory and statistics, the marginal distribution of a subset of a collection of random variables is the probability distribution of the variables contained in the subset. It gives the probabilities of various values of the variable ...
p_X(x) completely determines the
joint distribution A joint or articulation (or articular surface) is the connection made between bones, ossicles, or other hard structures in the body which link an animal's skeletal system into a functional whole.Saladin, Ken. Anatomy & Physiology. 7th ed. McGraw- ...
p_(x,y) due to the identity :\ p_(x,y)=p_(y, x)\,p_X(x) which, in turn, induces a
mutual information In probability theory and information theory, the mutual information (MI) of two random variables is a measure of the mutual Statistical dependence, dependence between the two variables. More specifically, it quantifies the "Information conten ...
I(X;Y). The channel capacity is defined as :\ C = \sup_ I(X;Y)\, where the
supremum In mathematics, the infimum (abbreviated inf; : infima) of a subset S of a partially ordered set P is the greatest element in P that is less than or equal to each element of S, if such an element exists. If the infimum of S exists, it is unique, ...
is taken over all possible choices of p_X(x).


Additivity of channel capacity

Channel capacity is additive over independent channels. It means that using two independent channels in a combined manner provides the same theoretical capacity as using them independently. More formally, let p_ and p_ be two independent channels modelled as above; p_ having an input alphabet \mathcal_ and an output alphabet \mathcal_. Idem for p_. We define the product channel p_\times p_2 as \forall (x_, x_) \in (\mathcal_, \mathcal_),\;(y_, y_) \in (\mathcal_, \mathcal_),\; (p_\times p_)((y_, y_) , (x_,x_))=p_(y_, x_)p_(y_, x_) This theorem states: C(p_\times p_) = C(p_) + C(p_)


Shannon capacity of a graph

If ''G'' is an
undirected graph In discrete mathematics, particularly in graph theory, a graph is a structure consisting of a set of objects where some pairs of the objects are in some sense "related". The objects are represented by abstractions called '' vertices'' (also call ...
, it can be used to define a communications channel in which the symbols are the graph vertices, and two codewords may be confused with each other if their symbols in each position are equal or adjacent. The computational complexity of finding the Shannon capacity of such a channel remains open, but it can be upper bounded by another important graph invariant, the Lovász number.


Noisy-channel coding theorem

The
noisy-channel coding theorem In information theory, the noisy-channel coding theorem (sometimes Shannon's theorem or Shannon's limit), establishes that for any given degree of noise contamination of a communication channel, it is possible (in theory) to communicate discrete ...
states that for any error probability ε > 0 and for any transmission rate ''R'' less than the channel capacity ''C'', there is an encoding and decoding scheme transmitting data at rate ''R'' whose error probability is less than ε, for a sufficiently large block length. Also, for any rate greater than the channel capacity, the probability of error at the receiver goes to 0.5 as the block length goes to infinity.


Example application

An application of the channel capacity concept to an
additive white Gaussian noise Additive white Gaussian noise (AWGN) is a basic noise model used in information theory to mimic the effect of many random processes that occur in nature. The modifiers denote specific characteristics: * ''Additive'' because it is added to any nois ...
(AWGN) channel with ''B'' Hz bandwidth and
signal-to-noise ratio Signal-to-noise ratio (SNR or S/N) is a measure used in science and engineering that compares the level of a desired signal to the level of background noise. SNR is defined as the ratio of signal power to noise power, often expressed in deci ...
''S/N'' is the
Shannon–Hartley theorem In information theory, the Shannon–Hartley theorem tells the maximum rate at which information can be transmitted over a communications channel of a specified bandwidth in the presence of noise. It is an application of the noisy-channel coding ...
: : C = B \log_2 \left( 1+\frac \right)\ ''C'' is measured in
bits per second In telecommunications and computing, bit rate (bitrate or as a variable ''R'') is the number of bits that are conveyed or processed per unit of time. The bit rate is expressed in the unit bit per second (symbol: bit/s), often in conjunction ...
if the
logarithm In mathematics, the logarithm of a number is the exponent by which another fixed value, the base, must be raised to produce that number. For example, the logarithm of to base is , because is to the rd power: . More generally, if , the ...
is taken in base 2, or nats per second if the
natural logarithm The natural logarithm of a number is its logarithm to the base of a logarithm, base of the e (mathematical constant), mathematical constant , which is an Irrational number, irrational and Transcendental number, transcendental number approxima ...
is used, assuming ''B'' is in
hertz The hertz (symbol: Hz) is the unit of frequency in the International System of Units (SI), often described as being equivalent to one event (or Cycle per second, cycle) per second. The hertz is an SI derived unit whose formal expression in ter ...
; the signal and noise powers ''S'' and ''N'' are expressed in a linear power unit (like watts or volts2). Since ''S/N'' figures are often cited in dB, a conversion may be needed. For example, a signal-to-noise ratio of 30 dB corresponds to a linear power ratio of 10^ = 10^3 = 1000.


Channel capacity estimation

To determine the channel capacity, it is necessary to find the capacity-achieving distribution p_X(x) and evaluate the
mutual information In probability theory and information theory, the mutual information (MI) of two random variables is a measure of the mutual Statistical dependence, dependence between the two variables. More specifically, it quantifies the "Information conten ...
I(X;Y). Research has mostly focused on studying additive noise channels under certain power constraints and noise distributions, as analytical methods are not feasible in the majority of other scenarios. Hence, alternative approaches such as, investigation on the input support, relaxations and capacity bounds, have been proposed in the literature. The capacity of a discrete memoryless channel can be computed using the Blahut-Arimoto algorithm.
Deep learning Deep learning is a subset of machine learning that focuses on utilizing multilayered neural networks to perform tasks such as classification, regression, and representation learning. The field takes inspiration from biological neuroscience a ...
can be used to estimate the channel capacity. In fact, the channel capacity and the capacity-achieving distribution of any discrete-time continuous memoryless vector channel can be obtained using CORTICAL, a cooperative framework inspired by generative adversarial networks. CORTICAL consists of two cooperative networks: a generator with the objective of learning to sample from the capacity-achieving input distribution, and a discriminator with the objective to learn to distinguish between paired and unpaired channel input-output samples and estimates I(X;Y).


Channel capacity in wireless communications

This section focuses on the single-antenna, point-to-point scenario. For channel capacity in systems with multiple antennas, see the article on
MIMO In radio, multiple-input and multiple-output (MIMO) () is a method for multiplying the capacity of a radio link using multiple transmission and receiving antennas to exploit multipath propagation. MIMO has become an essential element of wirel ...
.


Bandlimited AWGN channel

If the average received power is \bar the total bandwidth is W in Hertz, and the noise power spectral density is N_0 /Hz the AWGN channel capacity is :C_=W\log_2\left(1+\frac\right) its/s where \frac is the received signal-to-noise ratio (SNR). This result is known as the Shannon–Hartley theorem. When the SNR is large (SNR ≫ 0 dB), the capacity C\approx W\log_2 \frac is logarithmic in power and approximately linear in bandwidth. This is called the ''bandwidth-limited regime''. When the SNR is small (SNR ≪ 0 dB), the capacity C\approx \frac is linear in power but insensitive to bandwidth. This is called the ''power-limited regime''. The bandwidth-limited regime and power-limited regime are illustrated in the figure.


Frequency-selective AWGN channel

The capacity of the frequency-selective channel is given by so-called water filling power allocation, :C_=\sum_^ \log_2 \left(1+\frac \right), where P_n^*=\max \left\ and , \bar_n, ^2 is the gain of subchannel n, with \lambda chosen to meet the power constraint.


Slow-fading channel

In a slow-fading channel, where the coherence time is greater than the latency requirement, there is no definite capacity as the maximum rate of reliable communications supported by the channel, \log_2 (1+, h, ^2 SNR), depends on the random channel gain , h, ^2, which is unknown to the transmitter. If the transmitter encodes data at rate R its/s/Hz there is a non-zero probability that the decoding error probability cannot be made arbitrarily small, :p_=\mathbb(\log(1+, h, ^2 SNR), in which case the system is said to be in outage. With a non-zero probability that the channel is in deep fade, the capacity of the slow-fading channel in strict sense is zero. However, it is possible to determine the largest value of R such that the outage probability p_ is less than \epsilon. This value is known as the \epsilon-outage capacity.


Fast-fading channel

In a fast-fading channel, where the latency requirement is greater than the coherence time and the codeword length spans many coherence periods, one can average over many independent channel fades by coding over a large number of coherence time intervals. Thus, it is possible to achieve a reliable rate of communication of \mathbb(\log_2 (1+, h, ^2 SNR)) its/s/Hzand it is meaningful to speak of this value as the capacity of the fast-fading channel.


Feedback Capacity

Feedback capacity is the greatest rate at which
information Information is an Abstraction, abstract concept that refers to something which has the power Communication, to inform. At the most fundamental level, it pertains to the Interpretation (philosophy), interpretation (perhaps Interpretation (log ...
can be reliably transmitted, per unit time, over a point-to-point
communication channel A communication channel refers either to a physical transmission medium such as a wire, or to a logical connection over a multiplexed medium such as a radio channel in telecommunications and computer networking. A channel is used for infor ...
in which the receiver feeds back the channel outputs to the transmitter. Information-theoretic analysis of communication systems that incorporate feedback is more complicated and challenging than without feedback. Possibly, this was the reason C.E. Shannon chose feedback as the subject of the first Shannon Lecture, delivered at the 1973 IEEE International Symposium on Information Theory in Ashkelon, Israel. The feedback capacity is characterized by the maximum of the directed information between the channel inputs and the channel outputs, where the maximization is with respect to the causal conditioning of the input given the output. The directed information was coined by James Massey in 1990, who showed that its an upper bound on feedback capacity. For memoryless channels, Shannon showed that feedback does not increase the capacity, and the feedback capacity coincides with the channel capacity characterized by the
mutual information In probability theory and information theory, the mutual information (MI) of two random variables is a measure of the mutual Statistical dependence, dependence between the two variables. More specifically, it quantifies the "Information conten ...
between the input and the output. The feedback capacity is known as a closed-form expression only for several examples such as the trapdoor channel, Ising channel,. For some other channels, it is characterized through constant-size optimization problems such as the binary erasure channel with a no-consecutive-ones input constraint, NOST channel. The basic mathematical model for a communication system is the following: Here is the formal definition of each element (where the only difference with respect to the nonfeedback capacity is the encoder definition): * W is the message to be transmitted, taken in an
alphabet An alphabet is a standard set of letter (alphabet), letters written to represent particular sounds in a spoken language. Specifically, letters largely correspond to phonemes as the smallest sound segments that can distinguish one word from a ...
\mathcal; * X is the channel input symbol (X^n is a sequence of n symbols) taken in an
alphabet An alphabet is a standard set of letter (alphabet), letters written to represent particular sounds in a spoken language. Specifically, letters largely correspond to phonemes as the smallest sound segments that can distinguish one word from a ...
\mathcal; * Y is the channel output symbol (Y^n is a sequence of n symbols) taken in an alphabet \mathcal; * \hat is the estimate of the transmitted message; * f_i: \mathcal \times \mathcal^ \to \mathcal is the encoding function at time i, for a block of length n; * p(y_i, x^i,y^) = p_(y_i, x^i,y^) is the noisy channel at time i, which is modeled by a
conditional probability distribution In probability theory and statistics, the conditional probability distribution is a probability distribution that describes the probability of an outcome given the occurrence of a particular event. Given two jointly distributed random variables X ...
; and, * \hat: \mathcal^n \to \mathcal is the decoding function for a block of length n. That is, for each time i there exists a feedback of the previous output Y_ such that the encoder has access to all previous outputs Y^ . An (2^,n) code is a pair of encoding and decoding mappings with \mathcal= ,2,\dots, 2^/math>, and W is uniformly distributed. A rate R is said to be achievable if there exists a sequence of codes (2^,n) such that the ''average probability of error:'' P_e^\triangleq \Pr (\hat\neq W) tends to zero as n\to \infty. The feedback capacity is denoted by C_, and is defined as the supremum over all achievable rates.


Main results on feedback capacity

Let X and Y be modeled as random variables. The causal conditioning P(y^n, , x^n) \triangleq \prod_^n P(y_i, y^,x^) describes the given channel. The choice of the causally conditional distribution P(x^n, , y^) \triangleq \prod_^n P(x_i, x^,y^) determines the
joint distribution A joint or articulation (or articular surface) is the connection made between bones, ossicles, or other hard structures in the body which link an animal's skeletal system into a functional whole.Saladin, Ken. Anatomy & Physiology. 7th ed. McGraw- ...
p_(x^n,y^n) due to the chain rule for causal conditioning P(y^n, x^n) = P(y^n, , x^n) P(x^n, , y^) which, in turn, induces a directed information I(X^N \rightarrow Y^N)=\mathbf E\left \log \frac \right/math>. The feedback capacity is given by : \ C_ = \lim_ \frac \sup_ I(X^n \to Y^n)\, , where the
supremum In mathematics, the infimum (abbreviated inf; : infima) of a subset S of a partially ordered set P is the greatest element in P that is less than or equal to each element of S, if such an element exists. If the infimum of S exists, it is unique, ...
is taken over all possible choices of P_(x^n, , y^).


Gaussian feedback capacity

When the Gaussian noise is colored, the channel has memory. Consider for instance the simple case on an
autoregressive model In statistics, econometrics, and signal processing, an autoregressive (AR) model is a representation of a type of random process; as such, it can be used to describe certain time-varying processes in nature, economics, behavior, etc. The autoregre ...
noise process z_i = z_+w_i where w_i\sim N(0,1) is an i.i.d. process.


Solution techniques

The feedback capacity is difficult to solve in the general case. There are some techniques that are related to control theory and Markov decision processes if the channel is discrete.


See also

*
Bandwidth (computing) In computing, bandwidth is the maximum rate of data transfer across a given path. Bandwidth may be characterized as network bandwidth, data bandwidth, or digital bandwidth. This definition of ''bandwidth'' is in contrast to the field of signal ...
*
Bandwidth (signal processing) Bandwidth is the difference between the upper and lower Frequency, frequencies in a continuous Frequency band, band of frequencies. It is typically measured in unit of measurement, unit of hertz (symbol Hz). It may refer more specifically to ...
*
Bit rate In telecommunications and computing, bit rate (bitrate or as a variable ''R'') is the number of bits that are conveyed or processed per unit of time. The bit rate is expressed in the unit bit per second (symbol: bit/s), often in conjunction ...
*
Code rate In telecommunication and information theory, the code rate (or information rateHuffman, W. Cary, and Pless, Vera, ''Fundamentals of Error-Correcting Codes'', Cambridge, 2003.) of a forward error correction code is the proportion of the data-stre ...
*
Error exponent In information theory, the error exponent of a channel code or source code over the block length of the code is the rate at which the error probability decays exponentially with the block length of the code. Formally, it is defined as the limiti ...
* Nyquist rate * Negentropy * Redundancy * Sender,
Data compression In information theory, data compression, source coding, or bit-rate reduction is the process of encoding information using fewer bits than the original representation. Any particular compression is either lossy or lossless. Lossless compressi ...
, Receiver *
Shannon–Hartley theorem In information theory, the Shannon–Hartley theorem tells the maximum rate at which information can be transmitted over a communications channel of a specified bandwidth in the presence of noise. It is an application of the noisy-channel coding ...
*
Spectral efficiency Spectral efficiency, spectrum efficiency or bandwidth efficiency refers to the information rate that can be transmitted over a given bandwidth in a specific communication system. It is a measure of how efficiently a limited frequency spectrum i ...
*
Throughput Network throughput (or just throughput, when in context) refers to the rate of message delivery over a communication channel in a communication network, such as Ethernet or packet radio. The data that these messages contain may be delivered ov ...
* Shannon capacity of a graph


Advanced Communication Topics

*
MIMO In radio, multiple-input and multiple-output (MIMO) () is a method for multiplying the capacity of a radio link using multiple transmission and receiving antennas to exploit multipath propagation. MIMO has become an essential element of wirel ...
* Cooperative diversity


External links

*
AWGN Channel Capacity with various constraints on the channel input (interactive demonstration)


References

{{More citations needed, date=January 2008 Information theory Telecommunication theory Television terminology