A binary symmetric channel (or BSC
p) is a common
communications channel model used in
coding theory
Coding theory is the study of the properties of codes and their respective fitness for specific applications. Codes are used for data compression, cryptography, error detection and correction, data transmission and computer data storage, data sto ...
and
information theory
Information theory is the mathematical study of the quantification (science), quantification, Data storage, storage, and telecommunications, communication of information. The field was established and formalized by Claude Shannon in the 1940s, ...
. In this model, a transmitter wishes to send a
bit (a zero or a one), and the receiver will receive a bit. The bit will be "flipped" with a "crossover
probability
Probability is a branch of mathematics and statistics concerning events and numerical descriptions of how likely they are to occur. The probability of an event is a number between 0 and 1; the larger the probability, the more likely an e ...
" of ''p'', and otherwise is received correctly. This model can be applied to varied communication channels such as telephone lines or
disk drive storage.
The
noisy-channel coding theorem
In information theory, the noisy-channel coding theorem (sometimes Shannon's theorem or Shannon's limit), establishes that for any given degree of noise contamination of a communication channel, it is possible (in theory) to communicate discrete ...
applies to BSC
p, saying that information can be transmitted at any rate up to the
channel capacity
Channel capacity, in electrical engineering, computer science, and information theory, is the theoretical maximum rate at which information can be reliably transmitted over a communication channel.
Following the terms of the noisy-channel coding ...
with arbitrarily low error. The channel capacity is
bits, where
is the
binary entropy function
Binary may refer to:
Science and technology Mathematics
* Binary number, a representation of numbers using only two values (0 and 1) for each digit
* Binary function, a function that takes two arguments
* Binary operation, a mathematical op ...
. Codes including Forney's code have been designed to transmit information efficiently across the channel.
Definition
A binary symmetric channel with crossover probability
, denoted by BSC
p, is a channel with binary input and binary output and probability of error
. That is, if
is the transmitted
random variable
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a Mathematics, mathematical formalization of a quantity or object which depends on randomness, random events. The term 'random variable' in its mathema ...
and
the received variable, then the channel is characterized by the
conditional probabilities:
:
It is assumed that
. If
, then the receiver can swap the output (interpret 1 when it sees 0, and vice versa) and obtain an equivalent channel with crossover probability
.
Capacity
The
channel capacity
Channel capacity, in electrical engineering, computer science, and information theory, is the theoretical maximum rate at which information can be reliably transmitted over a communication channel.
Following the terms of the noisy-channel coding ...
of the binary symmetric channel, in
bits, is:
:
where
is the
binary entropy function
Binary may refer to:
Science and technology Mathematics
* Binary number, a representation of numbers using only two values (0 and 1) for each digit
* Binary function, a function that takes two arguments
* Binary operation, a mathematical op ...
, defined by:
:
:
Noisy-channel coding theorem
Shannon's
noisy-channel coding theorem
In information theory, the noisy-channel coding theorem (sometimes Shannon's theorem or Shannon's limit), establishes that for any given degree of noise contamination of a communication channel, it is possible (in theory) to communicate discrete ...
gives a result about the rate of information that can be transmitted through a communication channel with arbitrarily low error. We study the particular case of
.
The noise
that characterizes
is a
random variable
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a Mathematics, mathematical formalization of a quantity or object which depends on randomness, random events. The term 'random variable' in its mathema ...
consisting of n independent random bits (n is defined below) where each random bit is a
with probability
and a
with probability
. We indicate this by writing "
".
What this theorem actually implies is, a message when picked from
, encoded with a random encoding function
, and sent across a noisy
, there is a very high probability of recovering the original message by decoding, if
or in effect the rate of the channel is bounded by the quantity stated in the theorem. The decoding error probability is exponentially small.
Proof
The theorem can be proved directly with a
probabilistic method. Consider an encoding function
that is selected at random. This means that for each message
, the value
is selected at random (with equal probabilities). For a given encoding function
, the decoding function
is specified as follows: given any received codeword
, we find the message
such that the
Hamming distance
In information theory, the Hamming distance between two String (computer science), strings or vectors of equal length is the number of positions at which the corresponding symbols are different. In other words, it measures the minimum number ...
is as small as possible (with ties broken arbitrarily). (
is called a
maximum likelihood decoding function.)
The proof continues by showing that at least one such choice
satisfies the conclusion of theorem, by integration over the probabilities. Suppose
and
are fixed. First we show that, for a fixed
and
chosen randomly, the probability of failure over
noise is exponentially small in ''n''. At this point, the proof works for a fixed message
. Next we extend this result to work for all messages
. We achieve this by eliminating half of the codewords from the code with the argument that the proof for the decoding error probability holds for at least half of the codewords. The latter method is called expurgation. This gives the total process the name ''random coding with expurgation''.
:
:
Converse of Shannon's capacity theorem
The converse of the capacity theorem essentially states that
is the best rate one can achieve over a binary symmetric channel. Formally the theorem states:
The intuition behind the proof is however showing the number of errors to grow rapidly as the rate grows beyond the channel capacity. The idea is the sender generates messages of dimension
, while the channel
introduces transmission errors. When the capacity of the channel is
, the number of errors is typically
for a code of block length
. The maximum number of messages is
. The output of the channel on the other hand has
possible values. If there is any confusion between any two messages, it is likely that
. Hence we would have
, a case we would like to avoid to keep the decoding error probability exponentially small.
Codes
Very recently, a lot of work has been done and is also being done to design explicit error-correcting codes to achieve the capacities of several standard communication channels. The motivation behind designing such codes is to relate the rate of the code with the fraction of errors which it can correct.
The approach behind the design of codes which meet the channel capacities of
or the
binary erasure channel have been to correct a lesser number of errors with a high probability, and to achieve the highest possible rate. Shannon's theorem gives us the best rate which could be achieved over a
, but it does not give us an idea of any explicit codes which achieve that rate. In fact such codes are typically constructed to correct only a small fraction of errors with a high probability, but achieve a very good rate. The first such code was due to George D. Forney in 1966. The code is a concatenated code by concatenating two different kinds of codes.
Forney's code
Forney constructed a
concatenated code to achieve the capacity of the noisy-channel coding theorem for
. In his code,
* The outer code
is a code of block length
and rate
over the field
, and
. Additionally, we have a
decoding algorithm
for
which can correct up to
fraction of worst case errors and runs in
time.
* The inner code
is a code of block length
, dimension
, and a rate of
. Additionally, we have a decoding algorithm
for
with a
decoding error probability of at most
over
and runs in
time.
For the outer code
, a Reed-Solomon code would have been the first code to have come in mind. However, we would see that the construction of such a code cannot be done in
polynomial time. This is why a
binary linear code is used for
.
For the inner code
we find a
linear code
In coding theory, a linear code is an error-correcting code for which any linear combination of Code word (communication), codewords is also a codeword. Linear codes are traditionally partitioned into block codes and convolutional codes, although t ...
by exhaustively searching from the
linear code
In coding theory, a linear code is an error-correcting code for which any linear combination of Code word (communication), codewords is also a codeword. Linear codes are traditionally partitioned into block codes and convolutional codes, although t ...
of block length
and dimension
, whose rate meets the capacity of
, by the noisy-channel coding theorem.
The rate
which almost meets the
capacity. We further note that the encoding and decoding of
can be done in polynomial time with respect to
. As a matter of fact, encoding
takes time
. Further, the decoding algorithm described takes time
as long as
; and
.
Decoding error probability
A natural decoding algorithm for
is to:
* Assume
* Execute
on
Note that each block of code for
is considered a symbol for
. Now since the probability of error at any index
for
is at most
and the errors in
are independent, the expected number of errors for
is at most
by linearity of expectation. Now applying
Chernoff bound, we have bound error probability of more than
errors occurring to be
. Since the outer code
can correct at most
errors, this is the
decoding error probability of
. This when expressed in asymptotic terms, gives us an error probability of
. Thus the achieved decoding error probability of
is exponentially small as the noisy-channel coding theorem.
We have given a general technique to construct
. For more detailed descriptions on
and
please read the following references. Recently a few other codes have also been constructed for achieving the capacities.
LDPC codes have been considered for this purpose for their faster decoding time.
[Richardson and Urbanke]
Applications
The binary symmetric channel can model a
disk drive used for memory storage: the channel input represents a bit being written to the disk and the output corresponds to the bit later being read. Error could arise from the magnetization flipping, background noise or the writing head making an error. Other objects which the binary symmetric channel can model include a telephone or radio communication line or
cell division
Cell division is the process by which a parent cell (biology), cell divides into two daughter cells. Cell division usually occurs as part of a larger cell cycle in which the cell grows and replicates its chromosome(s) before dividing. In eukar ...
, from which the daughter cells contain
DNA
Deoxyribonucleic acid (; DNA) is a polymer composed of two polynucleotide chains that coil around each other to form a double helix. The polymer carries genetic instructions for the development, functioning, growth and reproduction of al ...
information from their parent cell.
This channel is often used by theorists because it is one of the simplest
noisy channels to analyze. Many problems in
communication theory
Communication theory is a proposed description of communication phenomena, the relationships among them, a storyline describing these relationships, and an argument for these three elements. Communication theory provides a way of talking about a ...
can be
reduced to a BSC. Conversely, being able to transmit effectively over the BSC can give rise to solutions for more complicated channels.
See also
*
Z channel
Notes
References
*
* G. David Forney
Concatenated Codes MIT Press, Cambridge, MA, 1966.
* Venkat Guruswamy's course o
Error-Correcting Codes: Constructions and Algorithms], Autumn 2006.
* {{cite book , last=MacKay, first=David J.C. , author-link=David J. C. MacKay, url=http://www.inference.phy.cam.ac.uk/mackay/itila/book.html, title=Information Theory, Inference, and Learning Algorithms, publisher=Cambridge University Press, year=2003, isbn=0-521-64298-1
* Atri Rudra's course on Error Correcting Codes: Combinatorics, Algorithms, and Applications (Fall 2007), Lecture
91029 an
30
* Madhu Sudan's course on Algorithmic Introduction to Coding Theory (Fall 2001), Lectur
1an
2
A mathematical theory of communicationC. E Shannon, ACM SIGMOBILE Mobile Computing and Communications Review.
Modern Coding Theoryby Tom Richardson and Rudiger Urbanke., Cambridge University Press
Coding theory