HOME

TheInfoList



OR:

In the
mathematics Mathematics is an area of knowledge that includes the topics of numbers, formulas and related structures, shapes and the spaces in which they are contained, and quantities and their changes. These topics are represented in modern mathematics ...
of
coding theory Coding theory is the study of the properties of codes and their respective fitness for specific applications. Codes are used for data compression, cryptography, error detection and correction, data transmission and data storage. Codes are stud ...
, the Plotkin bound, named after Morris Plotkin, is a limit (or bound) on the maximum possible number of codewords in
binary Binary may refer to: Science and technology Mathematics * Binary number, a representation of numbers using only two digits (0 and 1) * Binary function, a function that takes two arguments * Binary operation, a mathematical operation that t ...
code In communications and information processing, code is a system of rules to convert information—such as a letter, word, sound, image, or gesture—into another form, sometimes shortened or secret, for communication through a communication ...
s of given length ''n'' and given minimum distance ''d''.


Statement of the bound

A code is considered "binary" if the codewords use symbols from the binary
alphabet An alphabet is a standardized set of basic written graphemes (called letters) that represent the phonemes of certain spoken languages. Not all writing systems represent language in this way; in a syllabary, each character represents a syll ...
\. In particular, if all codewords have a fixed length ''n'', then the binary code has length ''n''. Equivalently, in this case the codewords can be considered elements of
vector space In mathematics and physics, a vector space (also called a linear space) is a set whose elements, often called ''vectors'', may be added together and multiplied ("scaled") by numbers called '' scalars''. Scalars are often real numbers, but can ...
\mathbb_2^n over the
finite field In mathematics, a finite field or Galois field (so-named in honor of Évariste Galois) is a field that contains a finite number of elements. As with any field, a finite field is a set on which the operations of multiplication, addition, subtr ...
\mathbb_2. Let d be the minimum distance of C, i.e. :d = \min_ d(x,y) where d(x,y) is the
Hamming distance In information theory, the Hamming distance between two strings of equal length is the number of positions at which the corresponding symbols are different. In other words, it measures the minimum number of ''substitutions'' required to chan ...
between x and y. The expression A_(n,d) represents the maximum number of possible codewords in a binary code of length n and minimum distance d. The Plotkin bound places a limit on this expression. Theorem (Plotkin bound): i) If d is even and 2d > n , then : A_(n,d) \leq 2 \left\lfloor\frac\right\rfloor. ii) If d is odd and 2d+1 > n , then : A_(n,d) \leq 2 \left\lfloor\frac\right\rfloor. iii) If d is even, then : A_(2d,d) \leq 4d. iv) If d is odd, then : A_(2d+1,d) \leq 4d+4 where \left\lfloor ~ \right\rfloor denotes the
floor function In mathematics and computer science, the floor function is the function that takes as input a real number , and gives as output the greatest integer less than or equal to , denoted or . Similarly, the ceiling function maps to the least int ...
.


Proof of case i

Let d(x,y) be the
Hamming distance In information theory, the Hamming distance between two strings of equal length is the number of positions at which the corresponding symbols are different. In other words, it measures the minimum number of ''substitutions'' required to chan ...
of x and y, and M be the number of elements in C (thus, M is equal to A_(n,d)). The bound is proved by bounding the quantity \sum_ d(x,y) in two different ways. On the one hand, there are M choices for x and for each such choice, there are M-1 choices for y. Since by definition d(x,y) \geq d for all x and y ( x\neq y ), it follows that : \sum_ d(x,y) \geq M(M-1) d. On the other hand, let A be an M \times n matrix whose rows are the elements of C. Let s_i be the number of zeros contained in the i'th column of A. This means that the i'th column contains M-s_i ones. Each choice of a zero and a one in the same column contributes exactly 2 (because d(x,y)=d(y,x)) to the sum \sum_ d(x,y) and therefore : \sum_ d(x,y) = \sum_^n 2s_i (M-s_i). The quantity on the right is maximized if and only if s_i = M/2 holds for all i (at this point of the proof we ignore the fact, that the s_i are integers), then : \sum_ d(x,y) \leq \frac n M^2. Combining the upper and lower bounds for \sum_ d(x,y) that we have just derived, : M(M-1) d \leq \frac n M^2 which given that 2d>n is equivalent to : M \leq \frac. Since M is even, it follows that : M \leq 2 \left\lfloor \frac \right\rfloor. This completes the proof of the bound.


See also

*
Singleton bound In coding theory, the Singleton bound, named after Richard Collom Singleton, is a relatively crude upper bound on the size of an arbitrary block code C with block length n, size M and minimum distance d. It is also known as the Joshibound. proved b ...
*
Hamming bound In mathematics and computer science, in the field of coding theory, the Hamming bound is a limit on the parameters of an arbitrary block code: it is also known as the sphere-packing bound or the volume bound from an interpretation in terms of pack ...
* Elias-Bassalygo bound * Gilbert-Varshamov bound *
Johnson bound In applied mathematics, the Johnson bound (named after Selmer Martin Johnson) is a limit on the size of error-correcting codes, as used in coding theory for data transmission or communications. Definition Let C be a ''q''-ary code of length n, i ...
*
Griesmer bound In the mathematics of coding theory, the Griesmer bound, named after James Hugo Griesmer, is a bound on the length of linear binary codes of dimension ''k'' and minimum distance ''d''. There is also a very similar version for non-binary codes. Sta ...
* Diamond code


References

* {{cite journal , title=Binary codes with specified minimum distance , author-first=Morris , author-last=Plotkin , journal=
IRE Transactions on Information Theory ''IEEE Transactions on Information Theory'' is a monthly peer-reviewed scientific journal published by the IEEE Information Theory Society. It covers information theory and the mathematics of communications. It was established in 1953 as ''IRE Tran ...
, volume=6 , pages=445–450 , date=1960 , issue=4 , doi=10.1109/TIT.1960.1057584 Coding theory Articles containing proofs