Shannon Capacity Of A Graph
   HOME

TheInfoList



OR:

In
graph theory In mathematics, graph theory is the study of ''graphs'', which are mathematical structures used to model pairwise relations between objects. A graph in this context is made up of '' vertices'' (also called ''nodes'' or ''points'') which are conne ...
, the Shannon capacity of a graph is a
graph invariant Graph may refer to: Mathematics *Graph (discrete mathematics), a structure made of vertices and edges **Graph theory, the study of such graphs and their properties * Graph (topology), a topological space resembling a graph in the sense of discr ...
defined from the number of independent sets of
strong graph product In graph theory, the strong product is a way of combining two graphs to make a larger graph. Two vertices are adjacent in the strong product when they come from pairs of vertices in the factor graphs that are either adjacent or identical. The str ...
s. It is named after American mathematician
Claude Shannon Claude Elwood Shannon (April 30, 1916 – February 24, 2001) was an American people, American mathematician, electrical engineering, electrical engineer, and cryptography, cryptographer known as a "father of information theory". As a 21-year-o ...
. It measures the Shannon capacity of a
communications channel A communication channel refers either to a physical transmission medium such as a wire, or to a logical connection over a multiplexed medium such as a radio channel in telecommunications and computer networking. A channel is used for informat ...
defined from the graph, and is upper bounded by the
Lovász number In graph theory, the Lovász number of a graph is a real number that is an upper bound on the Shannon capacity of the graph. It is also known as Lovász theta function and is commonly denoted by \vartheta(G), using a script form of the Greek letter ...
, which can be computed in
polynomial time In computer science, the time complexity is the computational complexity that describes the amount of computer time it takes to run an algorithm. Time complexity is commonly estimated by counting the number of elementary operations performed by ...
. However, the
computational complexity In computer science, the computational complexity or simply complexity of an algorithm is the amount of resources required to run it. Particular focus is given to computation time (generally measured by the number of needed elementary operations) ...
of the Shannon capacity itself remains unknown.


Graph models of communication channels

The Shannon capacity models the amount of information that can be transmitted across a noisy communication channel in which certain signal values can be confused with each other. In this application, the confusion graph or confusability graph describes the pairs of values that can be confused. For instance, suppose that a communications channel has five discrete signal values, any one of which can be transmitted in a single time step. These values may be modeled mathematically as the five numbers 0, 1, 2, 3, or 4 in
modular arithmetic In mathematics, modular arithmetic is a system of arithmetic for integers, where numbers "wrap around" when reaching a certain value, called the modulus. The modern approach to modular arithmetic was developed by Carl Friedrich Gauss in his book ...
modulo 5. However, suppose that when a value i is sent across the channel, the value that is received is i+\xi (mod 5) where \xi represents the noise on the channel and may be any real number in the open interval from −1 to 1. Thus, if the recipient receives a value such as 3.6, it is impossible to determine whether it was originally transmitted as a 3 or as a 4; the two values 3 and 4 can be confused with each other. This situation can be modeled by a graph, a
cycle Cycle, cycles, or cyclic may refer to: Anthropology and social sciences * Cyclic history, a theory of history * Cyclical theory, a theory of American political history associated with Arthur Schlesinger, Sr. * Social cycle, various cycles in soc ...
C_5 of length 5, in which the vertices correspond to the five values that can be transmitted and the edges of the graph represent values that can be confused with each other. For this example, it is possible to choose two values that can be transmitted in each time step without ambiguity, for instance, the values 1 and 3. These values are far enough apart that they can't be confused with each other: when the recipient receives a value x between 0 and 2, it can deduce that the value that was sent must have been 1, and when the recipient receives a value x in between 2 and 4, it can deduce that the value that was sent must have been 3. In this way, in n steps of communication, the sender can communicate up to 2^n different messages. Two is the maximum number of values that the recipient can distinguish from each other: every subset of three or more of the values 0, 1, 2, 3, 4 includes at least one pair that can be confused with each other. Even though the channel has five values that can be sent per time step, effectively only two of them can be used with this coding scheme. However, more complicated coding schemes allow a greater amount of information to be sent across the same channel, by using codewords of length greater than one. For instance, suppose that in two consecutive steps the sender transmits one of the five
code word In communication, a code word is an element of a standardized code or protocol. Each code word is assembled in accordance with the specific rules of the code and assigned a unique meaning. Code words are typically used for reasons of reliability, ...
s "11", "23", "35", "54", or "42". (Here, the quotation marks indicate that these words should be interpreted as
strings String or strings may refer to: *String (structure), a long flexible structure made from threads twisted together, which is used to tie, bind, or hang other objects Arts, entertainment, and media Films * ''Strings'' (1991 film), a Canadian anim ...
of symbols, not as decimal numbers.) Each pair of these code words includes at least one position where its values differ by two or more modulo 5; for instance, "11" and "23" differ by two in their second position, while "23" and "42" differ by two in their first position. Therefore, a recipient of one of these code words will always be able to determine unambiguously which one was sent: no two of these code words can be confused with each other. By using this method, in n steps of communication, the sender can communicate up to 5^ messages, significantly more than the 2^n that could be transmitted with the simpler one-digit code. The effective number of values that can be transmitted per unit time step is (5^)^=\sqrt5. In graph-theoretic terms, this means that the Shannon capacity of the 5-cycle is at least \sqrt5. As showed, this bound is tight: it is not possible to find a more complicated system of code words that allows even more different messages to be sent in the same amount of time, so the Shannon capacity of the 5-cycle is


Relation to independent sets

If a graph G represents a set of symbols and the pairs of symbols that can be confused with each other, then a subset S of symbols avoids all confusable pairs if and only if S is an independent set in the graph, a subset of vertices that does not include both endpoints of any edge. The maximum possible size of a subset of the symbols that can all be distinguished from each other is the
independence number Independence is a condition of a person, nation, country, or Sovereign state, state in which residents and population, or some portion thereof, exercise self-government, and usually sovereignty, over its territory. The opposite of independ ...
\alpha(G) of the graph, the size of its
maximum independent set In graph theory, an independent set, stable set, coclique or anticlique is a set of vertices in a graph, no two of which are adjacent. That is, it is a set S of vertices such that for every two vertices in S, there is no edge connecting the two ...
. For instance, '\alpha(C_5)=2: the 5-cycle has independent sets of two vertices, but not larger. For codewords of longer lengths, one can use independent sets in larger graphs to describe the sets of codewords that can be transmitted without confusion. For instance, for the same example of five symbols whose confusion graph is C_5, there are 25 strings of length two that can be used in a length-2 coding scheme. These strings may be represented by the vertices of a graph with 25 vertices. In this graph, each vertex has eight neighbors, the eight strings that it can be confused with. A subset of length-two strings forms a code with no possible confusion if and only if it corresponds to an independent set of this graph. The set of code words forms one of these independent sets, of maximum size. If G is a graph representing the signals and confusable pairs of a channel, then the graph representing the length-two codewords and their confusable pairs is G \boxtimes G, where the symbol \boxtimes represents the
strong product of graphs In graph theory, the strong product is a way of combining two graphs to make a larger graph. Two vertices are adjacent in the strong product when they come from pairs of vertices in the factor graphs that are either adjacent or identical. The str ...
. This is a graph that has a vertex for each pair (u,v) of a vertex in the first argument of the product and a vertex in the second argument of the product. Two distinct pairs (u_1,v_1) and (u_2,v_2) are adjacent in the strong product if and only if u_1 and u_2 are identical or adjacent, and v_1 and v_2 are identical or adjacent. More generally, the codewords of length k can be represented by the graph G_k, the k-fold strong product of G with itself, and the maximum number of codewords of this length that can be transmitted without confusion is given by the independence number \alpha(G_k). The effective number of signals transmitted per unit time step is the kth root of this number, \alpha(G_k)^. Using these concepts, the Shannon capacity may be defined as : \Theta(G) = \sup_k \sqrt = \lim_ \sqrt the limit (as k becomes arbitrarily large) of the effective number of signals per time step of arbitrarily long confusion-free codes.


Computational complexity

The
computational complexity In computer science, the computational complexity or simply complexity of an algorithm is the amount of resources required to run it. Particular focus is given to computation time (generally measured by the number of needed elementary operations) ...
of the Shannon capacity is unknown, and even the value of the Shannon capacity for certain small graphs such as C_7 (a
cycle graph In graph theory, a cycle graph or circular graph is a graph that consists of a single cycle, or in other words, some number of vertices (at least 3, if the graph is simple) connected in a closed chain. The cycle graph with vertices is called ...
of seven vertices) remains unknown. A natural approach to this problem would be to compute a finite number of powers of the given graph G , find their independence numbers, and infer from these numbers some information about the limiting behavior of the sequence from which the Shannon capacity is defined. However (even ignoring the computational difficulty of computing the independence numbers of these graphs, an
NP-hard In computational complexity theory, NP-hardness ( non-deterministic polynomial-time hardness) is the defining property of a class of problems that are informally "at least as hard as the hardest problems in NP". A simple example of an NP-hard pr ...
problem) the unpredictable behavior of the sequence of independence numbers of powers of G implies that this approach cannot be used to accurately approximate the Shannon capacity.


Upper bounds

In part because the Shannon capacity is difficult to compute, researchers have looked for other graph invariants that are easy to compute and that provide bounds on the Shannon capacity.


Lovász number

The
Lovász number In graph theory, the Lovász number of a graph is a real number that is an upper bound on the Shannon capacity of the graph. It is also known as Lovász theta function and is commonly denoted by \vartheta(G), using a script form of the Greek letter ...
(''G'') is a different graph invariant, that can be computed numerically to high accuracy in
polynomial time In computer science, the time complexity is the computational complexity that describes the amount of computer time it takes to run an algorithm. Time complexity is commonly estimated by counting the number of elementary operations performed by ...
by an algorithm based on the
ellipsoid method In mathematical optimization, the ellipsoid method is an iterative method for convex optimization, minimizing convex functions. When specialized to solving feasible linear optimization problems with rational data, the ellipsoid method is an algor ...
. The Shannon capacity of a graph ''G'' is bounded from below by α(''G''), and from above by (''G'').. In some cases, (''G'') and the Shannon capacity coincide; for instance, for the graph of a
pentagon In geometry, a pentagon (from the Greek πέντε ''pente'' meaning ''five'' and γωνία ''gonia'' meaning ''angle'') is any five-sided polygon or 5-gon. The sum of the internal angles in a simple pentagon is 540°. A pentagon may be simpl ...
, both are equal to . However, there exist other graphs for which the Shannon capacity and the Lovász number differ..


Haemers' bound

Haemers provided another upper bound on the Shannon capacity, which is sometimes better than Lovász bound:. The definition here corrects a typo in this paper. : \Theta(G) \leq R(G) = \min_ \operatorname(B), where ''B'' is an ''n'' × ''n'' matrix over some
field Field may refer to: Expanses of open ground * Field (agriculture), an area of land used for agricultural purposes * Airfield, an aerodrome that lacks the infrastructure of an airport * Battlefield * Lawn, an area of mowed grass * Meadow, a grass ...
, such that ''bii'' ≠ 0 and ''bij'' = 0 if vertices ''i'' and ''j'' are not adjacent.


References

{{reflist Graph invariants Information theory