HOME

TheInfoList



OR:

In
theoretical computer science Theoretical computer science is a subfield of computer science and mathematics that focuses on the Abstraction, abstract and mathematical foundations of computation. It is difficult to circumscribe the theoretical areas precisely. The Associati ...
, communication complexity studies the amount of communication required to solve a problem when the input to the problem is distributed among two or more parties. The study of communication complexity was first introduced by Andrew Yao in 1979, while studying the problem of computation distributed among several machines. The problem is usually stated as follows: two parties (traditionally called
Alice and Bob Alice and Bob are fictional characters commonly used as placeholders in discussions about cryptography, cryptographic systems and Cryptographic protocol, protocols, and in other science and engineering literature where there are several partici ...
) each receive a (potentially different) n- bit string x and y. The goal is for Alice to compute the value of a certain function, f(x, y), that depends on both x and y, with the least amount of
communication Communication is commonly defined as the transmission of information. Its precise definition is disputed and there are disagreements about whether Intention, unintentional or failed transmissions are included and whether communication not onl ...
between them. While Alice and Bob can always succeed by having Bob send his whole n-bit string to Alice (who then computes the function f), the idea here is to find clever ways of calculating ''f'' with fewer than n bits of communication. Note that, unlike in
computational complexity theory In theoretical computer science and mathematics, computational complexity theory focuses on classifying computational problems according to their resource usage, and explores the relationships between these classifications. A computational problem ...
, communication complexity is not concerned with the amount of computation performed by Alice or Bob, or the size of the
memory Memory is the faculty of the mind by which data or information is encoded, stored, and retrieved when needed. It is the retention of information over time for the purpose of influencing future action. If past events could not be remembe ...
used, as we generally assume nothing about the computational power of either Alice or Bob. This abstract problem with two parties (called two-party communication complexity), and its general form with more than two parties, is relevant in many contexts. In VLSI circuit design, for example, one seeks to minimize energy used by decreasing the amount of electric signals passed between the different components during a distributed computation. The problem is also relevant in the study of data structures and in the optimization of computer networks. For surveys of the field, see the textbooks by and .


Formal definition

Let f: X \times Y \rightarrow Z where we assume in the typical case that X=Y=\^n and Z=\. Alice holds an n-bit string x \in X while Bob holds an n-bit string y \in Y. By communicating to each other one bit at a time (adopting some ''communication protocol'' which is agreed upon in advance), Alice and Bob wish to compute the value of f(x,y) such that at least one party knows the value at the end of the communication. At this point the answer can be communicated back so that at the cost of one extra bit, both parties will know the answer. The worst case communication complexity of this communication problem of computing f, denoted as D(f) , is then defined to be : D(f) = minimum number of bits exchanged between Alice and Bob in the worst case. As observed above, for any function f: \^n \times \^n \rightarrow \, we have D(f) \leq n. Using the above definition, it is useful to think of the function f as a
matrix Matrix (: matrices or matrixes) or MATRIX may refer to: Science and mathematics * Matrix (mathematics), a rectangular array of numbers, symbols or expressions * Matrix (logic), part of a formula in prenex normal form * Matrix (biology), the m ...
A (called the ''input matrix'' or ''communication matrix'') where the rows are indexed by x \in X and columns by y \in Y. The entries of the matrix are A_=f(x,y). Initially both Alice and Bob have a copy of the entire matrix A (assuming the function f is known to both parties). Then, the problem of computing the function value can be rephrased as "zeroing-in" on the corresponding matrix entry. This problem can be solved if either Alice or Bob knows both x and y. At the start of communication, the number of choices for the value of the function on the inputs is the size of matrix, i.e. 2^. Then, as and when each party communicates a bit to the other, the number of choices for the answer reduces as this eliminates a set of rows/columns resulting in a
submatrix In mathematics, a matrix (: matrices) is a rectangular array or table of numbers, symbols, or expressions, with elements or entries arranged in rows and columns, which is used to represent a mathematical object or property of such an object. ...
of A. More formally, a set R \subseteq X \times Y is called a ''(combinatorial) rectangle'' if whenever (x_1,y_1) \in R and (x_2,y_2) \in R then (x_1,y_2) \in R. Equivalently, R is a combinatorial rectangle if it can be expressed as R = M \times N for some M \subseteq X and N \subseteq Y. Consider the case when k bits are already exchanged between the parties. Now, for a particular h \in \^k, let us define a matrix :T_ = \ Then, T_ \subseteq X \times Y, and it is not hard to show that T_ is a combinatorial rectangle in A.


Example: ''EQ''

We consider the case where Alice and Bob try to determine whether or not their input strings are equal. Formally, define the ''Equality'' function, denoted EQ : \^n \times \^n \rightarrow \, by EQ(x, y) = 1 if x = y. As we demonstrate below, any deterministic communication protocol solving EQ requires n bits of communication in the worst case. As a warm-up example, consider the simple case of x, y \in \^3. The equality function in this case can be represented by the matrix below. The rows represent all the possibilities of x, the columns those of y. In this table, the function only evaluates to 1 when x equals y (i.e., on the diagonal). It is also fairly easy to see how communicating a single bit divides someone's possibilities in half. When the first bit of y is 1, consider only half of the columns (where y can equal 100, 101, 110, or 111).


Theorem: ''D(EQ) = n''

Proof. Assume that D(EQ) \leq n-1. This means that there exists x \neq x' such that (x, x) and (x', x') have the same communication transcript h. Since this transcript defines a rectangle, f(x, x') must also be 1. By definition x \neq x' and we know that equality is only true for (a, b) when a = b. This yields a contradiction. This technique of proving deterministic communication lower bounds is called the ''fooling set'' technique.


Randomized communication complexity

In the above definition, we are concerned with the number of bits that must be ''deterministically'' transmitted between two parties. If both the parties are given access to a random number generator, can they determine the value of f with much less information exchanged? Yao, in his seminal paper answers this question by defining randomized communication complexity. A randomized protocol R for a function f has two-sided error. : \Pr (x,y) = 0> \frac, \textrm\, f(x,y) = 0 : \Pr (x,y) = 1> \frac, \textrm\, f(x,y) = 1 A randomized protocol is a deterministic protocol that uses an extra random string in addition to its normal input. There are two models for this: a ''public string'' is a random string that is known by both parties beforehand, while a ''private string'' is generated by one party and must be communicated to the other party. A theorem presented below shows that any public string protocol can be simulated by a private string protocol that uses ''O(log n)'' additional bits compared to the original. In the probability inequalities above, the outcome of the protocol is understood to depend ''only'' on the random string; both strings ''x'' and ''y'' remain fixed. In other words, if ''R''(''x'',''y'') yields ''g''(''x'',''y'',''r'') when using random string ''r'', then ''g''(''x'',''y'',''r'') = ''f''(''x'',''y'') for at least 2/3 of all choices for the string ''r''. The randomized complexity is simply defined as the number of bits exchanged in such a protocol. Note that it is also possible to define a randomized protocol with one-sided error, and the complexity is defined similarly.


Example: EQ

Returning to the previous example of ''EQ'', if certainty is not required, Alice and Bob can check for equality using only messages. Consider the following protocol: Assume that Alice and Bob both have access to the same random string z \in \^n. Alice computes z \cdot x and sends this bit (call it ''b'') to Bob. (The (\cdot) is the
dot product In mathematics, the dot product or scalar productThe term ''scalar product'' means literally "product with a Scalar (mathematics), scalar as a result". It is also used for other symmetric bilinear forms, for example in a pseudo-Euclidean space. N ...
in
GF(2) (also denoted \mathbb F_2, or \mathbb Z/2\mathbb Z) is the finite field with two elements. is the Field (mathematics), field with the smallest possible number of elements, and is unique if the additive identity and the multiplicative identity ...
.) Then Bob compares ''b'' to z \cdot y. If they are the same, then Bob accepts, saying ''x'' equals ''y''. Otherwise, he rejects. Clearly, if x = y, then z \cdot x = z \cdot y, so Prob_z ccept= 1. If ''x'' does not equal ''y'', it is still possible that z \cdot x = z \cdot y, which would give Bob the wrong answer. How does this happen? If ''x'' and ''y'' are not equal, they must differ in some locations: :\begin x = c_1 c_2 \ldots p \ldots p' \ldots x_n \\ y = c_1 c_2 \ldots q \ldots q' \ldots y_n \\ z = z_1 z_2 \ldots z_i \ldots z_j \ldots z_n \end Where and agree, z_i * x_i = z_i * c_i = z_i * y_i so those terms affect the dot products equally. We can safely ignore those terms and look only at where and differ. Furthermore, we can swap the bits x_i and y_i without changing whether or not the dot products are equal. This means we can swap bits so that contains only zeros and contains only ones: :\begin x' = 0 0 \ldots 0 \\ y' = 1 1 \ldots 1 \\ z' = z_1 z_2 \ldots z_ \end Note that z' \cdot x' = 0 and z' \cdot y' = \Sigma_i z'_i. Now, the question becomes: for some random string z', what is the probability that \Sigma_i z'_i = 0? Since each z'_i is equally likely to be or , this probability is just 1/2. Thus, when does not equal , Prob_z ccept= 1/2. The algorithm can be repeated many times to increase its accuracy. This fits the requirements for a randomized communication algorithm. This shows that ''if Alice and Bob share a random string of length n'', they can send one bit to each other to compute EQ(x,y). In the next section, it is shown that Alice and Bob can exchange only bits that are as good as sharing a random string of length ''n''. Once that is shown, it follows that ''EQ'' can be computed in messages.


Example: GH

For yet another example of randomized communication complexity, we turn to an example known as the '' gap-Hamming problem'' (abbreviated ''GH''). Formally, Alice and Bob both maintain binary messages, x,y \in \^n and would like to determine if the strings are very similar or if they are not very similar. In particular, they would like to find a communication protocol requiring the transmission of as few bits as possible to compute the following partial Boolean function, : \text_n(x, y) := \begin -1 & \langle x, y \rangle \leq \sqrt \\ +1 & \langle x, y \rangle \geq \sqrt. \end Clearly, they must communicate all their bits if the protocol is to be deterministic (this is because, if there is a deterministic, strict subset of indices that Alice and Bob relay to one another, then imagine having a pair of strings that on that set disagree in \sqrt - 1 positions. If another disagreement occurs in any position that is not relayed, then this affects the result of \text_n(x, y), and hence would result in an incorrect procedure. A natural question one then asks is, if we're permitted to err 1/3 of the time (over random instances x, y drawn uniformly at random from \^n ), then can we get away with a protocol with fewer bits? It turns out that the answer somewhat surprisingly is no, due to a result of Chakrabarti and Regev in 2012: they show that for random instances, any procedure which is correct at least 2/3 of the time must send \Omega(n) bits worth of communication, which is to say essentially all of them.


Public coins versus private coins

Creating random protocols becomes easier when both parties have access to the same random string, known as a shared string protocol. However, even in cases where the two parties do not share a random string, it is still possible to use private string protocols with only a small communication cost. Any shared string random protocol using any number of random string can be simulated by a private string protocol that uses an extra ''O(log n)'' bits. Intuitively, we can find some set of strings that has enough randomness in it to run the random protocol with only a small increase in error. This set can be shared beforehand, and instead of drawing a random string, Alice and Bob need only agree on which string to choose from the shared set. This set is small enough that the choice can be communicated efficiently. A formal proof follows. Consider some random protocol ''P'' with a maximum error rate of 0.1. Let R be 100n strings of length ''n'', numbered r_1, r_2, \dots, r_. Given such an R, define a new protocol P'_R which randomly picks some r_i and then runs ''P'' using r_i as the shared random string. It takes ''O''(log 100''n'') = ''O''(log ''n'') bits to communicate the choice of r_i. Let us define p(x,y) and p'_R(x,y) to be the probabilities that P and P'_R compute the correct value for the input (x,y). For a fixed (x,y), we can use Hoeffding's inequality to get the following equation: :\Pr_R exists (x,y):\ , p'_R(x,y) - p(x,y), \geq 0.1\leq \sum_ \Pr_R R, only one bit of classical communication is enough for Alice to know the evaluation f(x,y) in the worst case scenario for any
Boolean function In mathematics, a Boolean function is a function whose arguments and result assume values from a two-element set (usually , or ). Alternative names are switching function, used especially in older computer science literature, and truth functi ...
f. '' The surprising fact of a collapse of communication complexity is that the function f can have arbitrarily large entry size, but still the number of communication bit is constant to a single one. Some resources are shown to be non-collapsing, such as quantum correlations or more generally almost-quantum correlations, whereas on the contrary some other resources are shown to collapse randomized communication complexity, such as the PR-box, or some noisy PR-boxes satisfying some conditions.


Distributional Complexity

One approach to studying randomized communication complexity is through distributional complexity. Given a joint distribution \mu on the inputs of both players, the corresponding distributional complexity of a function f is the minimum cost of a ''deterministic'' protocol R such that \Pr (x,y) = R(x,y)\ge 2/3, where the inputs are sampled according to \mu. Yao's minimax principle (a special case of von Neumann's
minimax theorem In the mathematical area of game theory and of convex optimization, a minimax theorem is a theorem that claims that : \max_ \min_ f(x,y) = \min_ \max_f(x,y) under certain conditions on the sets X and Y and on the function f. It is always true that ...
) states that the randomized communication complexity of a function equals its maximum distributional complexity, where the maximum is taken over all joint distributions of the inputs (not necessarily product distributions!). Yao's principle can be used to prove lower bounds on the randomized communication complexity of a function: design the appropriate joint distribution, and prove a lower bound on the distributional complexity. Since distributional complexity concerns deterministic protocols, this could be easier than proving a lower bound on randomized protocols directly. As an example, let us consider the ''disjointness'' function DISJ: each of the inputs is interpreted as a subset of \, and DISJ(,)=1 if the two sets are disjoint. Razborov proved an \Omega(n) lower bound on the randomized communication complexity by considering the following distribution: with probability 3/4, sample two random disjoint sets of size n/4, and with probability 1/4, sample two random sets of size n/4 with a unique intersection.


Information Complexity

A powerful approach to the study of distributional complexity is information complexity. Initiated by Bar-Yossef, Jayram, Kumar and Sivakumar, the approach was codified in work of Barak, Braverman, Chen and Rao and by Braverman and Rao. The (internal) information complexity of a (possibly randomized) protocol with respect to a distribution is defined as follows. Let (X,Y) \sim \mu be random inputs sampled according to , and let be the transcript of when run on the inputs X,Y. The information complexity of the protocol is : \operatorname_\mu(R) = I(\Pi;Y, X) + I(\Pi;X, Y), where denotes conditional mutual information. The first summand measures the amount of information that Alice learns about Bob's input from the transcript, and the second measures the amount of information that Bob learns about Alice's input. The -error information complexity of a function with respect to a distribution is the infimal information complexity of a protocol for whose error (with respect to ) is at most . Braverman and Rao proved that information equals amortized communication. This means that the cost for solving independent copies of is roughly times the information complexity of . This is analogous to the well-known interpretation of
Shannon entropy Shannon may refer to: People * Shannon (given name) * Shannon (surname) * Shannon (American singer), stage name of singer Brenda Shannon Greene (born 1958) * Shannon (South Korean singer), British-South Korean singer and actress Shannon Arrum ...
as the amortized bit-length required to transmit data from a given information source. Braverman and Rao's proof uses a technique known as "protocol compression", in which an information-efficient protocol is "compressed" into a communication-efficient protocol. The techniques of information complexity enable the computation of the exact (up to first order) communication complexity of set disjointness to be 1.4923\ldots n. Information complexity techniques have also been used to analyze extended formulations, proving an essentially optimal lower bound on the complexity of algorithms based on
linear programming Linear programming (LP), also called linear optimization, is a method to achieve the best outcome (such as maximum profit or lowest cost) in a mathematical model whose requirements and objective are represented by linear function#As a polynomia ...
which approximately solve the maximum clique problem. Omri Weinstein's 2015 survey surveys the subject.


Quantum communication complexity

Quantum communication complexity tries to quantify the communication reduction possible by using quantum effects during a distributed computation. At least three quantum generalizations of communication complexity have been proposed; for a survey see the suggested text by G. Brassard. The first one is the qubit-communication model, where the parties can use quantum communication instead of classical communication, for example by exchanging
photons A photon () is an elementary particle that is a quantum of the electromagnetic field, including electromagnetic radiation such as light and radio waves, and the force carrier for the electromagnetic force. Photons are massless particles that ...
through an
optical fiber An optical fiber, or optical fibre, is a flexible glass or plastic fiber that can transmit light from one end to the other. Such fibers find wide usage in fiber-optic communications, where they permit transmission over longer distances and at ...
. In a second model the communication is still performed with classical bits, but the parties are allowed to manipulate an unlimited supply of quantum entangled states as part of their protocols. By doing measurements on their entangled states, the parties can save on classical communication during a distributed computation (see an application in Collapse of Randomized Communication Complexity). The third model involves access to previously shared entanglement in addition to
qubit In quantum computing, a qubit () or quantum bit is a basic unit of quantum information—the quantum version of the classic binary bit physically realized with a two-state device. A qubit is a two-state (or two-level) quantum-mechanical syste ...
communication, and is the least explored of the three quantum models.


Nondeterministic communication complexity

In nondeterministic communication complexity, Alice and Bob have access to an oracle. After receiving the oracle's word, the parties communicate to deduce f(x,y). The nondeterministic communication complexity is then the maximum over all pairs (x,y) over the sum of number of bits exchanged and the coding length of the oracle word. Viewed differently, this amounts to covering all 1-entries of the 0/1-matrix by combinatorial 1-rectangles (i.e., non-contiguous, non-convex submatrices, whose entries are all one (see Kushilevitz and Nisan or Dietzfelbinger et al.)). The nondeterministic communication complexity is the binary logarithm of the ''rectangle covering number'' of the matrix: the minimum number of combinatorial 1-rectangles required to cover all 1-entries of the matrix, without covering any 0-entries. Nondeterministic communication complexity occurs as a means to obtaining lower bounds for deterministic communication complexity (see Dietzfelbinger et al.), but also in the theory of nonnegative matrices, where it gives a lower bound on the nonnegative rank of a nonnegative matrix.


Unbounded-error communication complexity

In the unbounded-error setting, Alice and Bob have access to a private coin and their own inputs (x, y). In this setting, Alice succeeds if she responds with the correct value of f(x, y) with probability strictly greater than 1/2. In other words, if Alice's responses have ''any'' non-zero correlation to the true value of f(x, y), then the protocol is considered valid. Note that the requirement that the coin is ''private'' is essential. In particular, if the number of public bits shared between Alice and Bob are not counted against the communication complexity, it is easy to argue that computing any function has O(1) communication complexity. On the other hand, both models are equivalent if the number of public bits used by Alice and Bob is counted against the protocol's total communication. Though subtle, lower bounds on this model are extremely strong. More specifically, it is clear that any bound on problems of this class immediately imply equivalent bounds on problems in the deterministic model and the private and public coin models, but such bounds also hold immediately for nondeterministic communication models and quantum communication models. Forster was the first to prove explicit lower bounds for this class, showing that computing the inner product \langle x, y \rangle requires at least \Omega(n) bits of communication, though an earlier result of Alon, Frankl, and Rödl proved that the communication complexity for almost all Boolean functions f: \^n \times \^n \to \ is \Omega(n).


Lifting

Lifting is a general technique in complexity theory in which a lower bound on a simple measure of complexity is "lifted" to a lower bound on a more difficult measure. This technique was pioneered in the context of communication complexity by Raz and McKenzie, who proved the first query-to-communication lifting theorem, and used the result to separate the monotone NC hierarchy. Given a function f\colon \^n \to \ and a gadget g\colon \^a \times \^b \to \, their composition f \circ g\colon \^ \times \^ \to \ is defined as follows: : (f \circ g)(x,y) = f(g(x_ \cdots x_, y_ \cdots y_), \dots, g(x_ \cdots x_, y_ \cdots y_)). In words, x is partitioned into n blocks of length a, and y is partitioned into n blocks of length b. The gadget is applied n times on the blocks, and the outputs are fed into f. Diagrammatically: In this diagram, each of the inputs \mathbf_1,\dots,\mathbf_n is bits long, and each of the inputs \mathbf_1,\dots,\mathbf_n is bits long. A
decision tree A decision tree is a decision support system, decision support recursive partitioning structure that uses a Tree (graph theory), tree-like Causal model, model of decisions and their possible consequences, including probability, chance event ou ...
of depth \Delta for f can be translated to a communication protocol whose cost is \Delta \cdot D(g): each time the tree queries a bit, the corresponding value of g is computed using an optimal protocol for g. Raz and McKenzie showed that this is optimal up to a constant factor when g is the so-called "indexing gadget", in which x has length c \log n (for a large enough constant ), y has length n^c, and g(x,y) is the x-th bit of y. The proof of the Raz–McKenzie lifting theorem uses the method of simulation, in which a protocol for the composed function f \circ g is used to generate a decision tree for f. Göös, Pitassi and Watson gave an exposition of the original proof. Since then, several works have proved similar theorems with different gadgets, such as inner product. The smallest gadget which can be handled is the indexing gadget with c=1+\epsilon. Göös, Pitassi and Watson extended the Raz–McKenzie technique to randomized protocols. A simple modification of the Raz–McKenzie lifting theorem gives a lower bound of \Delta \cdot D(g) on the logarithm of the size of a protocol tree for computing f \circ g, where \Delta is the depth of the optimal decision tree for f. Garg, Göös, Kamath and Sokolov extended this to the DAG-like setting, and used their result to obtain monotone circuit lower bounds. The same technique has also yielded applications to
proof complexity In logic and theoretical computer science, and specifically proof theory and computational complexity theory, proof complexity is the field aiming to understand and analyse the computational resources that are required to prove or refute statements. ...
. A different type of lifting is exemplified by Sherstov's pattern matrix method, which gives a lower bound on the quantum communication complexity of f \circ g, where is a modified indexing gadget, in terms of the approximate degree of . The approximate degree of a Boolean function is the minimal degree of a polynomial which approximates the function on all Boolean points up to an additive error of 1/3. In contrast to the Raz–McKenzie proof which uses the method of simulation, Sherstov's proof takes a ''dual witness'' to the approximate degree of and gives a lower bound on the quantum query complexity of f \circ g using the ''generalized discrepancy method''. The dual witness for the approximate degree of is a lower bound witness for the approximate degree obtained via
LP duality LP or lp may stand for: Businesses and organizations *LP, Limited partnership in corporate law or a Limited Partner in a venture capital fund In politics *Labour Party (disambiguation), in several countries *Liberal Party, in several countries ...
. This dual witness is massaged into other objects constituting data for the generalized discrepancy method. Another example of this approach is the work of Pitassi and Robere, in which an ''algebraic gap'' is lifted to a lower bound on Razborov's ''rank measure''. The result is a strongly exponential lower bound on the monotone circuit complexity of an explicit function, obtained via the Karchmer–Wigderson characterization of monotone circuit size in terms of communication complexity.


Open problems

Considering a 0 or 1 input matrix M_f= (x,y), the minimum number of bits exchanged to compute f deterministically in the worst case, D(f), is known to be bounded from below by the logarithm of the rank of the matrix M_f. The log rank conjecture proposes that the communication complexity, D(f), is bounded from above by a constant power of the logarithm of the rank of M_f. Since D(f) is bounded from above and below by polynomials of log rank(M_f), we can say D(f) is polynomially related to log rank(M_f). Since the rank of a matrix is polynomial time computable in the size of the matrix, such an upper bound would allow the matrix's communication complexity to be approximated in polynomial time. Note, however, that the size of the matrix itself is exponential in the size of the input. For a randomized protocol, the number of bits exchanged in the worst case, R(f), was conjectured to be polynomially related to the following formula: : \log \min(\textrm(M'_f): M'_f\in \mathbb^, (M_f - M'_f)_\infty\leq 1/3). Such log rank conjectures are valuable because they reduce the question of a matrix's communication complexity to a question of linearly independent rows (columns) of the matrix. This particular version, called the Log-Approximate-Rank Conjecture, was recently refuted by Chattopadhyay, Mande and Sherif (2019) using a surprisingly simple counter-example. This reveals that the essence of the communication complexity problem, for example in the EQ case above, is figuring out where in the matrix the inputs are, in order to find out if they're equivalent.


Applications

Lower bounds in communication complexity can be used to prove lower bounds in decision tree complexity, VLSI circuits, data structures, streaming algorithms,
space–time tradeoff space–time trade-off, also known as time–memory trade-off or the algorithmic space-time continuum in computer science is a case where an algorithm or program trades increased space usage with decreased time. Here, ''space'' refers to the d ...
s for Turing machines and more. Conitzer and Sandholm studied the communication complexity of some common voting rules, which are essential in political and non political organizations. Compilation complexity is a closely related notion, which can be seen as a single-round communication complexity.


See also

* Gap-Hamming problem


Notes


References

* * * Brassard, G. Quantum communication complexity: a survey
https://arxiv.org/abs/quant-ph/0101005
* Dietzfelbinger, M., J. Hromkovic, J., and G. Schnitger,
A comparison of two lower-bound methods for communication complexity
, Theoret. Comput. Sci. 168, 1996. 39-51. * Raz, Ran. "Circuit and Communication Complexity." In Computational Complexity Theory. Steven Rudich and Avi Wigderson, eds. American Mathematical Society Institute for Advanced Study, 2004. 129-137. * A. C. Yao, "Some Complexity Questions Related to Distributed Computing", Proc. of 11th STOC, pp. 209–213, 1979. 14 * I. Newman
Private vs. Common Random Bits in Communication Complexity
Information Processing Letters 39, 1991, pp. 67–71. {{DEFAULTSORT:Communication Complexity Information theory Computational complexity theory Quantum complexity theory>p'_R(x,y) - p(x,y), \geq 0.1\leq 2 \exp(-2(0.1)^2 \cdot 100n) < 2^ Thus when we don't have (x,y) fixed: :\Pr_R exists (x,y):\ , p'_R(x,y) - p(x,y), \geq 0.1\leq \sum_ \Pr_R exists (x,y):\ , p'_R(x,y) - p(x,y), \geq 0.1\leq \sum_ \Pr_R p'_R(x,y) - p(x,y), \geq 0.1< \sum_ 2^ = 1 The last equality above holds because there are 2^ different pairs (x,y). Since the probability does not equal 1, there is some R_0 so that for all (x,y): :, p'_(x,y) - p(x,y), < 0.1 Since P has at most 0.1 error probability, P'_ can have at most 0.2 error probability.


Collapse of Randomized Communication Complexity

Let's say we additionally allow Alice and Bob to share some resource, for example a pair of entangled particles. Using that ressource, Alice and Bob can correlate their information and thus try to 'collapse' (or 'trivialize') communication complexity in the following sense. Definition. ''A resource R is said to be ''"collapsing"'' if, using that resource R, only one bit of classical communication is enough for Alice to know the evaluation f(x,y) in the worst case scenario for any
Boolean function In mathematics, a Boolean function is a function whose arguments and result assume values from a two-element set (usually , or ). Alternative names are switching function, used especially in older computer science literature, and truth functi ...
f. '' The surprising fact of a collapse of communication complexity is that the function f can have arbitrarily large entry size, but still the number of communication bit is constant to a single one. Some resources are shown to be non-collapsing, such as quantum correlations or more generally almost-quantum correlations, whereas on the contrary some other resources are shown to collapse randomized communication complexity, such as the PR-box, or some noisy PR-boxes satisfying some conditions.


Distributional Complexity

One approach to studying randomized communication complexity is through distributional complexity. Given a joint distribution \mu on the inputs of both players, the corresponding distributional complexity of a function f is the minimum cost of a ''deterministic'' protocol R such that \Pr (x,y) = R(x,y)\ge 2/3, where the inputs are sampled according to \mu. Yao's minimax principle (a special case of von Neumann's
minimax theorem In the mathematical area of game theory and of convex optimization, a minimax theorem is a theorem that claims that : \max_ \min_ f(x,y) = \min_ \max_f(x,y) under certain conditions on the sets X and Y and on the function f. It is always true that ...
) states that the randomized communication complexity of a function equals its maximum distributional complexity, where the maximum is taken over all joint distributions of the inputs (not necessarily product distributions!). Yao's principle can be used to prove lower bounds on the randomized communication complexity of a function: design the appropriate joint distribution, and prove a lower bound on the distributional complexity. Since distributional complexity concerns deterministic protocols, this could be easier than proving a lower bound on randomized protocols directly. As an example, let us consider the ''disjointness'' function DISJ: each of the inputs is interpreted as a subset of \, and DISJ(,)=1 if the two sets are disjoint. Razborov proved an \Omega(n) lower bound on the randomized communication complexity by considering the following distribution: with probability 3/4, sample two random disjoint sets of size n/4, and with probability 1/4, sample two random sets of size n/4 with a unique intersection.


Information Complexity

A powerful approach to the study of distributional complexity is information complexity. Initiated by Bar-Yossef, Jayram, Kumar and Sivakumar, the approach was codified in work of Barak, Braverman, Chen and Rao and by Braverman and Rao. The (internal) information complexity of a (possibly randomized) protocol with respect to a distribution is defined as follows. Let (X,Y) \sim \mu be random inputs sampled according to , and let be the transcript of when run on the inputs X,Y. The information complexity of the protocol is : \operatorname_\mu(R) = I(\Pi;Y, X) + I(\Pi;X, Y), where denotes conditional mutual information. The first summand measures the amount of information that Alice learns about Bob's input from the transcript, and the second measures the amount of information that Bob learns about Alice's input. The -error information complexity of a function with respect to a distribution is the infimal information complexity of a protocol for whose error (with respect to ) is at most . Braverman and Rao proved that information equals amortized communication. This means that the cost for solving independent copies of is roughly times the information complexity of . This is analogous to the well-known interpretation of
Shannon entropy Shannon may refer to: People * Shannon (given name) * Shannon (surname) * Shannon (American singer), stage name of singer Brenda Shannon Greene (born 1958) * Shannon (South Korean singer), British-South Korean singer and actress Shannon Arrum ...
as the amortized bit-length required to transmit data from a given information source. Braverman and Rao's proof uses a technique known as "protocol compression", in which an information-efficient protocol is "compressed" into a communication-efficient protocol. The techniques of information complexity enable the computation of the exact (up to first order) communication complexity of set disjointness to be 1.4923\ldots n. Information complexity techniques have also been used to analyze extended formulations, proving an essentially optimal lower bound on the complexity of algorithms based on
linear programming Linear programming (LP), also called linear optimization, is a method to achieve the best outcome (such as maximum profit or lowest cost) in a mathematical model whose requirements and objective are represented by linear function#As a polynomia ...
which approximately solve the maximum clique problem. Omri Weinstein's 2015 survey surveys the subject.


Quantum communication complexity

Quantum communication complexity tries to quantify the communication reduction possible by using quantum effects during a distributed computation. At least three quantum generalizations of communication complexity have been proposed; for a survey see the suggested text by G. Brassard. The first one is the qubit-communication model, where the parties can use quantum communication instead of classical communication, for example by exchanging
photons A photon () is an elementary particle that is a quantum of the electromagnetic field, including electromagnetic radiation such as light and radio waves, and the force carrier for the electromagnetic force. Photons are massless particles that ...
through an
optical fiber An optical fiber, or optical fibre, is a flexible glass or plastic fiber that can transmit light from one end to the other. Such fibers find wide usage in fiber-optic communications, where they permit transmission over longer distances and at ...
. In a second model the communication is still performed with classical bits, but the parties are allowed to manipulate an unlimited supply of quantum entangled states as part of their protocols. By doing measurements on their entangled states, the parties can save on classical communication during a distributed computation (see an application in Collapse of Randomized Communication Complexity). The third model involves access to previously shared entanglement in addition to
qubit In quantum computing, a qubit () or quantum bit is a basic unit of quantum information—the quantum version of the classic binary bit physically realized with a two-state device. A qubit is a two-state (or two-level) quantum-mechanical syste ...
communication, and is the least explored of the three quantum models.


Nondeterministic communication complexity

In nondeterministic communication complexity, Alice and Bob have access to an oracle. After receiving the oracle's word, the parties communicate to deduce f(x,y). The nondeterministic communication complexity is then the maximum over all pairs (x,y) over the sum of number of bits exchanged and the coding length of the oracle word. Viewed differently, this amounts to covering all 1-entries of the 0/1-matrix by combinatorial 1-rectangles (i.e., non-contiguous, non-convex submatrices, whose entries are all one (see Kushilevitz and Nisan or Dietzfelbinger et al.)). The nondeterministic communication complexity is the binary logarithm of the ''rectangle covering number'' of the matrix: the minimum number of combinatorial 1-rectangles required to cover all 1-entries of the matrix, without covering any 0-entries. Nondeterministic communication complexity occurs as a means to obtaining lower bounds for deterministic communication complexity (see Dietzfelbinger et al.), but also in the theory of nonnegative matrices, where it gives a lower bound on the nonnegative rank of a nonnegative matrix.


Unbounded-error communication complexity

In the unbounded-error setting, Alice and Bob have access to a private coin and their own inputs (x, y). In this setting, Alice succeeds if she responds with the correct value of f(x, y) with probability strictly greater than 1/2. In other words, if Alice's responses have ''any'' non-zero correlation to the true value of f(x, y), then the protocol is considered valid. Note that the requirement that the coin is ''private'' is essential. In particular, if the number of public bits shared between Alice and Bob are not counted against the communication complexity, it is easy to argue that computing any function has O(1) communication complexity. On the other hand, both models are equivalent if the number of public bits used by Alice and Bob is counted against the protocol's total communication. Though subtle, lower bounds on this model are extremely strong. More specifically, it is clear that any bound on problems of this class immediately imply equivalent bounds on problems in the deterministic model and the private and public coin models, but such bounds also hold immediately for nondeterministic communication models and quantum communication models. Forster was the first to prove explicit lower bounds for this class, showing that computing the inner product \langle x, y \rangle requires at least \Omega(n) bits of communication, though an earlier result of Alon, Frankl, and Rödl proved that the communication complexity for almost all Boolean functions f: \^n \times \^n \to \ is \Omega(n).


Lifting

Lifting is a general technique in complexity theory in which a lower bound on a simple measure of complexity is "lifted" to a lower bound on a more difficult measure. This technique was pioneered in the context of communication complexity by Raz and McKenzie, who proved the first query-to-communication lifting theorem, and used the result to separate the monotone NC hierarchy. Given a function f\colon \^n \to \ and a gadget g\colon \^a \times \^b \to \, their composition f \circ g\colon \^ \times \^ \to \ is defined as follows: : (f \circ g)(x,y) = f(g(x_ \cdots x_, y_ \cdots y_), \dots, g(x_ \cdots x_, y_ \cdots y_)). In words, x is partitioned into n blocks of length a, and y is partitioned into n blocks of length b. The gadget is applied n times on the blocks, and the outputs are fed into f. Diagrammatically: In this diagram, each of the inputs \mathbf_1,\dots,\mathbf_n is bits long, and each of the inputs \mathbf_1,\dots,\mathbf_n is bits long. A
decision tree A decision tree is a decision support system, decision support recursive partitioning structure that uses a Tree (graph theory), tree-like Causal model, model of decisions and their possible consequences, including probability, chance event ou ...
of depth \Delta for f can be translated to a communication protocol whose cost is \Delta \cdot D(g): each time the tree queries a bit, the corresponding value of g is computed using an optimal protocol for g. Raz and McKenzie showed that this is optimal up to a constant factor when g is the so-called "indexing gadget", in which x has length c \log n (for a large enough constant ), y has length n^c, and g(x,y) is the x-th bit of y. The proof of the Raz–McKenzie lifting theorem uses the method of simulation, in which a protocol for the composed function f \circ g is used to generate a decision tree for f. Göös, Pitassi and Watson gave an exposition of the original proof. Since then, several works have proved similar theorems with different gadgets, such as inner product. The smallest gadget which can be handled is the indexing gadget with c=1+\epsilon. Göös, Pitassi and Watson extended the Raz–McKenzie technique to randomized protocols. A simple modification of the Raz–McKenzie lifting theorem gives a lower bound of \Delta \cdot D(g) on the logarithm of the size of a protocol tree for computing f \circ g, where \Delta is the depth of the optimal decision tree for f. Garg, Göös, Kamath and Sokolov extended this to the DAG-like setting, and used their result to obtain monotone circuit lower bounds. The same technique has also yielded applications to
proof complexity In logic and theoretical computer science, and specifically proof theory and computational complexity theory, proof complexity is the field aiming to understand and analyse the computational resources that are required to prove or refute statements. ...
. A different type of lifting is exemplified by Sherstov's pattern matrix method, which gives a lower bound on the quantum communication complexity of f \circ g, where is a modified indexing gadget, in terms of the approximate degree of . The approximate degree of a Boolean function is the minimal degree of a polynomial which approximates the function on all Boolean points up to an additive error of 1/3. In contrast to the Raz–McKenzie proof which uses the method of simulation, Sherstov's proof takes a ''dual witness'' to the approximate degree of and gives a lower bound on the quantum query complexity of f \circ g using the ''generalized discrepancy method''. The dual witness for the approximate degree of is a lower bound witness for the approximate degree obtained via
LP duality LP or lp may stand for: Businesses and organizations *LP, Limited partnership in corporate law or a Limited Partner in a venture capital fund In politics *Labour Party (disambiguation), in several countries *Liberal Party, in several countries ...
. This dual witness is massaged into other objects constituting data for the generalized discrepancy method. Another example of this approach is the work of Pitassi and Robere, in which an ''algebraic gap'' is lifted to a lower bound on Razborov's ''rank measure''. The result is a strongly exponential lower bound on the monotone circuit complexity of an explicit function, obtained via the Karchmer–Wigderson characterization of monotone circuit size in terms of communication complexity.


Open problems

Considering a 0 or 1 input matrix M_f= (x,y), the minimum number of bits exchanged to compute f deterministically in the worst case, D(f), is known to be bounded from below by the logarithm of the rank of the matrix M_f. The log rank conjecture proposes that the communication complexity, D(f), is bounded from above by a constant power of the logarithm of the rank of M_f. Since D(f) is bounded from above and below by polynomials of log rank(M_f), we can say D(f) is polynomially related to log rank(M_f). Since the rank of a matrix is polynomial time computable in the size of the matrix, such an upper bound would allow the matrix's communication complexity to be approximated in polynomial time. Note, however, that the size of the matrix itself is exponential in the size of the input. For a randomized protocol, the number of bits exchanged in the worst case, R(f), was conjectured to be polynomially related to the following formula: : \log \min(\textrm(M'_f): M'_f\in \mathbb^, (M_f - M'_f)_\infty\leq 1/3). Such log rank conjectures are valuable because they reduce the question of a matrix's communication complexity to a question of linearly independent rows (columns) of the matrix. This particular version, called the Log-Approximate-Rank Conjecture, was recently refuted by Chattopadhyay, Mande and Sherif (2019) using a surprisingly simple counter-example. This reveals that the essence of the communication complexity problem, for example in the EQ case above, is figuring out where in the matrix the inputs are, in order to find out if they're equivalent.


Applications

Lower bounds in communication complexity can be used to prove lower bounds in decision tree complexity, VLSI circuits, data structures, streaming algorithms,
space–time tradeoff space–time trade-off, also known as time–memory trade-off or the algorithmic space-time continuum in computer science is a case where an algorithm or program trades increased space usage with decreased time. Here, ''space'' refers to the d ...
s for Turing machines and more. Conitzer and Sandholm studied the communication complexity of some common voting rules, which are essential in political and non political organizations. Compilation complexity is a closely related notion, which can be seen as a single-round communication complexity.


See also

* Gap-Hamming problem


Notes


References

* * * Brassard, G. Quantum communication complexity: a survey
https://arxiv.org/abs/quant-ph/0101005
* Dietzfelbinger, M., J. Hromkovic, J., and G. Schnitger,
A comparison of two lower-bound methods for communication complexity
, Theoret. Comput. Sci. 168, 1996. 39-51. * Raz, Ran. "Circuit and Communication Complexity." In Computational Complexity Theory. Steven Rudich and Avi Wigderson, eds. American Mathematical Society Institute for Advanced Study, 2004. 129-137. * A. C. Yao, "Some Complexity Questions Related to Distributed Computing", Proc. of 11th STOC, pp. 209–213, 1979. 14 * I. Newman
Private vs. Common Random Bits in Communication Complexity
Information Processing Letters 39, 1991, pp. 67–71. {{DEFAULTSORT:Communication Complexity Information theory Computational complexity theory Quantum complexity theory>p'_R(x,y) - p(x,y), \geq 0.1< \sum_ 2^ = 1 The last equality above holds because there are 2^ different pairs (x,y). Since the probability does not equal 1, there is some R_0 so that for all (x,y): :, p'_(x,y) - p(x,y), < 0.1 Since P has at most 0.1 error probability, P'_ can have at most 0.2 error probability.


Collapse of Randomized Communication Complexity

Let's say we additionally allow Alice and Bob to share some resource, for example a pair of entangled particles. Using that ressource, Alice and Bob can correlate their information and thus try to 'collapse' (or 'trivialize') communication complexity in the following sense. Definition. ''A resource R is said to be ''"collapsing"'' if, using that resource R, only one bit of classical communication is enough for Alice to know the evaluation f(x,y) in the worst case scenario for any
Boolean function In mathematics, a Boolean function is a function whose arguments and result assume values from a two-element set (usually , or ). Alternative names are switching function, used especially in older computer science literature, and truth functi ...
f. '' The surprising fact of a collapse of communication complexity is that the function f can have arbitrarily large entry size, but still the number of communication bit is constant to a single one. Some resources are shown to be non-collapsing, such as quantum correlations or more generally almost-quantum correlations, whereas on the contrary some other resources are shown to collapse randomized communication complexity, such as the PR-box, or some noisy PR-boxes satisfying some conditions.


Distributional Complexity

One approach to studying randomized communication complexity is through distributional complexity. Given a joint distribution \mu on the inputs of both players, the corresponding distributional complexity of a function f is the minimum cost of a ''deterministic'' protocol R such that \Pr (x,y) = R(x,y)\ge 2/3, where the inputs are sampled according to \mu. Yao's minimax principle (a special case of von Neumann's
minimax theorem In the mathematical area of game theory and of convex optimization, a minimax theorem is a theorem that claims that : \max_ \min_ f(x,y) = \min_ \max_f(x,y) under certain conditions on the sets X and Y and on the function f. It is always true that ...
) states that the randomized communication complexity of a function equals its maximum distributional complexity, where the maximum is taken over all joint distributions of the inputs (not necessarily product distributions!). Yao's principle can be used to prove lower bounds on the randomized communication complexity of a function: design the appropriate joint distribution, and prove a lower bound on the distributional complexity. Since distributional complexity concerns deterministic protocols, this could be easier than proving a lower bound on randomized protocols directly. As an example, let us consider the ''disjointness'' function DISJ: each of the inputs is interpreted as a subset of \, and DISJ(,)=1 if the two sets are disjoint. Razborov proved an \Omega(n) lower bound on the randomized communication complexity by considering the following distribution: with probability 3/4, sample two random disjoint sets of size n/4, and with probability 1/4, sample two random sets of size n/4 with a unique intersection.


Information Complexity

A powerful approach to the study of distributional complexity is information complexity. Initiated by Bar-Yossef, Jayram, Kumar and Sivakumar, the approach was codified in work of Barak, Braverman, Chen and Rao and by Braverman and Rao. The (internal) information complexity of a (possibly randomized) protocol with respect to a distribution is defined as follows. Let (X,Y) \sim \mu be random inputs sampled according to , and let be the transcript of when run on the inputs X,Y. The information complexity of the protocol is : \operatorname_\mu(R) = I(\Pi;Y, X) + I(\Pi;X, Y), where denotes conditional mutual information. The first summand measures the amount of information that Alice learns about Bob's input from the transcript, and the second measures the amount of information that Bob learns about Alice's input. The -error information complexity of a function with respect to a distribution is the infimal information complexity of a protocol for whose error (with respect to ) is at most . Braverman and Rao proved that information equals amortized communication. This means that the cost for solving independent copies of is roughly times the information complexity of . This is analogous to the well-known interpretation of
Shannon entropy Shannon may refer to: People * Shannon (given name) * Shannon (surname) * Shannon (American singer), stage name of singer Brenda Shannon Greene (born 1958) * Shannon (South Korean singer), British-South Korean singer and actress Shannon Arrum ...
as the amortized bit-length required to transmit data from a given information source. Braverman and Rao's proof uses a technique known as "protocol compression", in which an information-efficient protocol is "compressed" into a communication-efficient protocol. The techniques of information complexity enable the computation of the exact (up to first order) communication complexity of set disjointness to be 1.4923\ldots n. Information complexity techniques have also been used to analyze extended formulations, proving an essentially optimal lower bound on the complexity of algorithms based on
linear programming Linear programming (LP), also called linear optimization, is a method to achieve the best outcome (such as maximum profit or lowest cost) in a mathematical model whose requirements and objective are represented by linear function#As a polynomia ...
which approximately solve the maximum clique problem. Omri Weinstein's 2015 survey surveys the subject.


Quantum communication complexity

Quantum communication complexity tries to quantify the communication reduction possible by using quantum effects during a distributed computation. At least three quantum generalizations of communication complexity have been proposed; for a survey see the suggested text by G. Brassard. The first one is the qubit-communication model, where the parties can use quantum communication instead of classical communication, for example by exchanging
photons A photon () is an elementary particle that is a quantum of the electromagnetic field, including electromagnetic radiation such as light and radio waves, and the force carrier for the electromagnetic force. Photons are massless particles that ...
through an
optical fiber An optical fiber, or optical fibre, is a flexible glass or plastic fiber that can transmit light from one end to the other. Such fibers find wide usage in fiber-optic communications, where they permit transmission over longer distances and at ...
. In a second model the communication is still performed with classical bits, but the parties are allowed to manipulate an unlimited supply of quantum entangled states as part of their protocols. By doing measurements on their entangled states, the parties can save on classical communication during a distributed computation (see an application in Collapse of Randomized Communication Complexity). The third model involves access to previously shared entanglement in addition to
qubit In quantum computing, a qubit () or quantum bit is a basic unit of quantum information—the quantum version of the classic binary bit physically realized with a two-state device. A qubit is a two-state (or two-level) quantum-mechanical syste ...
communication, and is the least explored of the three quantum models.


Nondeterministic communication complexity

In nondeterministic communication complexity, Alice and Bob have access to an oracle. After receiving the oracle's word, the parties communicate to deduce f(x,y). The nondeterministic communication complexity is then the maximum over all pairs (x,y) over the sum of number of bits exchanged and the coding length of the oracle word. Viewed differently, this amounts to covering all 1-entries of the 0/1-matrix by combinatorial 1-rectangles (i.e., non-contiguous, non-convex submatrices, whose entries are all one (see Kushilevitz and Nisan or Dietzfelbinger et al.)). The nondeterministic communication complexity is the binary logarithm of the ''rectangle covering number'' of the matrix: the minimum number of combinatorial 1-rectangles required to cover all 1-entries of the matrix, without covering any 0-entries. Nondeterministic communication complexity occurs as a means to obtaining lower bounds for deterministic communication complexity (see Dietzfelbinger et al.), but also in the theory of nonnegative matrices, where it gives a lower bound on the nonnegative rank of a nonnegative matrix.


Unbounded-error communication complexity

In the unbounded-error setting, Alice and Bob have access to a private coin and their own inputs (x, y). In this setting, Alice succeeds if she responds with the correct value of f(x, y) with probability strictly greater than 1/2. In other words, if Alice's responses have ''any'' non-zero correlation to the true value of f(x, y), then the protocol is considered valid. Note that the requirement that the coin is ''private'' is essential. In particular, if the number of public bits shared between Alice and Bob are not counted against the communication complexity, it is easy to argue that computing any function has O(1) communication complexity. On the other hand, both models are equivalent if the number of public bits used by Alice and Bob is counted against the protocol's total communication. Though subtle, lower bounds on this model are extremely strong. More specifically, it is clear that any bound on problems of this class immediately imply equivalent bounds on problems in the deterministic model and the private and public coin models, but such bounds also hold immediately for nondeterministic communication models and quantum communication models. Forster was the first to prove explicit lower bounds for this class, showing that computing the inner product \langle x, y \rangle requires at least \Omega(n) bits of communication, though an earlier result of Alon, Frankl, and Rödl proved that the communication complexity for almost all Boolean functions f: \^n \times \^n \to \ is \Omega(n).


Lifting

Lifting is a general technique in complexity theory in which a lower bound on a simple measure of complexity is "lifted" to a lower bound on a more difficult measure. This technique was pioneered in the context of communication complexity by Raz and McKenzie, who proved the first query-to-communication lifting theorem, and used the result to separate the monotone NC hierarchy. Given a function f\colon \^n \to \ and a gadget g\colon \^a \times \^b \to \, their composition f \circ g\colon \^ \times \^ \to \ is defined as follows: : (f \circ g)(x,y) = f(g(x_ \cdots x_, y_ \cdots y_), \dots, g(x_ \cdots x_, y_ \cdots y_)). In words, x is partitioned into n blocks of length a, and y is partitioned into n blocks of length b. The gadget is applied n times on the blocks, and the outputs are fed into f. Diagrammatically: In this diagram, each of the inputs \mathbf_1,\dots,\mathbf_n is bits long, and each of the inputs \mathbf_1,\dots,\mathbf_n is bits long. A
decision tree A decision tree is a decision support system, decision support recursive partitioning structure that uses a Tree (graph theory), tree-like Causal model, model of decisions and their possible consequences, including probability, chance event ou ...
of depth \Delta for f can be translated to a communication protocol whose cost is \Delta \cdot D(g): each time the tree queries a bit, the corresponding value of g is computed using an optimal protocol for g. Raz and McKenzie showed that this is optimal up to a constant factor when g is the so-called "indexing gadget", in which x has length c \log n (for a large enough constant ), y has length n^c, and g(x,y) is the x-th bit of y. The proof of the Raz–McKenzie lifting theorem uses the method of simulation, in which a protocol for the composed function f \circ g is used to generate a decision tree for f. Göös, Pitassi and Watson gave an exposition of the original proof. Since then, several works have proved similar theorems with different gadgets, such as inner product. The smallest gadget which can be handled is the indexing gadget with c=1+\epsilon. Göös, Pitassi and Watson extended the Raz–McKenzie technique to randomized protocols. A simple modification of the Raz–McKenzie lifting theorem gives a lower bound of \Delta \cdot D(g) on the logarithm of the size of a protocol tree for computing f \circ g, where \Delta is the depth of the optimal decision tree for f. Garg, Göös, Kamath and Sokolov extended this to the DAG-like setting, and used their result to obtain monotone circuit lower bounds. The same technique has also yielded applications to
proof complexity In logic and theoretical computer science, and specifically proof theory and computational complexity theory, proof complexity is the field aiming to understand and analyse the computational resources that are required to prove or refute statements. ...
. A different type of lifting is exemplified by Sherstov's pattern matrix method, which gives a lower bound on the quantum communication complexity of f \circ g, where is a modified indexing gadget, in terms of the approximate degree of . The approximate degree of a Boolean function is the minimal degree of a polynomial which approximates the function on all Boolean points up to an additive error of 1/3. In contrast to the Raz–McKenzie proof which uses the method of simulation, Sherstov's proof takes a ''dual witness'' to the approximate degree of and gives a lower bound on the quantum query complexity of f \circ g using the ''generalized discrepancy method''. The dual witness for the approximate degree of is a lower bound witness for the approximate degree obtained via
LP duality LP or lp may stand for: Businesses and organizations *LP, Limited partnership in corporate law or a Limited Partner in a venture capital fund In politics *Labour Party (disambiguation), in several countries *Liberal Party, in several countries ...
. This dual witness is massaged into other objects constituting data for the generalized discrepancy method. Another example of this approach is the work of Pitassi and Robere, in which an ''algebraic gap'' is lifted to a lower bound on Razborov's ''rank measure''. The result is a strongly exponential lower bound on the monotone circuit complexity of an explicit function, obtained via the Karchmer–Wigderson characterization of monotone circuit size in terms of communication complexity.


Open problems

Considering a 0 or 1 input matrix M_f= (x,y), the minimum number of bits exchanged to compute f deterministically in the worst case, D(f), is known to be bounded from below by the logarithm of the rank of the matrix M_f. The log rank conjecture proposes that the communication complexity, D(f), is bounded from above by a constant power of the logarithm of the rank of M_f. Since D(f) is bounded from above and below by polynomials of log rank(M_f), we can say D(f) is polynomially related to log rank(M_f). Since the rank of a matrix is polynomial time computable in the size of the matrix, such an upper bound would allow the matrix's communication complexity to be approximated in polynomial time. Note, however, that the size of the matrix itself is exponential in the size of the input. For a randomized protocol, the number of bits exchanged in the worst case, R(f), was conjectured to be polynomially related to the following formula: : \log \min(\textrm(M'_f): M'_f\in \mathbb^, (M_f - M'_f)_\infty\leq 1/3). Such log rank conjectures are valuable because they reduce the question of a matrix's communication complexity to a question of linearly independent rows (columns) of the matrix. This particular version, called the Log-Approximate-Rank Conjecture, was recently refuted by Chattopadhyay, Mande and Sherif (2019) using a surprisingly simple counter-example. This reveals that the essence of the communication complexity problem, for example in the EQ case above, is figuring out where in the matrix the inputs are, in order to find out if they're equivalent.


Applications

Lower bounds in communication complexity can be used to prove lower bounds in decision tree complexity, VLSI circuits, data structures, streaming algorithms,
space–time tradeoff space–time trade-off, also known as time–memory trade-off or the algorithmic space-time continuum in computer science is a case where an algorithm or program trades increased space usage with decreased time. Here, ''space'' refers to the d ...
s for Turing machines and more. Conitzer and Sandholm studied the communication complexity of some common voting rules, which are essential in political and non political organizations. Compilation complexity is a closely related notion, which can be seen as a single-round communication complexity.


See also

* Gap-Hamming problem


Notes


References

* * * Brassard, G. Quantum communication complexity: a survey
https://arxiv.org/abs/quant-ph/0101005
* Dietzfelbinger, M., J. Hromkovic, J., and G. Schnitger,
A comparison of two lower-bound methods for communication complexity
, Theoret. Comput. Sci. 168, 1996. 39-51. * Raz, Ran. "Circuit and Communication Complexity." In Computational Complexity Theory. Steven Rudich and Avi Wigderson, eds. American Mathematical Society Institute for Advanced Study, 2004. 129-137. * A. C. Yao, "Some Complexity Questions Related to Distributed Computing", Proc. of 11th STOC, pp. 209–213, 1979. 14 * I. Newman
Private vs. Common Random Bits in Communication Complexity
Information Processing Letters 39, 1991, pp. 67–71. {{DEFAULTSORT:Communication Complexity Information theory Computational complexity theory Quantum complexity theory>p'_R(x,y) - p(x,y), \geq 0.1< \sum_ 2^ = 1 The last equality above holds because there are 2^ different pairs (x,y). Since the probability does not equal 1, there is some R_0 so that for all (x,y): :, p'_(x,y) - p(x,y), < 0.1 Since P has at most 0.1 error probability, P'_ can have at most 0.2 error probability.


Collapse of Randomized Communication Complexity

Let's say we additionally allow Alice and Bob to share some resource, for example a pair of entangled particles. Using that ressource, Alice and Bob can correlate their information and thus try to 'collapse' (or 'trivialize') communication complexity in the following sense. Definition. ''A resource R is said to be ''"collapsing"'' if, using that resource R, only one bit of classical communication is enough for Alice to know the evaluation f(x,y) in the worst case scenario for any
Boolean function In mathematics, a Boolean function is a function whose arguments and result assume values from a two-element set (usually , or ). Alternative names are switching function, used especially in older computer science literature, and truth functi ...
f. '' The surprising fact of a collapse of communication complexity is that the function f can have arbitrarily large entry size, but still the number of communication bit is constant to a single one. Some resources are shown to be non-collapsing, such as quantum correlations or more generally almost-quantum correlations, whereas on the contrary some other resources are shown to collapse randomized communication complexity, such as the PR-box, or some noisy PR-boxes satisfying some conditions.


Distributional Complexity

One approach to studying randomized communication complexity is through distributional complexity. Given a joint distribution \mu on the inputs of both players, the corresponding distributional complexity of a function f is the minimum cost of a ''deterministic'' protocol R such that \Pr (x,y) = R(x,y)\ge 2/3, where the inputs are sampled according to \mu. Yao's minimax principle (a special case of von Neumann's
minimax theorem In the mathematical area of game theory and of convex optimization, a minimax theorem is a theorem that claims that : \max_ \min_ f(x,y) = \min_ \max_f(x,y) under certain conditions on the sets X and Y and on the function f. It is always true that ...
) states that the randomized communication complexity of a function equals its maximum distributional complexity, where the maximum is taken over all joint distributions of the inputs (not necessarily product distributions!). Yao's principle can be used to prove lower bounds on the randomized communication complexity of a function: design the appropriate joint distribution, and prove a lower bound on the distributional complexity. Since distributional complexity concerns deterministic protocols, this could be easier than proving a lower bound on randomized protocols directly. As an example, let us consider the ''disjointness'' function DISJ: each of the inputs is interpreted as a subset of \, and DISJ(,)=1 if the two sets are disjoint. Razborov proved an \Omega(n) lower bound on the randomized communication complexity by considering the following distribution: with probability 3/4, sample two random disjoint sets of size n/4, and with probability 1/4, sample two random sets of size n/4 with a unique intersection.


Information Complexity

A powerful approach to the study of distributional complexity is information complexity. Initiated by Bar-Yossef, Jayram, Kumar and Sivakumar, the approach was codified in work of Barak, Braverman, Chen and Rao and by Braverman and Rao. The (internal) information complexity of a (possibly randomized) protocol with respect to a distribution is defined as follows. Let (X,Y) \sim \mu be random inputs sampled according to , and let be the transcript of when run on the inputs X,Y. The information complexity of the protocol is : \operatorname_\mu(R) = I(\Pi;Y, X) + I(\Pi;X, Y), where denotes conditional mutual information. The first summand measures the amount of information that Alice learns about Bob's input from the transcript, and the second measures the amount of information that Bob learns about Alice's input. The -error information complexity of a function with respect to a distribution is the infimal information complexity of a protocol for whose error (with respect to ) is at most . Braverman and Rao proved that information equals amortized communication. This means that the cost for solving independent copies of is roughly times the information complexity of . This is analogous to the well-known interpretation of
Shannon entropy Shannon may refer to: People * Shannon (given name) * Shannon (surname) * Shannon (American singer), stage name of singer Brenda Shannon Greene (born 1958) * Shannon (South Korean singer), British-South Korean singer and actress Shannon Arrum ...
as the amortized bit-length required to transmit data from a given information source. Braverman and Rao's proof uses a technique known as "protocol compression", in which an information-efficient protocol is "compressed" into a communication-efficient protocol. The techniques of information complexity enable the computation of the exact (up to first order) communication complexity of set disjointness to be 1.4923\ldots n. Information complexity techniques have also been used to analyze extended formulations, proving an essentially optimal lower bound on the complexity of algorithms based on
linear programming Linear programming (LP), also called linear optimization, is a method to achieve the best outcome (such as maximum profit or lowest cost) in a mathematical model whose requirements and objective are represented by linear function#As a polynomia ...
which approximately solve the maximum clique problem. Omri Weinstein's 2015 survey surveys the subject.


Quantum communication complexity

Quantum communication complexity tries to quantify the communication reduction possible by using quantum effects during a distributed computation. At least three quantum generalizations of communication complexity have been proposed; for a survey see the suggested text by G. Brassard. The first one is the qubit-communication model, where the parties can use quantum communication instead of classical communication, for example by exchanging
photons A photon () is an elementary particle that is a quantum of the electromagnetic field, including electromagnetic radiation such as light and radio waves, and the force carrier for the electromagnetic force. Photons are massless particles that ...
through an
optical fiber An optical fiber, or optical fibre, is a flexible glass or plastic fiber that can transmit light from one end to the other. Such fibers find wide usage in fiber-optic communications, where they permit transmission over longer distances and at ...
. In a second model the communication is still performed with classical bits, but the parties are allowed to manipulate an unlimited supply of quantum entangled states as part of their protocols. By doing measurements on their entangled states, the parties can save on classical communication during a distributed computation (see an application in Collapse of Randomized Communication Complexity). The third model involves access to previously shared entanglement in addition to
qubit In quantum computing, a qubit () or quantum bit is a basic unit of quantum information—the quantum version of the classic binary bit physically realized with a two-state device. A qubit is a two-state (or two-level) quantum-mechanical syste ...
communication, and is the least explored of the three quantum models.


Nondeterministic communication complexity

In nondeterministic communication complexity, Alice and Bob have access to an oracle. After receiving the oracle's word, the parties communicate to deduce f(x,y). The nondeterministic communication complexity is then the maximum over all pairs (x,y) over the sum of number of bits exchanged and the coding length of the oracle word. Viewed differently, this amounts to covering all 1-entries of the 0/1-matrix by combinatorial 1-rectangles (i.e., non-contiguous, non-convex submatrices, whose entries are all one (see Kushilevitz and Nisan or Dietzfelbinger et al.)). The nondeterministic communication complexity is the binary logarithm of the ''rectangle covering number'' of the matrix: the minimum number of combinatorial 1-rectangles required to cover all 1-entries of the matrix, without covering any 0-entries. Nondeterministic communication complexity occurs as a means to obtaining lower bounds for deterministic communication complexity (see Dietzfelbinger et al.), but also in the theory of nonnegative matrices, where it gives a lower bound on the nonnegative rank of a nonnegative matrix.


Unbounded-error communication complexity

In the unbounded-error setting, Alice and Bob have access to a private coin and their own inputs (x, y). In this setting, Alice succeeds if she responds with the correct value of f(x, y) with probability strictly greater than 1/2. In other words, if Alice's responses have ''any'' non-zero correlation to the true value of f(x, y), then the protocol is considered valid. Note that the requirement that the coin is ''private'' is essential. In particular, if the number of public bits shared between Alice and Bob are not counted against the communication complexity, it is easy to argue that computing any function has O(1) communication complexity. On the other hand, both models are equivalent if the number of public bits used by Alice and Bob is counted against the protocol's total communication. Though subtle, lower bounds on this model are extremely strong. More specifically, it is clear that any bound on problems of this class immediately imply equivalent bounds on problems in the deterministic model and the private and public coin models, but such bounds also hold immediately for nondeterministic communication models and quantum communication models. Forster was the first to prove explicit lower bounds for this class, showing that computing the inner product \langle x, y \rangle requires at least \Omega(n) bits of communication, though an earlier result of Alon, Frankl, and Rödl proved that the communication complexity for almost all Boolean functions f: \^n \times \^n \to \ is \Omega(n).


Lifting

Lifting is a general technique in complexity theory in which a lower bound on a simple measure of complexity is "lifted" to a lower bound on a more difficult measure. This technique was pioneered in the context of communication complexity by Raz and McKenzie, who proved the first query-to-communication lifting theorem, and used the result to separate the monotone NC hierarchy. Given a function f\colon \^n \to \ and a gadget g\colon \^a \times \^b \to \, their composition f \circ g\colon \^ \times \^ \to \ is defined as follows: : (f \circ g)(x,y) = f(g(x_ \cdots x_, y_ \cdots y_), \dots, g(x_ \cdots x_, y_ \cdots y_)). In words, x is partitioned into n blocks of length a, and y is partitioned into n blocks of length b. The gadget is applied n times on the blocks, and the outputs are fed into f. Diagrammatically: In this diagram, each of the inputs \mathbf_1,\dots,\mathbf_n is bits long, and each of the inputs \mathbf_1,\dots,\mathbf_n is bits long. A
decision tree A decision tree is a decision support system, decision support recursive partitioning structure that uses a Tree (graph theory), tree-like Causal model, model of decisions and their possible consequences, including probability, chance event ou ...
of depth \Delta for f can be translated to a communication protocol whose cost is \Delta \cdot D(g): each time the tree queries a bit, the corresponding value of g is computed using an optimal protocol for g. Raz and McKenzie showed that this is optimal up to a constant factor when g is the so-called "indexing gadget", in which x has length c \log n (for a large enough constant ), y has length n^c, and g(x,y) is the x-th bit of y. The proof of the Raz–McKenzie lifting theorem uses the method of simulation, in which a protocol for the composed function f \circ g is used to generate a decision tree for f. Göös, Pitassi and Watson gave an exposition of the original proof. Since then, several works have proved similar theorems with different gadgets, such as inner product. The smallest gadget which can be handled is the indexing gadget with c=1+\epsilon. Göös, Pitassi and Watson extended the Raz–McKenzie technique to randomized protocols. A simple modification of the Raz–McKenzie lifting theorem gives a lower bound of \Delta \cdot D(g) on the logarithm of the size of a protocol tree for computing f \circ g, where \Delta is the depth of the optimal decision tree for f. Garg, Göös, Kamath and Sokolov extended this to the DAG-like setting, and used their result to obtain monotone circuit lower bounds. The same technique has also yielded applications to
proof complexity In logic and theoretical computer science, and specifically proof theory and computational complexity theory, proof complexity is the field aiming to understand and analyse the computational resources that are required to prove or refute statements. ...
. A different type of lifting is exemplified by Sherstov's pattern matrix method, which gives a lower bound on the quantum communication complexity of f \circ g, where is a modified indexing gadget, in terms of the approximate degree of . The approximate degree of a Boolean function is the minimal degree of a polynomial which approximates the function on all Boolean points up to an additive error of 1/3. In contrast to the Raz–McKenzie proof which uses the method of simulation, Sherstov's proof takes a ''dual witness'' to the approximate degree of and gives a lower bound on the quantum query complexity of f \circ g using the ''generalized discrepancy method''. The dual witness for the approximate degree of is a lower bound witness for the approximate degree obtained via
LP duality LP or lp may stand for: Businesses and organizations *LP, Limited partnership in corporate law or a Limited Partner in a venture capital fund In politics *Labour Party (disambiguation), in several countries *Liberal Party, in several countries ...
. This dual witness is massaged into other objects constituting data for the generalized discrepancy method. Another example of this approach is the work of Pitassi and Robere, in which an ''algebraic gap'' is lifted to a lower bound on Razborov's ''rank measure''. The result is a strongly exponential lower bound on the monotone circuit complexity of an explicit function, obtained via the Karchmer–Wigderson characterization of monotone circuit size in terms of communication complexity.


Open problems

Considering a 0 or 1 input matrix M_f= (x,y), the minimum number of bits exchanged to compute f deterministically in the worst case, D(f), is known to be bounded from below by the logarithm of the rank of the matrix M_f. The log rank conjecture proposes that the communication complexity, D(f), is bounded from above by a constant power of the logarithm of the rank of M_f. Since D(f) is bounded from above and below by polynomials of log rank(M_f), we can say D(f) is polynomially related to log rank(M_f). Since the rank of a matrix is polynomial time computable in the size of the matrix, such an upper bound would allow the matrix's communication complexity to be approximated in polynomial time. Note, however, that the size of the matrix itself is exponential in the size of the input. For a randomized protocol, the number of bits exchanged in the worst case, R(f), was conjectured to be polynomially related to the following formula: : \log \min(\textrm(M'_f): M'_f\in \mathbb^, (M_f - M'_f)_\infty\leq 1/3). Such log rank conjectures are valuable because they reduce the question of a matrix's communication complexity to a question of linearly independent rows (columns) of the matrix. This particular version, called the Log-Approximate-Rank Conjecture, was recently refuted by Chattopadhyay, Mande and Sherif (2019) using a surprisingly simple counter-example. This reveals that the essence of the communication complexity problem, for example in the EQ case above, is figuring out where in the matrix the inputs are, in order to find out if they're equivalent.


Applications

Lower bounds in communication complexity can be used to prove lower bounds in decision tree complexity, VLSI circuits, data structures, streaming algorithms,
space–time tradeoff space–time trade-off, also known as time–memory trade-off or the algorithmic space-time continuum in computer science is a case where an algorithm or program trades increased space usage with decreased time. Here, ''space'' refers to the d ...
s for Turing machines and more. Conitzer and Sandholm studied the communication complexity of some common voting rules, which are essential in political and non political organizations. Compilation complexity is a closely related notion, which can be seen as a single-round communication complexity.


See also

* Gap-Hamming problem


Notes


References

* * * Brassard, G. Quantum communication complexity: a survey
https://arxiv.org/abs/quant-ph/0101005
* Dietzfelbinger, M., J. Hromkovic, J., and G. Schnitger,
A comparison of two lower-bound methods for communication complexity
, Theoret. Comput. Sci. 168, 1996. 39-51. * Raz, Ran. "Circuit and Communication Complexity." In Computational Complexity Theory. Steven Rudich and Avi Wigderson, eds. American Mathematical Society Institute for Advanced Study, 2004. 129-137. * A. C. Yao, "Some Complexity Questions Related to Distributed Computing", Proc. of 11th STOC, pp. 209–213, 1979. 14 * I. Newman
Private vs. Common Random Bits in Communication Complexity
Information Processing Letters 39, 1991, pp. 67–71. {{DEFAULTSORT:Communication Complexity Information theory Computational complexity theory Quantum complexity theory>p'_R(x,y) - p(x,y), \geq 0.1\leq 2 \exp(-2(0.1)^2 \cdot 100n) < 2^ Thus when we don't have (x,y) fixed: :\Pr_R exists (x,y):\ , p'_R(x,y) - p(x,y), \geq 0.1\leq \sum_ \Pr_R exists (x,y):\ , p'_R(x,y) - p(x,y), \geq 0.1\leq \sum_ \Pr_R R, only one bit of classical communication is enough for Alice to know the evaluation f(x,y) in the worst case scenario for any
Boolean function In mathematics, a Boolean function is a function whose arguments and result assume values from a two-element set (usually , or ). Alternative names are switching function, used especially in older computer science literature, and truth functi ...
f. '' The surprising fact of a collapse of communication complexity is that the function f can have arbitrarily large entry size, but still the number of communication bit is constant to a single one. Some resources are shown to be non-collapsing, such as quantum correlations or more generally almost-quantum correlations, whereas on the contrary some other resources are shown to collapse randomized communication complexity, such as the PR-box, or some noisy PR-boxes satisfying some conditions.


Distributional Complexity

One approach to studying randomized communication complexity is through distributional complexity. Given a joint distribution \mu on the inputs of both players, the corresponding distributional complexity of a function f is the minimum cost of a ''deterministic'' protocol R such that \Pr (x,y) = R(x,y)\ge 2/3, where the inputs are sampled according to \mu. Yao's minimax principle (a special case of von Neumann's
minimax theorem In the mathematical area of game theory and of convex optimization, a minimax theorem is a theorem that claims that : \max_ \min_ f(x,y) = \min_ \max_f(x,y) under certain conditions on the sets X and Y and on the function f. It is always true that ...
) states that the randomized communication complexity of a function equals its maximum distributional complexity, where the maximum is taken over all joint distributions of the inputs (not necessarily product distributions!). Yao's principle can be used to prove lower bounds on the randomized communication complexity of a function: design the appropriate joint distribution, and prove a lower bound on the distributional complexity. Since distributional complexity concerns deterministic protocols, this could be easier than proving a lower bound on randomized protocols directly. As an example, let us consider the ''disjointness'' function DISJ: each of the inputs is interpreted as a subset of \, and DISJ(,)=1 if the two sets are disjoint. Razborov proved an \Omega(n) lower bound on the randomized communication complexity by considering the following distribution: with probability 3/4, sample two random disjoint sets of size n/4, and with probability 1/4, sample two random sets of size n/4 with a unique intersection.


Information Complexity

A powerful approach to the study of distributional complexity is information complexity. Initiated by Bar-Yossef, Jayram, Kumar and Sivakumar, the approach was codified in work of Barak, Braverman, Chen and Rao and by Braverman and Rao. The (internal) information complexity of a (possibly randomized) protocol with respect to a distribution is defined as follows. Let (X,Y) \sim \mu be random inputs sampled according to , and let be the transcript of when run on the inputs X,Y. The information complexity of the protocol is : \operatorname_\mu(R) = I(\Pi;Y, X) + I(\Pi;X, Y), where denotes conditional mutual information. The first summand measures the amount of information that Alice learns about Bob's input from the transcript, and the second measures the amount of information that Bob learns about Alice's input. The -error information complexity of a function with respect to a distribution is the infimal information complexity of a protocol for whose error (with respect to ) is at most . Braverman and Rao proved that information equals amortized communication. This means that the cost for solving independent copies of is roughly times the information complexity of . This is analogous to the well-known interpretation of
Shannon entropy Shannon may refer to: People * Shannon (given name) * Shannon (surname) * Shannon (American singer), stage name of singer Brenda Shannon Greene (born 1958) * Shannon (South Korean singer), British-South Korean singer and actress Shannon Arrum ...
as the amortized bit-length required to transmit data from a given information source. Braverman and Rao's proof uses a technique known as "protocol compression", in which an information-efficient protocol is "compressed" into a communication-efficient protocol. The techniques of information complexity enable the computation of the exact (up to first order) communication complexity of set disjointness to be 1.4923\ldots n. Information complexity techniques have also been used to analyze extended formulations, proving an essentially optimal lower bound on the complexity of algorithms based on
linear programming Linear programming (LP), also called linear optimization, is a method to achieve the best outcome (such as maximum profit or lowest cost) in a mathematical model whose requirements and objective are represented by linear function#As a polynomia ...
which approximately solve the maximum clique problem. Omri Weinstein's 2015 survey surveys the subject.


Quantum communication complexity

Quantum communication complexity tries to quantify the communication reduction possible by using quantum effects during a distributed computation. At least three quantum generalizations of communication complexity have been proposed; for a survey see the suggested text by G. Brassard. The first one is the qubit-communication model, where the parties can use quantum communication instead of classical communication, for example by exchanging
photons A photon () is an elementary particle that is a quantum of the electromagnetic field, including electromagnetic radiation such as light and radio waves, and the force carrier for the electromagnetic force. Photons are massless particles that ...
through an
optical fiber An optical fiber, or optical fibre, is a flexible glass or plastic fiber that can transmit light from one end to the other. Such fibers find wide usage in fiber-optic communications, where they permit transmission over longer distances and at ...
. In a second model the communication is still performed with classical bits, but the parties are allowed to manipulate an unlimited supply of quantum entangled states as part of their protocols. By doing measurements on their entangled states, the parties can save on classical communication during a distributed computation (see an application in Collapse of Randomized Communication Complexity). The third model involves access to previously shared entanglement in addition to
qubit In quantum computing, a qubit () or quantum bit is a basic unit of quantum information—the quantum version of the classic binary bit physically realized with a two-state device. A qubit is a two-state (or two-level) quantum-mechanical syste ...
communication, and is the least explored of the three quantum models.


Nondeterministic communication complexity

In nondeterministic communication complexity, Alice and Bob have access to an oracle. After receiving the oracle's word, the parties communicate to deduce f(x,y). The nondeterministic communication complexity is then the maximum over all pairs (x,y) over the sum of number of bits exchanged and the coding length of the oracle word. Viewed differently, this amounts to covering all 1-entries of the 0/1-matrix by combinatorial 1-rectangles (i.e., non-contiguous, non-convex submatrices, whose entries are all one (see Kushilevitz and Nisan or Dietzfelbinger et al.)). The nondeterministic communication complexity is the binary logarithm of the ''rectangle covering number'' of the matrix: the minimum number of combinatorial 1-rectangles required to cover all 1-entries of the matrix, without covering any 0-entries. Nondeterministic communication complexity occurs as a means to obtaining lower bounds for deterministic communication complexity (see Dietzfelbinger et al.), but also in the theory of nonnegative matrices, where it gives a lower bound on the nonnegative rank of a nonnegative matrix.


Unbounded-error communication complexity

In the unbounded-error setting, Alice and Bob have access to a private coin and their own inputs (x, y). In this setting, Alice succeeds if she responds with the correct value of f(x, y) with probability strictly greater than 1/2. In other words, if Alice's responses have ''any'' non-zero correlation to the true value of f(x, y), then the protocol is considered valid. Note that the requirement that the coin is ''private'' is essential. In particular, if the number of public bits shared between Alice and Bob are not counted against the communication complexity, it is easy to argue that computing any function has O(1) communication complexity. On the other hand, both models are equivalent if the number of public bits used by Alice and Bob is counted against the protocol's total communication. Though subtle, lower bounds on this model are extremely strong. More specifically, it is clear that any bound on problems of this class immediately imply equivalent bounds on problems in the deterministic model and the private and public coin models, but such bounds also hold immediately for nondeterministic communication models and quantum communication models. Forster was the first to prove explicit lower bounds for this class, showing that computing the inner product \langle x, y \rangle requires at least \Omega(n) bits of communication, though an earlier result of Alon, Frankl, and Rödl proved that the communication complexity for almost all Boolean functions f: \^n \times \^n \to \ is \Omega(n).


Lifting

Lifting is a general technique in complexity theory in which a lower bound on a simple measure of complexity is "lifted" to a lower bound on a more difficult measure. This technique was pioneered in the context of communication complexity by Raz and McKenzie, who proved the first query-to-communication lifting theorem, and used the result to separate the monotone NC hierarchy. Given a function f\colon \^n \to \ and a gadget g\colon \^a \times \^b \to \, their composition f \circ g\colon \^ \times \^ \to \ is defined as follows: : (f \circ g)(x,y) = f(g(x_ \cdots x_, y_ \cdots y_), \dots, g(x_ \cdots x_, y_ \cdots y_)). In words, x is partitioned into n blocks of length a, and y is partitioned into n blocks of length b. The gadget is applied n times on the blocks, and the outputs are fed into f. Diagrammatically: In this diagram, each of the inputs \mathbf_1,\dots,\mathbf_n is bits long, and each of the inputs \mathbf_1,\dots,\mathbf_n is bits long. A
decision tree A decision tree is a decision support system, decision support recursive partitioning structure that uses a Tree (graph theory), tree-like Causal model, model of decisions and their possible consequences, including probability, chance event ou ...
of depth \Delta for f can be translated to a communication protocol whose cost is \Delta \cdot D(g): each time the tree queries a bit, the corresponding value of g is computed using an optimal protocol for g. Raz and McKenzie showed that this is optimal up to a constant factor when g is the so-called "indexing gadget", in which x has length c \log n (for a large enough constant ), y has length n^c, and g(x,y) is the x-th bit of y. The proof of the Raz–McKenzie lifting theorem uses the method of simulation, in which a protocol for the composed function f \circ g is used to generate a decision tree for f. Göös, Pitassi and Watson gave an exposition of the original proof. Since then, several works have proved similar theorems with different gadgets, such as inner product. The smallest gadget which can be handled is the indexing gadget with c=1+\epsilon. Göös, Pitassi and Watson extended the Raz–McKenzie technique to randomized protocols. A simple modification of the Raz–McKenzie lifting theorem gives a lower bound of \Delta \cdot D(g) on the logarithm of the size of a protocol tree for computing f \circ g, where \Delta is the depth of the optimal decision tree for f. Garg, Göös, Kamath and Sokolov extended this to the DAG-like setting, and used their result to obtain monotone circuit lower bounds. The same technique has also yielded applications to
proof complexity In logic and theoretical computer science, and specifically proof theory and computational complexity theory, proof complexity is the field aiming to understand and analyse the computational resources that are required to prove or refute statements. ...
. A different type of lifting is exemplified by Sherstov's pattern matrix method, which gives a lower bound on the quantum communication complexity of f \circ g, where is a modified indexing gadget, in terms of the approximate degree of . The approximate degree of a Boolean function is the minimal degree of a polynomial which approximates the function on all Boolean points up to an additive error of 1/3. In contrast to the Raz–McKenzie proof which uses the method of simulation, Sherstov's proof takes a ''dual witness'' to the approximate degree of and gives a lower bound on the quantum query complexity of f \circ g using the ''generalized discrepancy method''. The dual witness for the approximate degree of is a lower bound witness for the approximate degree obtained via
LP duality LP or lp may stand for: Businesses and organizations *LP, Limited partnership in corporate law or a Limited Partner in a venture capital fund In politics *Labour Party (disambiguation), in several countries *Liberal Party, in several countries ...
. This dual witness is massaged into other objects constituting data for the generalized discrepancy method. Another example of this approach is the work of Pitassi and Robere, in which an ''algebraic gap'' is lifted to a lower bound on Razborov's ''rank measure''. The result is a strongly exponential lower bound on the monotone circuit complexity of an explicit function, obtained via the Karchmer–Wigderson characterization of monotone circuit size in terms of communication complexity.


Open problems

Considering a 0 or 1 input matrix M_f= (x,y), the minimum number of bits exchanged to compute f deterministically in the worst case, D(f), is known to be bounded from below by the logarithm of the rank of the matrix M_f. The log rank conjecture proposes that the communication complexity, D(f), is bounded from above by a constant power of the logarithm of the rank of M_f. Since D(f) is bounded from above and below by polynomials of log rank(M_f), we can say D(f) is polynomially related to log rank(M_f). Since the rank of a matrix is polynomial time computable in the size of the matrix, such an upper bound would allow the matrix's communication complexity to be approximated in polynomial time. Note, however, that the size of the matrix itself is exponential in the size of the input. For a randomized protocol, the number of bits exchanged in the worst case, R(f), was conjectured to be polynomially related to the following formula: : \log \min(\textrm(M'_f): M'_f\in \mathbb^, (M_f - M'_f)_\infty\leq 1/3). Such log rank conjectures are valuable because they reduce the question of a matrix's communication complexity to a question of linearly independent rows (columns) of the matrix. This particular version, called the Log-Approximate-Rank Conjecture, was recently refuted by Chattopadhyay, Mande and Sherif (2019) using a surprisingly simple counter-example. This reveals that the essence of the communication complexity problem, for example in the EQ case above, is figuring out where in the matrix the inputs are, in order to find out if they're equivalent.


Applications

Lower bounds in communication complexity can be used to prove lower bounds in decision tree complexity, VLSI circuits, data structures, streaming algorithms,
space–time tradeoff space–time trade-off, also known as time–memory trade-off or the algorithmic space-time continuum in computer science is a case where an algorithm or program trades increased space usage with decreased time. Here, ''space'' refers to the d ...
s for Turing machines and more. Conitzer and Sandholm studied the communication complexity of some common voting rules, which are essential in political and non political organizations. Compilation complexity is a closely related notion, which can be seen as a single-round communication complexity.


See also

* Gap-Hamming problem


Notes


References

* * * Brassard, G. Quantum communication complexity: a survey
https://arxiv.org/abs/quant-ph/0101005
* Dietzfelbinger, M., J. Hromkovic, J., and G. Schnitger,
A comparison of two lower-bound methods for communication complexity
, Theoret. Comput. Sci. 168, 1996. 39-51. * Raz, Ran. "Circuit and Communication Complexity." In Computational Complexity Theory. Steven Rudich and Avi Wigderson, eds. American Mathematical Society Institute for Advanced Study, 2004. 129-137. * A. C. Yao, "Some Complexity Questions Related to Distributed Computing", Proc. of 11th STOC, pp. 209–213, 1979. 14 * I. Newman
Private vs. Common Random Bits in Communication Complexity
Information Processing Letters 39, 1991, pp. 67–71. {{DEFAULTSORT:Communication Complexity Information theory Computational complexity theory Quantum complexity theory>p'_R(x,y) - p(x,y), \geq 0.1\leq 2 \exp(-2(0.1)^2 \cdot 100n) < 2^ Thus when we don't have (x,y) fixed: :\Pr_R exists (x,y):\ , p'_R(x,y) - p(x,y), \geq 0.1\leq \sum_ \Pr_R exists (x,y):\ , p'_R(x,y) - p(x,y), \geq 0.1\leq \sum_ \Pr_R p'_R(x,y) - p(x,y), \geq 0.1< \sum_ 2^ = 1 The last equality above holds because there are 2^ different pairs (x,y). Since the probability does not equal 1, there is some R_0 so that for all (x,y): :, p'_(x,y) - p(x,y), < 0.1 Since P has at most 0.1 error probability, P'_ can have at most 0.2 error probability.


Collapse of Randomized Communication Complexity

Let's say we additionally allow Alice and Bob to share some resource, for example a pair of entangled particles. Using that ressource, Alice and Bob can correlate their information and thus try to 'collapse' (or 'trivialize') communication complexity in the following sense. Definition. ''A resource R is said to be ''"collapsing"'' if, using that resource R, only one bit of classical communication is enough for Alice to know the evaluation f(x,y) in the worst case scenario for any
Boolean function In mathematics, a Boolean function is a function whose arguments and result assume values from a two-element set (usually , or ). Alternative names are switching function, used especially in older computer science literature, and truth functi ...
f. '' The surprising fact of a collapse of communication complexity is that the function f can have arbitrarily large entry size, but still the number of communication bit is constant to a single one. Some resources are shown to be non-collapsing, such as quantum correlations or more generally almost-quantum correlations, whereas on the contrary some other resources are shown to collapse randomized communication complexity, such as the PR-box, or some noisy PR-boxes satisfying some conditions.


Distributional Complexity

One approach to studying randomized communication complexity is through distributional complexity. Given a joint distribution \mu on the inputs of both players, the corresponding distributional complexity of a function f is the minimum cost of a ''deterministic'' protocol R such that \Pr (x,y) = R(x,y)\ge 2/3, where the inputs are sampled according to \mu. Yao's minimax principle (a special case of von Neumann's
minimax theorem In the mathematical area of game theory and of convex optimization, a minimax theorem is a theorem that claims that : \max_ \min_ f(x,y) = \min_ \max_f(x,y) under certain conditions on the sets X and Y and on the function f. It is always true that ...
) states that the randomized communication complexity of a function equals its maximum distributional complexity, where the maximum is taken over all joint distributions of the inputs (not necessarily product distributions!). Yao's principle can be used to prove lower bounds on the randomized communication complexity of a function: design the appropriate joint distribution, and prove a lower bound on the distributional complexity. Since distributional complexity concerns deterministic protocols, this could be easier than proving a lower bound on randomized protocols directly. As an example, let us consider the ''disjointness'' function DISJ: each of the inputs is interpreted as a subset of \, and DISJ(,)=1 if the two sets are disjoint. Razborov proved an \Omega(n) lower bound on the randomized communication complexity by considering the following distribution: with probability 3/4, sample two random disjoint sets of size n/4, and with probability 1/4, sample two random sets of size n/4 with a unique intersection.


Information Complexity

A powerful approach to the study of distributional complexity is information complexity. Initiated by Bar-Yossef, Jayram, Kumar and Sivakumar, the approach was codified in work of Barak, Braverman, Chen and Rao and by Braverman and Rao. The (internal) information complexity of a (possibly randomized) protocol with respect to a distribution is defined as follows. Let (X,Y) \sim \mu be random inputs sampled according to , and let be the transcript of when run on the inputs X,Y. The information complexity of the protocol is : \operatorname_\mu(R) = I(\Pi;Y, X) + I(\Pi;X, Y), where denotes conditional mutual information. The first summand measures the amount of information that Alice learns about Bob's input from the transcript, and the second measures the amount of information that Bob learns about Alice's input. The -error information complexity of a function with respect to a distribution is the infimal information complexity of a protocol for whose error (with respect to ) is at most . Braverman and Rao proved that information equals amortized communication. This means that the cost for solving independent copies of is roughly times the information complexity of . This is analogous to the well-known interpretation of
Shannon entropy Shannon may refer to: People * Shannon (given name) * Shannon (surname) * Shannon (American singer), stage name of singer Brenda Shannon Greene (born 1958) * Shannon (South Korean singer), British-South Korean singer and actress Shannon Arrum ...
as the amortized bit-length required to transmit data from a given information source. Braverman and Rao's proof uses a technique known as "protocol compression", in which an information-efficient protocol is "compressed" into a communication-efficient protocol. The techniques of information complexity enable the computation of the exact (up to first order) communication complexity of set disjointness to be 1.4923\ldots n. Information complexity techniques have also been used to analyze extended formulations, proving an essentially optimal lower bound on the complexity of algorithms based on
linear programming Linear programming (LP), also called linear optimization, is a method to achieve the best outcome (such as maximum profit or lowest cost) in a mathematical model whose requirements and objective are represented by linear function#As a polynomia ...
which approximately solve the maximum clique problem. Omri Weinstein's 2015 survey surveys the subject.


Quantum communication complexity

Quantum communication complexity tries to quantify the communication reduction possible by using quantum effects during a distributed computation. At least three quantum generalizations of communication complexity have been proposed; for a survey see the suggested text by G. Brassard. The first one is the qubit-communication model, where the parties can use quantum communication instead of classical communication, for example by exchanging
photons A photon () is an elementary particle that is a quantum of the electromagnetic field, including electromagnetic radiation such as light and radio waves, and the force carrier for the electromagnetic force. Photons are massless particles that ...
through an
optical fiber An optical fiber, or optical fibre, is a flexible glass or plastic fiber that can transmit light from one end to the other. Such fibers find wide usage in fiber-optic communications, where they permit transmission over longer distances and at ...
. In a second model the communication is still performed with classical bits, but the parties are allowed to manipulate an unlimited supply of quantum entangled states as part of their protocols. By doing measurements on their entangled states, the parties can save on classical communication during a distributed computation (see an application in Collapse of Randomized Communication Complexity). The third model involves access to previously shared entanglement in addition to
qubit In quantum computing, a qubit () or quantum bit is a basic unit of quantum information—the quantum version of the classic binary bit physically realized with a two-state device. A qubit is a two-state (or two-level) quantum-mechanical syste ...
communication, and is the least explored of the three quantum models.


Nondeterministic communication complexity

In nondeterministic communication complexity, Alice and Bob have access to an oracle. After receiving the oracle's word, the parties communicate to deduce f(x,y). The nondeterministic communication complexity is then the maximum over all pairs (x,y) over the sum of number of bits exchanged and the coding length of the oracle word. Viewed differently, this amounts to covering all 1-entries of the 0/1-matrix by combinatorial 1-rectangles (i.e., non-contiguous, non-convex submatrices, whose entries are all one (see Kushilevitz and Nisan or Dietzfelbinger et al.)). The nondeterministic communication complexity is the binary logarithm of the ''rectangle covering number'' of the matrix: the minimum number of combinatorial 1-rectangles required to cover all 1-entries of the matrix, without covering any 0-entries. Nondeterministic communication complexity occurs as a means to obtaining lower bounds for deterministic communication complexity (see Dietzfelbinger et al.), but also in the theory of nonnegative matrices, where it gives a lower bound on the nonnegative rank of a nonnegative matrix.


Unbounded-error communication complexity

In the unbounded-error setting, Alice and Bob have access to a private coin and their own inputs (x, y). In this setting, Alice succeeds if she responds with the correct value of f(x, y) with probability strictly greater than 1/2. In other words, if Alice's responses have ''any'' non-zero correlation to the true value of f(x, y), then the protocol is considered valid. Note that the requirement that the coin is ''private'' is essential. In particular, if the number of public bits shared between Alice and Bob are not counted against the communication complexity, it is easy to argue that computing any function has O(1) communication complexity. On the other hand, both models are equivalent if the number of public bits used by Alice and Bob is counted against the protocol's total communication. Though subtle, lower bounds on this model are extremely strong. More specifically, it is clear that any bound on problems of this class immediately imply equivalent bounds on problems in the deterministic model and the private and public coin models, but such bounds also hold immediately for nondeterministic communication models and quantum communication models. Forster was the first to prove explicit lower bounds for this class, showing that computing the inner product \langle x, y \rangle requires at least \Omega(n) bits of communication, though an earlier result of Alon, Frankl, and Rödl proved that the communication complexity for almost all Boolean functions f: \^n \times \^n \to \ is \Omega(n).


Lifting

Lifting is a general technique in complexity theory in which a lower bound on a simple measure of complexity is "lifted" to a lower bound on a more difficult measure. This technique was pioneered in the context of communication complexity by Raz and McKenzie, who proved the first query-to-communication lifting theorem, and used the result to separate the monotone NC hierarchy. Given a function f\colon \^n \to \ and a gadget g\colon \^a \times \^b \to \, their composition f \circ g\colon \^ \times \^ \to \ is defined as follows: : (f \circ g)(x,y) = f(g(x_ \cdots x_, y_ \cdots y_), \dots, g(x_ \cdots x_, y_ \cdots y_)). In words, x is partitioned into n blocks of length a, and y is partitioned into n blocks of length b. The gadget is applied n times on the blocks, and the outputs are fed into f. Diagrammatically: In this diagram, each of the inputs \mathbf_1,\dots,\mathbf_n is bits long, and each of the inputs \mathbf_1,\dots,\mathbf_n is bits long. A
decision tree A decision tree is a decision support system, decision support recursive partitioning structure that uses a Tree (graph theory), tree-like Causal model, model of decisions and their possible consequences, including probability, chance event ou ...
of depth \Delta for f can be translated to a communication protocol whose cost is \Delta \cdot D(g): each time the tree queries a bit, the corresponding value of g is computed using an optimal protocol for g. Raz and McKenzie showed that this is optimal up to a constant factor when g is the so-called "indexing gadget", in which x has length c \log n (for a large enough constant ), y has length n^c, and g(x,y) is the x-th bit of y. The proof of the Raz–McKenzie lifting theorem uses the method of simulation, in which a protocol for the composed function f \circ g is used to generate a decision tree for f. Göös, Pitassi and Watson gave an exposition of the original proof. Since then, several works have proved similar theorems with different gadgets, such as inner product. The smallest gadget which can be handled is the indexing gadget with c=1+\epsilon. Göös, Pitassi and Watson extended the Raz–McKenzie technique to randomized protocols. A simple modification of the Raz–McKenzie lifting theorem gives a lower bound of \Delta \cdot D(g) on the logarithm of the size of a protocol tree for computing f \circ g, where \Delta is the depth of the optimal decision tree for f. Garg, Göös, Kamath and Sokolov extended this to the DAG-like setting, and used their result to obtain monotone circuit lower bounds. The same technique has also yielded applications to
proof complexity In logic and theoretical computer science, and specifically proof theory and computational complexity theory, proof complexity is the field aiming to understand and analyse the computational resources that are required to prove or refute statements. ...
. A different type of lifting is exemplified by Sherstov's pattern matrix method, which gives a lower bound on the quantum communication complexity of f \circ g, where is a modified indexing gadget, in terms of the approximate degree of . The approximate degree of a Boolean function is the minimal degree of a polynomial which approximates the function on all Boolean points up to an additive error of 1/3. In contrast to the Raz–McKenzie proof which uses the method of simulation, Sherstov's proof takes a ''dual witness'' to the approximate degree of and gives a lower bound on the quantum query complexity of f \circ g using the ''generalized discrepancy method''. The dual witness for the approximate degree of is a lower bound witness for the approximate degree obtained via
LP duality LP or lp may stand for: Businesses and organizations *LP, Limited partnership in corporate law or a Limited Partner in a venture capital fund In politics *Labour Party (disambiguation), in several countries *Liberal Party, in several countries ...
. This dual witness is massaged into other objects constituting data for the generalized discrepancy method. Another example of this approach is the work of Pitassi and Robere, in which an ''algebraic gap'' is lifted to a lower bound on Razborov's ''rank measure''. The result is a strongly exponential lower bound on the monotone circuit complexity of an explicit function, obtained via the Karchmer–Wigderson characterization of monotone circuit size in terms of communication complexity.


Open problems

Considering a 0 or 1 input matrix M_f= (x,y), the minimum number of bits exchanged to compute f deterministically in the worst case, D(f), is known to be bounded from below by the logarithm of the rank of the matrix M_f. The log rank conjecture proposes that the communication complexity, D(f), is bounded from above by a constant power of the logarithm of the rank of M_f. Since D(f) is bounded from above and below by polynomials of log rank(M_f), we can say D(f) is polynomially related to log rank(M_f). Since the rank of a matrix is polynomial time computable in the size of the matrix, such an upper bound would allow the matrix's communication complexity to be approximated in polynomial time. Note, however, that the size of the matrix itself is exponential in the size of the input. For a randomized protocol, the number of bits exchanged in the worst case, R(f), was conjectured to be polynomially related to the following formula: : \log \min(\textrm(M'_f): M'_f\in \mathbb^, (M_f - M'_f)_\infty\leq 1/3). Such log rank conjectures are valuable because they reduce the question of a matrix's communication complexity to a question of linearly independent rows (columns) of the matrix. This particular version, called the Log-Approximate-Rank Conjecture, was recently refuted by Chattopadhyay, Mande and Sherif (2019) using a surprisingly simple counter-example. This reveals that the essence of the communication complexity problem, for example in the EQ case above, is figuring out where in the matrix the inputs are, in order to find out if they're equivalent.


Applications

Lower bounds in communication complexity can be used to prove lower bounds in decision tree complexity, VLSI circuits, data structures, streaming algorithms,
space–time tradeoff space–time trade-off, also known as time–memory trade-off or the algorithmic space-time continuum in computer science is a case where an algorithm or program trades increased space usage with decreased time. Here, ''space'' refers to the d ...
s for Turing machines and more. Conitzer and Sandholm studied the communication complexity of some common voting rules, which are essential in political and non political organizations. Compilation complexity is a closely related notion, which can be seen as a single-round communication complexity.


See also

* Gap-Hamming problem


Notes


References

* * * Brassard, G. Quantum communication complexity: a survey
https://arxiv.org/abs/quant-ph/0101005
* Dietzfelbinger, M., J. Hromkovic, J., and G. Schnitger,
A comparison of two lower-bound methods for communication complexity
, Theoret. Comput. Sci. 168, 1996. 39-51. * Raz, Ran. "Circuit and Communication Complexity." In Computational Complexity Theory. Steven Rudich and Avi Wigderson, eds. American Mathematical Society Institute for Advanced Study, 2004. 129-137. * A. C. Yao, "Some Complexity Questions Related to Distributed Computing", Proc. of 11th STOC, pp. 209–213, 1979. 14 * I. Newman
Private vs. Common Random Bits in Communication Complexity
Information Processing Letters 39, 1991, pp. 67–71. {{DEFAULTSORT:Communication Complexity Information theory Computational complexity theory Quantum complexity theory>p'_R(x,y) - p(x,y), \geq 0.1< \sum_ 2^ = 1 The last equality above holds because there are 2^ different pairs (x,y). Since the probability does not equal 1, there is some R_0 so that for all (x,y): :, p'_(x,y) - p(x,y), < 0.1 Since P has at most 0.1 error probability, P'_ can have at most 0.2 error probability.


Collapse of Randomized Communication Complexity

Let's say we additionally allow Alice and Bob to share some resource, for example a pair of entangled particles. Using that ressource, Alice and Bob can correlate their information and thus try to 'collapse' (or 'trivialize') communication complexity in the following sense. Definition. ''A resource R is said to be ''"collapsing"'' if, using that resource R, only one bit of classical communication is enough for Alice to know the evaluation f(x,y) in the worst case scenario for any
Boolean function In mathematics, a Boolean function is a function whose arguments and result assume values from a two-element set (usually , or ). Alternative names are switching function, used especially in older computer science literature, and truth functi ...
f. '' The surprising fact of a collapse of communication complexity is that the function f can have arbitrarily large entry size, but still the number of communication bit is constant to a single one. Some resources are shown to be non-collapsing, such as quantum correlations or more generally almost-quantum correlations, whereas on the contrary some other resources are shown to collapse randomized communication complexity, such as the PR-box, or some noisy PR-boxes satisfying some conditions.


Distributional Complexity

One approach to studying randomized communication complexity is through distributional complexity. Given a joint distribution \mu on the inputs of both players, the corresponding distributional complexity of a function f is the minimum cost of a ''deterministic'' protocol R such that \Pr (x,y) = R(x,y)\ge 2/3, where the inputs are sampled according to \mu. Yao's minimax principle (a special case of von Neumann's
minimax theorem In the mathematical area of game theory and of convex optimization, a minimax theorem is a theorem that claims that : \max_ \min_ f(x,y) = \min_ \max_f(x,y) under certain conditions on the sets X and Y and on the function f. It is always true that ...
) states that the randomized communication complexity of a function equals its maximum distributional complexity, where the maximum is taken over all joint distributions of the inputs (not necessarily product distributions!). Yao's principle can be used to prove lower bounds on the randomized communication complexity of a function: design the appropriate joint distribution, and prove a lower bound on the distributional complexity. Since distributional complexity concerns deterministic protocols, this could be easier than proving a lower bound on randomized protocols directly. As an example, let us consider the ''disjointness'' function DISJ: each of the inputs is interpreted as a subset of \, and DISJ(,)=1 if the two sets are disjoint. Razborov proved an \Omega(n) lower bound on the randomized communication complexity by considering the following distribution: with probability 3/4, sample two random disjoint sets of size n/4, and with probability 1/4, sample two random sets of size n/4 with a unique intersection.


Information Complexity

A powerful approach to the study of distributional complexity is information complexity. Initiated by Bar-Yossef, Jayram, Kumar and Sivakumar, the approach was codified in work of Barak, Braverman, Chen and Rao and by Braverman and Rao. The (internal) information complexity of a (possibly randomized) protocol with respect to a distribution is defined as follows. Let (X,Y) \sim \mu be random inputs sampled according to , and let be the transcript of when run on the inputs X,Y. The information complexity of the protocol is : \operatorname_\mu(R) = I(\Pi;Y, X) + I(\Pi;X, Y), where denotes conditional mutual information. The first summand measures the amount of information that Alice learns about Bob's input from the transcript, and the second measures the amount of information that Bob learns about Alice's input. The -error information complexity of a function with respect to a distribution is the infimal information complexity of a protocol for whose error (with respect to ) is at most . Braverman and Rao proved that information equals amortized communication. This means that the cost for solving independent copies of is roughly times the information complexity of . This is analogous to the well-known interpretation of
Shannon entropy Shannon may refer to: People * Shannon (given name) * Shannon (surname) * Shannon (American singer), stage name of singer Brenda Shannon Greene (born 1958) * Shannon (South Korean singer), British-South Korean singer and actress Shannon Arrum ...
as the amortized bit-length required to transmit data from a given information source. Braverman and Rao's proof uses a technique known as "protocol compression", in which an information-efficient protocol is "compressed" into a communication-efficient protocol. The techniques of information complexity enable the computation of the exact (up to first order) communication complexity of set disjointness to be 1.4923\ldots n. Information complexity techniques have also been used to analyze extended formulations, proving an essentially optimal lower bound on the complexity of algorithms based on
linear programming Linear programming (LP), also called linear optimization, is a method to achieve the best outcome (such as maximum profit or lowest cost) in a mathematical model whose requirements and objective are represented by linear function#As a polynomia ...
which approximately solve the maximum clique problem. Omri Weinstein's 2015 survey surveys the subject.


Quantum communication complexity

Quantum communication complexity tries to quantify the communication reduction possible by using quantum effects during a distributed computation. At least three quantum generalizations of communication complexity have been proposed; for a survey see the suggested text by G. Brassard. The first one is the qubit-communication model, where the parties can use quantum communication instead of classical communication, for example by exchanging
photons A photon () is an elementary particle that is a quantum of the electromagnetic field, including electromagnetic radiation such as light and radio waves, and the force carrier for the electromagnetic force. Photons are massless particles that ...
through an
optical fiber An optical fiber, or optical fibre, is a flexible glass or plastic fiber that can transmit light from one end to the other. Such fibers find wide usage in fiber-optic communications, where they permit transmission over longer distances and at ...
. In a second model the communication is still performed with classical bits, but the parties are allowed to manipulate an unlimited supply of quantum entangled states as part of their protocols. By doing measurements on their entangled states, the parties can save on classical communication during a distributed computation (see an application in Collapse of Randomized Communication Complexity). The third model involves access to previously shared entanglement in addition to
qubit In quantum computing, a qubit () or quantum bit is a basic unit of quantum information—the quantum version of the classic binary bit physically realized with a two-state device. A qubit is a two-state (or two-level) quantum-mechanical syste ...
communication, and is the least explored of the three quantum models.


Nondeterministic communication complexity

In nondeterministic communication complexity, Alice and Bob have access to an oracle. After receiving the oracle's word, the parties communicate to deduce f(x,y). The nondeterministic communication complexity is then the maximum over all pairs (x,y) over the sum of number of bits exchanged and the coding length of the oracle word. Viewed differently, this amounts to covering all 1-entries of the 0/1-matrix by combinatorial 1-rectangles (i.e., non-contiguous, non-convex submatrices, whose entries are all one (see Kushilevitz and Nisan or Dietzfelbinger et al.)). The nondeterministic communication complexity is the binary logarithm of the ''rectangle covering number'' of the matrix: the minimum number of combinatorial 1-rectangles required to cover all 1-entries of the matrix, without covering any 0-entries. Nondeterministic communication complexity occurs as a means to obtaining lower bounds for deterministic communication complexity (see Dietzfelbinger et al.), but also in the theory of nonnegative matrices, where it gives a lower bound on the nonnegative rank of a nonnegative matrix.


Unbounded-error communication complexity

In the unbounded-error setting, Alice and Bob have access to a private coin and their own inputs (x, y). In this setting, Alice succeeds if she responds with the correct value of f(x, y) with probability strictly greater than 1/2. In other words, if Alice's responses have ''any'' non-zero correlation to the true value of f(x, y), then the protocol is considered valid. Note that the requirement that the coin is ''private'' is essential. In particular, if the number of public bits shared between Alice and Bob are not counted against the communication complexity, it is easy to argue that computing any function has O(1) communication complexity. On the other hand, both models are equivalent if the number of public bits used by Alice and Bob is counted against the protocol's total communication. Though subtle, lower bounds on this model are extremely strong. More specifically, it is clear that any bound on problems of this class immediately imply equivalent bounds on problems in the deterministic model and the private and public coin models, but such bounds also hold immediately for nondeterministic communication models and quantum communication models. Forster was the first to prove explicit lower bounds for this class, showing that computing the inner product \langle x, y \rangle requires at least \Omega(n) bits of communication, though an earlier result of Alon, Frankl, and Rödl proved that the communication complexity for almost all Boolean functions f: \^n \times \^n \to \ is \Omega(n).


Lifting

Lifting is a general technique in complexity theory in which a lower bound on a simple measure of complexity is "lifted" to a lower bound on a more difficult measure. This technique was pioneered in the context of communication complexity by Raz and McKenzie, who proved the first query-to-communication lifting theorem, and used the result to separate the monotone NC hierarchy. Given a function f\colon \^n \to \ and a gadget g\colon \^a \times \^b \to \, their composition f \circ g\colon \^ \times \^ \to \ is defined as follows: : (f \circ g)(x,y) = f(g(x_ \cdots x_, y_ \cdots y_), \dots, g(x_ \cdots x_, y_ \cdots y_)). In words, x is partitioned into n blocks of length a, and y is partitioned into n blocks of length b. The gadget is applied n times on the blocks, and the outputs are fed into f. Diagrammatically: In this diagram, each of the inputs \mathbf_1,\dots,\mathbf_n is bits long, and each of the inputs \mathbf_1,\dots,\mathbf_n is bits long. A
decision tree A decision tree is a decision support system, decision support recursive partitioning structure that uses a Tree (graph theory), tree-like Causal model, model of decisions and their possible consequences, including probability, chance event ou ...
of depth \Delta for f can be translated to a communication protocol whose cost is \Delta \cdot D(g): each time the tree queries a bit, the corresponding value of g is computed using an optimal protocol for g. Raz and McKenzie showed that this is optimal up to a constant factor when g is the so-called "indexing gadget", in which x has length c \log n (for a large enough constant ), y has length n^c, and g(x,y) is the x-th bit of y. The proof of the Raz–McKenzie lifting theorem uses the method of simulation, in which a protocol for the composed function f \circ g is used to generate a decision tree for f. Göös, Pitassi and Watson gave an exposition of the original proof. Since then, several works have proved similar theorems with different gadgets, such as inner product. The smallest gadget which can be handled is the indexing gadget with c=1+\epsilon. Göös, Pitassi and Watson extended the Raz–McKenzie technique to randomized protocols. A simple modification of the Raz–McKenzie lifting theorem gives a lower bound of \Delta \cdot D(g) on the logarithm of the size of a protocol tree for computing f \circ g, where \Delta is the depth of the optimal decision tree for f. Garg, Göös, Kamath and Sokolov extended this to the DAG-like setting, and used their result to obtain monotone circuit lower bounds. The same technique has also yielded applications to
proof complexity In logic and theoretical computer science, and specifically proof theory and computational complexity theory, proof complexity is the field aiming to understand and analyse the computational resources that are required to prove or refute statements. ...
. A different type of lifting is exemplified by Sherstov's pattern matrix method, which gives a lower bound on the quantum communication complexity of f \circ g, where is a modified indexing gadget, in terms of the approximate degree of . The approximate degree of a Boolean function is the minimal degree of a polynomial which approximates the function on all Boolean points up to an additive error of 1/3. In contrast to the Raz–McKenzie proof which uses the method of simulation, Sherstov's proof takes a ''dual witness'' to the approximate degree of and gives a lower bound on the quantum query complexity of f \circ g using the ''generalized discrepancy method''. The dual witness for the approximate degree of is a lower bound witness for the approximate degree obtained via
LP duality LP or lp may stand for: Businesses and organizations *LP, Limited partnership in corporate law or a Limited Partner in a venture capital fund In politics *Labour Party (disambiguation), in several countries *Liberal Party, in several countries ...
. This dual witness is massaged into other objects constituting data for the generalized discrepancy method. Another example of this approach is the work of Pitassi and Robere, in which an ''algebraic gap'' is lifted to a lower bound on Razborov's ''rank measure''. The result is a strongly exponential lower bound on the monotone circuit complexity of an explicit function, obtained via the Karchmer–Wigderson characterization of monotone circuit size in terms of communication complexity.


Open problems

Considering a 0 or 1 input matrix M_f= (x,y), the minimum number of bits exchanged to compute f deterministically in the worst case, D(f), is known to be bounded from below by the logarithm of the rank of the matrix M_f. The log rank conjecture proposes that the communication complexity, D(f), is bounded from above by a constant power of the logarithm of the rank of M_f. Since D(f) is bounded from above and below by polynomials of log rank(M_f), we can say D(f) is polynomially related to log rank(M_f). Since the rank of a matrix is polynomial time computable in the size of the matrix, such an upper bound would allow the matrix's communication complexity to be approximated in polynomial time. Note, however, that the size of the matrix itself is exponential in the size of the input. For a randomized protocol, the number of bits exchanged in the worst case, R(f), was conjectured to be polynomially related to the following formula: : \log \min(\textrm(M'_f): M'_f\in \mathbb^, (M_f - M'_f)_\infty\leq 1/3). Such log rank conjectures are valuable because they reduce the question of a matrix's communication complexity to a question of linearly independent rows (columns) of the matrix. This particular version, called the Log-Approximate-Rank Conjecture, was recently refuted by Chattopadhyay, Mande and Sherif (2019) using a surprisingly simple counter-example. This reveals that the essence of the communication complexity problem, for example in the EQ case above, is figuring out where in the matrix the inputs are, in order to find out if they're equivalent.


Applications

Lower bounds in communication complexity can be used to prove lower bounds in decision tree complexity, VLSI circuits, data structures, streaming algorithms,
space–time tradeoff space–time trade-off, also known as time–memory trade-off or the algorithmic space-time continuum in computer science is a case where an algorithm or program trades increased space usage with decreased time. Here, ''space'' refers to the d ...
s for Turing machines and more. Conitzer and Sandholm studied the communication complexity of some common voting rules, which are essential in political and non political organizations. Compilation complexity is a closely related notion, which can be seen as a single-round communication complexity.


See also

* Gap-Hamming problem


Notes


References

* * * Brassard, G. Quantum communication complexity: a survey
https://arxiv.org/abs/quant-ph/0101005
* Dietzfelbinger, M., J. Hromkovic, J., and G. Schnitger,
A comparison of two lower-bound methods for communication complexity
, Theoret. Comput. Sci. 168, 1996. 39-51. * Raz, Ran. "Circuit and Communication Complexity." In Computational Complexity Theory. Steven Rudich and Avi Wigderson, eds. American Mathematical Society Institute for Advanced Study, 2004. 129-137. * A. C. Yao, "Some Complexity Questions Related to Distributed Computing", Proc. of 11th STOC, pp. 209–213, 1979. 14 * I. Newman
Private vs. Common Random Bits in Communication Complexity
Information Processing Letters 39, 1991, pp. 67–71. {{DEFAULTSORT:Communication Complexity Information theory Computational complexity theory Quantum complexity theory>p'_R(x,y) - p(x,y), \geq 0.1< \sum_ 2^ = 1 The last equality above holds because there are 2^ different pairs (x,y). Since the probability does not equal 1, there is some R_0 so that for all (x,y): :, p'_(x,y) - p(x,y), < 0.1 Since P has at most 0.1 error probability, P'_ can have at most 0.2 error probability.


Collapse of Randomized Communication Complexity

Let's say we additionally allow Alice and Bob to share some resource, for example a pair of entangled particles. Using that ressource, Alice and Bob can correlate their information and thus try to 'collapse' (or 'trivialize') communication complexity in the following sense. Definition. ''A resource R is said to be ''"collapsing"'' if, using that resource R, only one bit of classical communication is enough for Alice to know the evaluation f(x,y) in the worst case scenario for any
Boolean function In mathematics, a Boolean function is a function whose arguments and result assume values from a two-element set (usually , or ). Alternative names are switching function, used especially in older computer science literature, and truth functi ...
f. '' The surprising fact of a collapse of communication complexity is that the function f can have arbitrarily large entry size, but still the number of communication bit is constant to a single one. Some resources are shown to be non-collapsing, such as quantum correlations or more generally almost-quantum correlations, whereas on the contrary some other resources are shown to collapse randomized communication complexity, such as the PR-box, or some noisy PR-boxes satisfying some conditions.


Distributional Complexity

One approach to studying randomized communication complexity is through distributional complexity. Given a joint distribution \mu on the inputs of both players, the corresponding distributional complexity of a function f is the minimum cost of a ''deterministic'' protocol R such that \Pr (x,y) = R(x,y)\ge 2/3, where the inputs are sampled according to \mu. Yao's minimax principle (a special case of von Neumann's
minimax theorem In the mathematical area of game theory and of convex optimization, a minimax theorem is a theorem that claims that : \max_ \min_ f(x,y) = \min_ \max_f(x,y) under certain conditions on the sets X and Y and on the function f. It is always true that ...
) states that the randomized communication complexity of a function equals its maximum distributional complexity, where the maximum is taken over all joint distributions of the inputs (not necessarily product distributions!). Yao's principle can be used to prove lower bounds on the randomized communication complexity of a function: design the appropriate joint distribution, and prove a lower bound on the distributional complexity. Since distributional complexity concerns deterministic protocols, this could be easier than proving a lower bound on randomized protocols directly. As an example, let us consider the ''disjointness'' function DISJ: each of the inputs is interpreted as a subset of \, and DISJ(,)=1 if the two sets are disjoint. Razborov proved an \Omega(n) lower bound on the randomized communication complexity by considering the following distribution: with probability 3/4, sample two random disjoint sets of size n/4, and with probability 1/4, sample two random sets of size n/4 with a unique intersection.


Information Complexity

A powerful approach to the study of distributional complexity is information complexity. Initiated by Bar-Yossef, Jayram, Kumar and Sivakumar, the approach was codified in work of Barak, Braverman, Chen and Rao and by Braverman and Rao. The (internal) information complexity of a (possibly randomized) protocol with respect to a distribution is defined as follows. Let (X,Y) \sim \mu be random inputs sampled according to , and let be the transcript of when run on the inputs X,Y. The information complexity of the protocol is : \operatorname_\mu(R) = I(\Pi;Y, X) + I(\Pi;X, Y), where denotes conditional mutual information. The first summand measures the amount of information that Alice learns about Bob's input from the transcript, and the second measures the amount of information that Bob learns about Alice's input. The -error information complexity of a function with respect to a distribution is the infimal information complexity of a protocol for whose error (with respect to ) is at most . Braverman and Rao proved that information equals amortized communication. This means that the cost for solving independent copies of is roughly times the information complexity of . This is analogous to the well-known interpretation of
Shannon entropy Shannon may refer to: People * Shannon (given name) * Shannon (surname) * Shannon (American singer), stage name of singer Brenda Shannon Greene (born 1958) * Shannon (South Korean singer), British-South Korean singer and actress Shannon Arrum ...
as the amortized bit-length required to transmit data from a given information source. Braverman and Rao's proof uses a technique known as "protocol compression", in which an information-efficient protocol is "compressed" into a communication-efficient protocol. The techniques of information complexity enable the computation of the exact (up to first order) communication complexity of set disjointness to be 1.4923\ldots n. Information complexity techniques have also been used to analyze extended formulations, proving an essentially optimal lower bound on the complexity of algorithms based on
linear programming Linear programming (LP), also called linear optimization, is a method to achieve the best outcome (such as maximum profit or lowest cost) in a mathematical model whose requirements and objective are represented by linear function#As a polynomia ...
which approximately solve the maximum clique problem. Omri Weinstein's 2015 survey surveys the subject.


Quantum communication complexity

Quantum communication complexity tries to quantify the communication reduction possible by using quantum effects during a distributed computation. At least three quantum generalizations of communication complexity have been proposed; for a survey see the suggested text by G. Brassard. The first one is the qubit-communication model, where the parties can use quantum communication instead of classical communication, for example by exchanging
photons A photon () is an elementary particle that is a quantum of the electromagnetic field, including electromagnetic radiation such as light and radio waves, and the force carrier for the electromagnetic force. Photons are massless particles that ...
through an
optical fiber An optical fiber, or optical fibre, is a flexible glass or plastic fiber that can transmit light from one end to the other. Such fibers find wide usage in fiber-optic communications, where they permit transmission over longer distances and at ...
. In a second model the communication is still performed with classical bits, but the parties are allowed to manipulate an unlimited supply of quantum entangled states as part of their protocols. By doing measurements on their entangled states, the parties can save on classical communication during a distributed computation (see an application in Collapse of Randomized Communication Complexity). The third model involves access to previously shared entanglement in addition to
qubit In quantum computing, a qubit () or quantum bit is a basic unit of quantum information—the quantum version of the classic binary bit physically realized with a two-state device. A qubit is a two-state (or two-level) quantum-mechanical syste ...
communication, and is the least explored of the three quantum models.


Nondeterministic communication complexity

In nondeterministic communication complexity, Alice and Bob have access to an oracle. After receiving the oracle's word, the parties communicate to deduce f(x,y). The nondeterministic communication complexity is then the maximum over all pairs (x,y) over the sum of number of bits exchanged and the coding length of the oracle word. Viewed differently, this amounts to covering all 1-entries of the 0/1-matrix by combinatorial 1-rectangles (i.e., non-contiguous, non-convex submatrices, whose entries are all one (see Kushilevitz and Nisan or Dietzfelbinger et al.)). The nondeterministic communication complexity is the binary logarithm of the ''rectangle covering number'' of the matrix: the minimum number of combinatorial 1-rectangles required to cover all 1-entries of the matrix, without covering any 0-entries. Nondeterministic communication complexity occurs as a means to obtaining lower bounds for deterministic communication complexity (see Dietzfelbinger et al.), but also in the theory of nonnegative matrices, where it gives a lower bound on the nonnegative rank of a nonnegative matrix.


Unbounded-error communication complexity

In the unbounded-error setting, Alice and Bob have access to a private coin and their own inputs (x, y). In this setting, Alice succeeds if she responds with the correct value of f(x, y) with probability strictly greater than 1/2. In other words, if Alice's responses have ''any'' non-zero correlation to the true value of f(x, y), then the protocol is considered valid. Note that the requirement that the coin is ''private'' is essential. In particular, if the number of public bits shared between Alice and Bob are not counted against the communication complexity, it is easy to argue that computing any function has O(1) communication complexity. On the other hand, both models are equivalent if the number of public bits used by Alice and Bob is counted against the protocol's total communication. Though subtle, lower bounds on this model are extremely strong. More specifically, it is clear that any bound on problems of this class immediately imply equivalent bounds on problems in the deterministic model and the private and public coin models, but such bounds also hold immediately for nondeterministic communication models and quantum communication models. Forster was the first to prove explicit lower bounds for this class, showing that computing the inner product \langle x, y \rangle requires at least \Omega(n) bits of communication, though an earlier result of Alon, Frankl, and Rödl proved that the communication complexity for almost all Boolean functions f: \^n \times \^n \to \ is \Omega(n).


Lifting

Lifting is a general technique in complexity theory in which a lower bound on a simple measure of complexity is "lifted" to a lower bound on a more difficult measure. This technique was pioneered in the context of communication complexity by Raz and McKenzie, who proved the first query-to-communication lifting theorem, and used the result to separate the monotone NC hierarchy. Given a function f\colon \^n \to \ and a gadget g\colon \^a \times \^b \to \, their composition f \circ g\colon \^ \times \^ \to \ is defined as follows: : (f \circ g)(x,y) = f(g(x_ \cdots x_, y_ \cdots y_), \dots, g(x_ \cdots x_, y_ \cdots y_)). In words, x is partitioned into n blocks of length a, and y is partitioned into n blocks of length b. The gadget is applied n times on the blocks, and the outputs are fed into f. Diagrammatically: In this diagram, each of the inputs \mathbf_1,\dots,\mathbf_n is bits long, and each of the inputs \mathbf_1,\dots,\mathbf_n is bits long. A
decision tree A decision tree is a decision support system, decision support recursive partitioning structure that uses a Tree (graph theory), tree-like Causal model, model of decisions and their possible consequences, including probability, chance event ou ...
of depth \Delta for f can be translated to a communication protocol whose cost is \Delta \cdot D(g): each time the tree queries a bit, the corresponding value of g is computed using an optimal protocol for g. Raz and McKenzie showed that this is optimal up to a constant factor when g is the so-called "indexing gadget", in which x has length c \log n (for a large enough constant ), y has length n^c, and g(x,y) is the x-th bit of y. The proof of the Raz–McKenzie lifting theorem uses the method of simulation, in which a protocol for the composed function f \circ g is used to generate a decision tree for f. Göös, Pitassi and Watson gave an exposition of the original proof. Since then, several works have proved similar theorems with different gadgets, such as inner product. The smallest gadget which can be handled is the indexing gadget with c=1+\epsilon. Göös, Pitassi and Watson extended the Raz–McKenzie technique to randomized protocols. A simple modification of the Raz–McKenzie lifting theorem gives a lower bound of \Delta \cdot D(g) on the logarithm of the size of a protocol tree for computing f \circ g, where \Delta is the depth of the optimal decision tree for f. Garg, Göös, Kamath and Sokolov extended this to the DAG-like setting, and used their result to obtain monotone circuit lower bounds. The same technique has also yielded applications to
proof complexity In logic and theoretical computer science, and specifically proof theory and computational complexity theory, proof complexity is the field aiming to understand and analyse the computational resources that are required to prove or refute statements. ...
. A different type of lifting is exemplified by Sherstov's pattern matrix method, which gives a lower bound on the quantum communication complexity of f \circ g, where is a modified indexing gadget, in terms of the approximate degree of . The approximate degree of a Boolean function is the minimal degree of a polynomial which approximates the function on all Boolean points up to an additive error of 1/3. In contrast to the Raz–McKenzie proof which uses the method of simulation, Sherstov's proof takes a ''dual witness'' to the approximate degree of and gives a lower bound on the quantum query complexity of f \circ g using the ''generalized discrepancy method''. The dual witness for the approximate degree of is a lower bound witness for the approximate degree obtained via
LP duality LP or lp may stand for: Businesses and organizations *LP, Limited partnership in corporate law or a Limited Partner in a venture capital fund In politics *Labour Party (disambiguation), in several countries *Liberal Party, in several countries ...
. This dual witness is massaged into other objects constituting data for the generalized discrepancy method. Another example of this approach is the work of Pitassi and Robere, in which an ''algebraic gap'' is lifted to a lower bound on Razborov's ''rank measure''. The result is a strongly exponential lower bound on the monotone circuit complexity of an explicit function, obtained via the Karchmer–Wigderson characterization of monotone circuit size in terms of communication complexity.


Open problems

Considering a 0 or 1 input matrix M_f= (x,y), the minimum number of bits exchanged to compute f deterministically in the worst case, D(f), is known to be bounded from below by the logarithm of the rank of the matrix M_f. The log rank conjecture proposes that the communication complexity, D(f), is bounded from above by a constant power of the logarithm of the rank of M_f. Since D(f) is bounded from above and below by polynomials of log rank(M_f), we can say D(f) is polynomially related to log rank(M_f). Since the rank of a matrix is polynomial time computable in the size of the matrix, such an upper bound would allow the matrix's communication complexity to be approximated in polynomial time. Note, however, that the size of the matrix itself is exponential in the size of the input. For a randomized protocol, the number of bits exchanged in the worst case, R(f), was conjectured to be polynomially related to the following formula: : \log \min(\textrm(M'_f): M'_f\in \mathbb^, (M_f - M'_f)_\infty\leq 1/3). Such log rank conjectures are valuable because they reduce the question of a matrix's communication complexity to a question of linearly independent rows (columns) of the matrix. This particular version, called the Log-Approximate-Rank Conjecture, was recently refuted by Chattopadhyay, Mande and Sherif (2019) using a surprisingly simple counter-example. This reveals that the essence of the communication complexity problem, for example in the EQ case above, is figuring out where in the matrix the inputs are, in order to find out if they're equivalent.


Applications

Lower bounds in communication complexity can be used to prove lower bounds in decision tree complexity, VLSI circuits, data structures, streaming algorithms,
space–time tradeoff space–time trade-off, also known as time–memory trade-off or the algorithmic space-time continuum in computer science is a case where an algorithm or program trades increased space usage with decreased time. Here, ''space'' refers to the d ...
s for Turing machines and more. Conitzer and Sandholm studied the communication complexity of some common voting rules, which are essential in political and non political organizations. Compilation complexity is a closely related notion, which can be seen as a single-round communication complexity.


See also

* Gap-Hamming problem


Notes


References

* * * Brassard, G. Quantum communication complexity: a survey
https://arxiv.org/abs/quant-ph/0101005
* Dietzfelbinger, M., J. Hromkovic, J., and G. Schnitger,
A comparison of two lower-bound methods for communication complexity
, Theoret. Comput. Sci. 168, 1996. 39-51. * Raz, Ran. "Circuit and Communication Complexity." In Computational Complexity Theory. Steven Rudich and Avi Wigderson, eds. American Mathematical Society Institute for Advanced Study, 2004. 129-137. * A. C. Yao, "Some Complexity Questions Related to Distributed Computing", Proc. of 11th STOC, pp. 209–213, 1979. 14 * I. Newman
Private vs. Common Random Bits in Communication Complexity
Information Processing Letters 39, 1991, pp. 67–71. {{DEFAULTSORT:Communication Complexity Information theory Computational complexity theory Quantum complexity theory>p'_R(x,y) - p(x,y), \geq 0.1< \sum_ 2^ = 1 The last equality above holds because there are 2^ different pairs (x,y). Since the probability does not equal 1, there is some R_0 so that for all (x,y): :, p'_(x,y) - p(x,y), < 0.1 Since P has at most 0.1 error probability, P'_ can have at most 0.2 error probability.


Collapse of Randomized Communication Complexity

Let's say we additionally allow Alice and Bob to share some resource, for example a pair of entangled particles. Using that ressource, Alice and Bob can correlate their information and thus try to 'collapse' (or 'trivialize') communication complexity in the following sense. Definition. ''A resource R is said to be ''"collapsing"'' if, using that resource R, only one bit of classical communication is enough for Alice to know the evaluation f(x,y) in the worst case scenario for any
Boolean function In mathematics, a Boolean function is a function whose arguments and result assume values from a two-element set (usually , or ). Alternative names are switching function, used especially in older computer science literature, and truth functi ...
f. '' The surprising fact of a collapse of communication complexity is that the function f can have arbitrarily large entry size, but still the number of communication bit is constant to a single one. Some resources are shown to be non-collapsing, such as quantum correlations or more generally almost-quantum correlations, whereas on the contrary some other resources are shown to collapse randomized communication complexity, such as the PR-box, or some noisy PR-boxes satisfying some conditions.


Distributional Complexity

One approach to studying randomized communication complexity is through distributional complexity. Given a joint distribution \mu on the inputs of both players, the corresponding distributional complexity of a function f is the minimum cost of a ''deterministic'' protocol R such that \Pr (x,y) = R(x,y)\ge 2/3, where the inputs are sampled according to \mu. Yao's minimax principle (a special case of von Neumann's
minimax theorem In the mathematical area of game theory and of convex optimization, a minimax theorem is a theorem that claims that : \max_ \min_ f(x,y) = \min_ \max_f(x,y) under certain conditions on the sets X and Y and on the function f. It is always true that ...
) states that the randomized communication complexity of a function equals its maximum distributional complexity, where the maximum is taken over all joint distributions of the inputs (not necessarily product distributions!). Yao's principle can be used to prove lower bounds on the randomized communication complexity of a function: design the appropriate joint distribution, and prove a lower bound on the distributional complexity. Since distributional complexity concerns deterministic protocols, this could be easier than proving a lower bound on randomized protocols directly. As an example, let us consider the ''disjointness'' function DISJ: each of the inputs is interpreted as a subset of \, and DISJ(,)=1 if the two sets are disjoint. Razborov proved an \Omega(n) lower bound on the randomized communication complexity by considering the following distribution: with probability 3/4, sample two random disjoint sets of size n/4, and with probability 1/4, sample two random sets of size n/4 with a unique intersection.


Information Complexity

A powerful approach to the study of distributional complexity is information complexity. Initiated by Bar-Yossef, Jayram, Kumar and Sivakumar, the approach was codified in work of Barak, Braverman, Chen and Rao and by Braverman and Rao. The (internal) information complexity of a (possibly randomized) protocol with respect to a distribution is defined as follows. Let (X,Y) \sim \mu be random inputs sampled according to , and let be the transcript of when run on the inputs X,Y. The information complexity of the protocol is : \operatorname_\mu(R) = I(\Pi;Y, X) + I(\Pi;X, Y), where denotes conditional mutual information. The first summand measures the amount of information that Alice learns about Bob's input from the transcript, and the second measures the amount of information that Bob learns about Alice's input. The -error information complexity of a function with respect to a distribution is the infimal information complexity of a protocol for whose error (with respect to ) is at most . Braverman and Rao proved that information equals amortized communication. This means that the cost for solving independent copies of is roughly times the information complexity of . This is analogous to the well-known interpretation of
Shannon entropy Shannon may refer to: People * Shannon (given name) * Shannon (surname) * Shannon (American singer), stage name of singer Brenda Shannon Greene (born 1958) * Shannon (South Korean singer), British-South Korean singer and actress Shannon Arrum ...
as the amortized bit-length required to transmit data from a given information source. Braverman and Rao's proof uses a technique known as "protocol compression", in which an information-efficient protocol is "compressed" into a communication-efficient protocol. The techniques of information complexity enable the computation of the exact (up to first order) communication complexity of set disjointness to be 1.4923\ldots n. Information complexity techniques have also been used to analyze extended formulations, proving an essentially optimal lower bound on the complexity of algorithms based on
linear programming Linear programming (LP), also called linear optimization, is a method to achieve the best outcome (such as maximum profit or lowest cost) in a mathematical model whose requirements and objective are represented by linear function#As a polynomia ...
which approximately solve the maximum clique problem. Omri Weinstein's 2015 survey surveys the subject.


Quantum communication complexity

Quantum communication complexity tries to quantify the communication reduction possible by using quantum effects during a distributed computation. At least three quantum generalizations of communication complexity have been proposed; for a survey see the suggested text by G. Brassard. The first one is the qubit-communication model, where the parties can use quantum communication instead of classical communication, for example by exchanging
photons A photon () is an elementary particle that is a quantum of the electromagnetic field, including electromagnetic radiation such as light and radio waves, and the force carrier for the electromagnetic force. Photons are massless particles that ...
through an
optical fiber An optical fiber, or optical fibre, is a flexible glass or plastic fiber that can transmit light from one end to the other. Such fibers find wide usage in fiber-optic communications, where they permit transmission over longer distances and at ...
. In a second model the communication is still performed with classical bits, but the parties are allowed to manipulate an unlimited supply of quantum entangled states as part of their protocols. By doing measurements on their entangled states, the parties can save on classical communication during a distributed computation (see an application in Collapse of Randomized Communication Complexity). The third model involves access to previously shared entanglement in addition to
qubit In quantum computing, a qubit () or quantum bit is a basic unit of quantum information—the quantum version of the classic binary bit physically realized with a two-state device. A qubit is a two-state (or two-level) quantum-mechanical syste ...
communication, and is the least explored of the three quantum models.


Nondeterministic communication complexity

In nondeterministic communication complexity, Alice and Bob have access to an oracle. After receiving the oracle's word, the parties communicate to deduce f(x,y). The nondeterministic communication complexity is then the maximum over all pairs (x,y) over the sum of number of bits exchanged and the coding length of the oracle word. Viewed differently, this amounts to covering all 1-entries of the 0/1-matrix by combinatorial 1-rectangles (i.e., non-contiguous, non-convex submatrices, whose entries are all one (see Kushilevitz and Nisan or Dietzfelbinger et al.)). The nondeterministic communication complexity is the binary logarithm of the ''rectangle covering number'' of the matrix: the minimum number of combinatorial 1-rectangles required to cover all 1-entries of the matrix, without covering any 0-entries. Nondeterministic communication complexity occurs as a means to obtaining lower bounds for deterministic communication complexity (see Dietzfelbinger et al.), but also in the theory of nonnegative matrices, where it gives a lower bound on the nonnegative rank of a nonnegative matrix.


Unbounded-error communication complexity

In the unbounded-error setting, Alice and Bob have access to a private coin and their own inputs (x, y). In this setting, Alice succeeds if she responds with the correct value of f(x, y) with probability strictly greater than 1/2. In other words, if Alice's responses have ''any'' non-zero correlation to the true value of f(x, y), then the protocol is considered valid. Note that the requirement that the coin is ''private'' is essential. In particular, if the number of public bits shared between Alice and Bob are not counted against the communication complexity, it is easy to argue that computing any function has O(1) communication complexity. On the other hand, both models are equivalent if the number of public bits used by Alice and Bob is counted against the protocol's total communication. Though subtle, lower bounds on this model are extremely strong. More specifically, it is clear that any bound on problems of this class immediately imply equivalent bounds on problems in the deterministic model and the private and public coin models, but such bounds also hold immediately for nondeterministic communication models and quantum communication models. Forster was the first to prove explicit lower bounds for this class, showing that computing the inner product \langle x, y \rangle requires at least \Omega(n) bits of communication, though an earlier result of Alon, Frankl, and Rödl proved that the communication complexity for almost all Boolean functions f: \^n \times \^n \to \ is \Omega(n).


Lifting

Lifting is a general technique in complexity theory in which a lower bound on a simple measure of complexity is "lifted" to a lower bound on a more difficult measure. This technique was pioneered in the context of communication complexity by Raz and McKenzie, who proved the first query-to-communication lifting theorem, and used the result to separate the monotone NC hierarchy. Given a function f\colon \^n \to \ and a gadget g\colon \^a \times \^b \to \, their composition f \circ g\colon \^ \times \^ \to \ is defined as follows: : (f \circ g)(x,y) = f(g(x_ \cdots x_, y_ \cdots y_), \dots, g(x_ \cdots x_, y_ \cdots y_)). In words, x is partitioned into n blocks of length a, and y is partitioned into n blocks of length b. The gadget is applied n times on the blocks, and the outputs are fed into f. Diagrammatically: In this diagram, each of the inputs \mathbf_1,\dots,\mathbf_n is bits long, and each of the inputs \mathbf_1,\dots,\mathbf_n is bits long. A
decision tree A decision tree is a decision support system, decision support recursive partitioning structure that uses a Tree (graph theory), tree-like Causal model, model of decisions and their possible consequences, including probability, chance event ou ...
of depth \Delta for f can be translated to a communication protocol whose cost is \Delta \cdot D(g): each time the tree queries a bit, the corresponding value of g is computed using an optimal protocol for g. Raz and McKenzie showed that this is optimal up to a constant factor when g is the so-called "indexing gadget", in which x has length c \log n (for a large enough constant ), y has length n^c, and g(x,y) is the x-th bit of y. The proof of the Raz–McKenzie lifting theorem uses the method of simulation, in which a protocol for the composed function f \circ g is used to generate a decision tree for f. Göös, Pitassi and Watson gave an exposition of the original proof. Since then, several works have proved similar theorems with different gadgets, such as inner product. The smallest gadget which can be handled is the indexing gadget with c=1+\epsilon. Göös, Pitassi and Watson extended the Raz–McKenzie technique to randomized protocols. A simple modification of the Raz–McKenzie lifting theorem gives a lower bound of \Delta \cdot D(g) on the logarithm of the size of a protocol tree for computing f \circ g, where \Delta is the depth of the optimal decision tree for f. Garg, Göös, Kamath and Sokolov extended this to the DAG-like setting, and used their result to obtain monotone circuit lower bounds. The same technique has also yielded applications to
proof complexity In logic and theoretical computer science, and specifically proof theory and computational complexity theory, proof complexity is the field aiming to understand and analyse the computational resources that are required to prove or refute statements. ...
. A different type of lifting is exemplified by Sherstov's pattern matrix method, which gives a lower bound on the quantum communication complexity of f \circ g, where is a modified indexing gadget, in terms of the approximate degree of . The approximate degree of a Boolean function is the minimal degree of a polynomial which approximates the function on all Boolean points up to an additive error of 1/3. In contrast to the Raz–McKenzie proof which uses the method of simulation, Sherstov's proof takes a ''dual witness'' to the approximate degree of and gives a lower bound on the quantum query complexity of f \circ g using the ''generalized discrepancy method''. The dual witness for the approximate degree of is a lower bound witness for the approximate degree obtained via
LP duality LP or lp may stand for: Businesses and organizations *LP, Limited partnership in corporate law or a Limited Partner in a venture capital fund In politics *Labour Party (disambiguation), in several countries *Liberal Party, in several countries ...
. This dual witness is massaged into other objects constituting data for the generalized discrepancy method. Another example of this approach is the work of Pitassi and Robere, in which an ''algebraic gap'' is lifted to a lower bound on Razborov's ''rank measure''. The result is a strongly exponential lower bound on the monotone circuit complexity of an explicit function, obtained via the Karchmer–Wigderson characterization of monotone circuit size in terms of communication complexity.


Open problems

Considering a 0 or 1 input matrix M_f= (x,y), the minimum number of bits exchanged to compute f deterministically in the worst case, D(f), is known to be bounded from below by the logarithm of the rank of the matrix M_f. The log rank conjecture proposes that the communication complexity, D(f), is bounded from above by a constant power of the logarithm of the rank of M_f. Since D(f) is bounded from above and below by polynomials of log rank(M_f), we can say D(f) is polynomially related to log rank(M_f). Since the rank of a matrix is polynomial time computable in the size of the matrix, such an upper bound would allow the matrix's communication complexity to be approximated in polynomial time. Note, however, that the size of the matrix itself is exponential in the size of the input. For a randomized protocol, the number of bits exchanged in the worst case, R(f), was conjectured to be polynomially related to the following formula: : \log \min(\textrm(M'_f): M'_f\in \mathbb^, (M_f - M'_f)_\infty\leq 1/3). Such log rank conjectures are valuable because they reduce the question of a matrix's communication complexity to a question of linearly independent rows (columns) of the matrix. This particular version, called the Log-Approximate-Rank Conjecture, was recently refuted by Chattopadhyay, Mande and Sherif (2019) using a surprisingly simple counter-example. This reveals that the essence of the communication complexity problem, for example in the EQ case above, is figuring out where in the matrix the inputs are, in order to find out if they're equivalent.


Applications

Lower bounds in communication complexity can be used to prove lower bounds in decision tree complexity, VLSI circuits, data structures, streaming algorithms,
space–time tradeoff space–time trade-off, also known as time–memory trade-off or the algorithmic space-time continuum in computer science is a case where an algorithm or program trades increased space usage with decreased time. Here, ''space'' refers to the d ...
s for Turing machines and more. Conitzer and Sandholm studied the communication complexity of some common voting rules, which are essential in political and non political organizations. Compilation complexity is a closely related notion, which can be seen as a single-round communication complexity.


See also

* Gap-Hamming problem


Notes


References

* * * Brassard, G. Quantum communication complexity: a survey
https://arxiv.org/abs/quant-ph/0101005
* Dietzfelbinger, M., J. Hromkovic, J., and G. Schnitger,
A comparison of two lower-bound methods for communication complexity
, Theoret. Comput. Sci. 168, 1996. 39-51. * Raz, Ran. "Circuit and Communication Complexity." In Computational Complexity Theory. Steven Rudich and Avi Wigderson, eds. American Mathematical Society Institute for Advanced Study, 2004. 129-137. * A. C. Yao, "Some Complexity Questions Related to Distributed Computing", Proc. of 11th STOC, pp. 209–213, 1979. 14 * I. Newman
Private vs. Common Random Bits in Communication Complexity
Information Processing Letters 39, 1991, pp. 67–71. {{DEFAULTSORT:Communication Complexity Information theory Computational complexity theory Quantum complexity theory