Expander Walk Sampling
   HOME

TheInfoList



OR:

In the
mathematical Mathematics is an area of knowledge that includes the topics of numbers, formulas and related structures, shapes and the spaces in which they are contained, and quantities and their changes. These topics are represented in modern mathematics ...
discipline of
graph theory In mathematics, graph theory is the study of ''graphs'', which are mathematical structures used to model pairwise relations between objects. A graph in this context is made up of '' vertices'' (also called ''nodes'' or ''points'') which are conne ...
, the expander walk sampling theorem intuitively states that sampling vertices in an
expander graph In graph theory, an expander graph is a sparse graph that has strong connectivity properties, quantified using vertex, edge or spectral expansion. Expander constructions have spawned research in pure and applied mathematics, with several applicati ...
by doing relatively short
random walk In mathematics, a random walk is a random process that describes a path that consists of a succession of random steps on some mathematical space. An elementary example of a random walk is the random walk on the integer number line \mathbb Z ...
can simulate sampling the vertices independently from a
uniform distribution Uniform distribution may refer to: * Continuous uniform distribution * Discrete uniform distribution * Uniform distribution (ecology) * Equidistributed sequence In mathematics, a sequence (''s''1, ''s''2, ''s''3, ...) of real numbers is said to be ...
. The earliest version of this theorem is due to , and the more general version is typically attributed to .


Statement

Let G=(V,E) be an n-vertex
expander graph In graph theory, an expander graph is a sparse graph that has strong connectivity properties, quantified using vertex, edge or spectral expansion. Expander constructions have spawned research in pure and applied mathematics, with several applicati ...
with positively weighted edges, and let A\subset V. Let P denote the
stochastic matrix In mathematics, a stochastic matrix is a square matrix used to describe the transitions of a Markov chain. Each of its entries is a nonnegative real number representing a probability. It is also called a probability matrix, transition matrix, ...
of the graph, and let \lambda_2 be the second largest
eigenvalue In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted b ...
of P. Let y_0, y_1, \ldots, y_ denote the vertices encountered in a (k-1)-step random walk on G starting at vertex y_0, and let \pi (A):= \lim_ \frac \sum_^ \mathbf_A(y_i). Where \mathbf_A(y)\begin 1, & \texty \in A \\ 0, & \text\end (It is well known that almost all trajectories y_0, y_1, \ldots, y_ converges to some limiting point, \pi (A), as k \rightarrow \infty .) The theorem states that for a weighted graph G=(V,E) and a random walk y_0, y_1, \ldots, y_ where y_0 is chosen by an initial distribution \mathbf, for all \gamma > 0, we have the following bound: :\Pr\left \frac \sum_^ \mathbf_A(y_i) - \pi(A)\bigg, \geq \gamma\right\leq C e^. Where C is dependant on \mathbf, G and A . The theorem gives a bound for the rate of convergence to \pi(A) with respect to the length of the random walk, hence giving a more efficient method to estimate \pi(A) compared to independent sampling the vertices of G.


Proof

In order to prove the theorem, we provide a few definitions followed by three lemmas. Let \it_ be the weight of the edge xy\in E(G) and let \it_x=\sum_\it_. Denote by \pi(x):=\it_x/\sum_ \it_y. Let \frac be the matrix with entries\frac , and let N_=, , \frac, , _. Let D=\text(1/\it_i ) and M=(\it_). Let P(r)=PE_r where P is the stochastic matrix, E_r=\text(e^) and r \ge 0 . Then: :P = \sqrtS\sqrt \qquad \text \qquad P(r) = \sqrtS(r)\sqrt Where S:=\sqrtM\sqrt \text S(r) := \sqrtM\sqrt. As S and S(r) are symmetric, they have real eigenvalues. Therefore, as the eigenvalues of S(r) and P(r) are equal, the eigenvalues of P(r) are real. Let \lambda(r) and \lambda_2(r) be the first and second largest eigenvalue of P(r) respectively. For convenience of notation, let t_k=\frac \sum_^ \mathbf_A(y_i), \epsilon=\lambda-\lambda_2 , \epsilon_r=\lambda(r)-\lambda_2(r) , and let \mathbf be the all-1 vector. Lemma 1 \Pr\left _k- \pi(A) \ge \gamma\right\leq e^(\mathbfP(r)^k\mathbf)/\lambda(r)^k Proof: By Markov’s inequality, \begin \Pr\left _k \ge \pi(A) +\gamma\right=\Pr ^\ge e^leq e^E_\mathbfe^ \end Where E_\mathbf is the expectation of x_0 chosen according to the probability distribution \mathbf. As this can be interpreted by summing over all possible trajectories x_0,x_1,.. .,x_k, hence: E_e^=\sum_e^\mathbb(x_0)\Pi_^kp_=\mathbfP(r)^k\mathbf Combining the two results proves the lemma. Lemma 2 For 0\le r \le 1, :(\mathbfP(r)^k\mathbf)/\lambda(r)^k\le (1+r)N_ Proof: As eigenvalues of P(r) and S(r) are equal, \begin (\mathbfP(r)^k\mathbf)/\lambda(r)^k&= (\mathbfP\sqrtS(r)^k \sqrt\mathbf)/\lambda(r)^k\\ &\le e^, , \frac, , _2, , S(r)^k, , _2, , \sqrt, , _2/\lambda(r)^k\\ &\le e^N_\\ &\le (1+r)N_\qquad \square \end Lemma 3 If r is a real number such that 0\le e^r-1\le \epsilon/4, :\log\lambda(r)\le r\pi(A)+5r^2/\epsilon Proof summary: We Taylor expand \log \lambda(y) about point r=z to get: :\log\lambda(r)= \log\lambda(z)+m_z(r-z)+(r-z)^2\int_0^1 (1-t)V_dt Where m_x \text V_x are first and second derivatives of \log \lambda(r) at r=x. We show that m_0=\lim_t_k=\pi(A). We then prove that (i) \epsilon_r\ge 3\epsilon/4 by matrix manipulation, and then prove (ii)V_r\le 10/\epsilon using (i) and Cauchy’s estimate from complex analysis. The results combine to show that :\begin \log\lambda(r)= \log\lambda(0)+m_0r+r^2\int_0^1 (1-t)V_dt \le r\pi(A)+5r^2/\epsilon \end :A line to line proof can be found in Gilman (199

Proof of theorem Combining lemma 2 and lemma 3, we get that :\Pr _k-\pi(A)\ge \gammale(1+r)N_e^ Interpreting the exponent on the right hand side of the inequality as a quadratic in r and minimising the expression, we see that :\Pr _k-\pi(A)\ge \gammale(1+\gamma\epsilon/10)N_e^ A similar bound :\Pr _k-\pi(A)\le - \gammale (1+\gamma\epsilon/10)N_e^ holds, hence setting C=2(1+\gamma\epsilon/10)N_ gives the desired result.


Uses

This theorem is useful in randomness reduction in the study of
derandomization A randomized algorithm is an algorithm that employs a degree of randomness as part of its logic or procedure. The algorithm typically uses uniformly random bits as an auxiliary input to guide its behavior, in the hope of achieving good performan ...
. Sampling from an expander walk is an example of a randomness-efficient
sampler Sampler may refer to: * Sampler (signal), a digital signal processing device that converts a continuous signal to a discrete signal * Sampler (needlework), a handstitched piece of embroidery used to demonstrate skill in needlework * Sampler (surna ...
. Note that the number of
bit The bit is the most basic unit of information in computing and digital communications. The name is a portmanteau of binary digit. The bit represents a logical state with one of two possible values. These values are most commonly represente ...
s used in sampling k independent samples from f is k \log n, whereas if we sample from an infinite family of constant-degree expanders this costs only \log n + O(k). Such families exist and are efficiently constructible, e.g. the
Ramanujan graph In the mathematical field of spectral graph theory, a Ramanujan graph is a regular graph whose spectral gap is almost as large as possible (see extremal graph theory). Such graphs are excellent spectral expanders. AMurty's survey papernotes, Ramanu ...
s of Lubotzky-Phillips- Sarnak.


References

* * * {{refend Sampling (statistics)