Professional Documents
Culture Documents
G, called sparsier, with O(nlog n) edges (or fewer) on the same set of vertices in such a way that every cut
in sparsier
G is within a factor of (1 ) of the corresponding cut in G for some constant > 0. It was
shown that, for all x 0, 1
n
,
[x
T
Lx x
T
Lx[ x
T
Lx, (1)
where L and
L are the Laplacian of the graph G and its sparsier
G, respectively.
Spielman and Teng [46, 47] devised a sampling scheme to construct a spectral sparsier with O(nlog
c
n)
edges for some (large) constant c in O(m polylog(n)) time. A spectral sparsier
G for graph G is a a
sparsier satisfying equation (1) for all x R
n
. In [48] Spielman and Srivastava gave a dierent sampling
1
polylog(n)) (see [45]), together with
an approximation guarantee within a quadratic root of the optimal conductance (where denotes the spectral
gap of the normalized Laplacian). The running time can be further improved to O(n
2
polylog(n)) by using
Spielman-Teng solver for linear systems [47]. Another approach for the balanced cut problem is by using
commodity ows [7, 41]. In [7] the approximation guarantee is within a factor of log n of the optimal in
[7]), which was further reduced to log
1/2
n in [6] but the running time is still O(n
2
polylog(n)). In another
direction, Spielman and Teng [46, 48] introduced local graph partitioning which yields a cut near the specied
seeds with running time only depending on the volume of the output. Their local partitioning algorithm has
an approximation guarantee similar to the spectral method by using a mixing result on random walks [38, 39].
Andersen, Chung and Lang [3] used PageRank vectors to give a local partitioning algorithm with improved
approximation guarantee and running time. Recently, Andersen and Peres use involving sets instead of
PageRank to further improved the running time [4].
Our balanced-cut algorithm consisting of two parts. First we use PageRank vectors to sparsify the graph.
Then we use the known PageRank partitioning algorithm to nd a balanced cut. Both parts have the same
complexity as computing the PageRank vectors. Consequently, the complexity for our PageRank balanced-cut
algorithm is O(mlog
2
n/+n polylog(n)) for any input graph on n vertices and m edges. The balanced-cut
algorithm here can be viewed as an application of graph sparsication.
2 A sharp PageRank approximation algorithm
We consider an undirected, weighted graph G = (V, E, w) with n vertices and m edges where the edge
weights w(u, v) = w(v, u) 0 and the edge set E consists of all pairs (u, v) with w(u, v) > 0. The weighted
degree d(u) of vertex u is the sum of w(u, v) over all v, i.e., d(u) =
v
w(u, v).
A typical random walk is dened by its transition probability matrix P satisfying P(u, v) = w(u, v)/d(u).
We may write P = D
1
A, where A is the weighted adjacency matrix satisfying A(u, v) = w(u, v) for all
pairs of u, v V and D is the diagonal matrix of weighted degree. We here consider the lazy walk Z on G,
dened by
Z =
I +P
2
.
In [3] PageRank vector p satises a recurrence relation involving a seed vector s (as a probability distribution)
and a positive jumping constant < 1 ( or transportation constant):
p = s + (1 )pZ
where p and s are taken to be row vectors. In this paper, we consider the PageRank p as a discrete Greens
function satisfying
p = s(I (1 )Z)
1
= s(I +L)
1
3
where = 2/(1 ) and L = I P. Note that the usual Greens function is associated with the pseudo
inverse of L. Another way to express the recurrence of PageRank in terms of is the following: For a positive
value > 0, the (personalized) PageRank vector pr
,s
with a seed vector s is the unique solution of the
following:
pr
,s
=
2 +
s +
2
2 +
pr
,s
Z.
If a seed vector is the characteristic function
u
of a single vertex u, we may write pr
,u
= pr
,u
if there is
no confusion. It is easy to check that
vV
pr
,s
(v) = 1 since
vV
s(v) = 1.
The PageRank approximation algorithms in [3] contains the following steps, called push and Approx-
imatePR, which serve as subroutines later in the sharp approximate PageRank algorithm. For a vertex u,
a current approximate PageRank vector p and a residual vector r, the push step is as follows:
push(u):
Let p
= p and r
(u) = p(u) +
2+
r(u)
2. r
(u) =
r(u)
2+
3. For each vertex v such that (u, v) E:
r
(v) = r(v) +
r(u)
(2+)d(u)
.
Lemma 1 ([3]). Let p
and r
= pr
,sr
.
ApproximatePR(s, , ):
1. Let p = 0 and r = s.
2. While r(u) d(u) for some vertex u:
(a) Pick any vertex u where r(u) d(u).
(b) Apply push(u).
3. Return p and r.
Theorem 1 ([3]). For any s with |s|
1
1, and any constant (0, 1], the algorithm ApproximatePR(s, , )
computes an approximate PageRank vector p = pr
,sr
such that the residual vector r satises [r(v)/d(v)[
for all v V . The running time of the algorithm is O(
2+
).
We will improve the estimate error bound for the above algorithm by the following iterated process:
SharpApproximatePR(s, , ):
1. Let
= 1, r = s and p = 0.
2. While
> :
(a) Set
/2.
(b) Let p
and r
).
(c) Let p = p +p
and r = r
.
3. Return p and r.
4
Theorem 2. For any constant (0, 1], > 0 and any seed vector s, to approximate pr
,s
the algorithm
SharpApproximatePR(s, , ) computes approximate PageRank vector p = pr
,sr
such that the residual
vector r satises [r(v)/d(v)[ for all v V . The running time of the algorithm is O(
(2+)mlog(1/)
).
Proof. Clearly, the algorithm returns an approximate PageRank vector p = pr
,sr
where the residual vector
satises that r(v) d(u) at every vertex.
To bound the running time, we consider one xed round of the second step. Let T denote the total number
of push operations performed by ApproximatePR(r, ,
), and let d
i
be the degree of vertex involved in the
ith push operation. When the ith push operation was performed, the PageRank at this vertex was at least
d
i
, so |r|
1
decreased by at least
d
i
2 +
.
Since |r|
1
is at most 2
vV
d(v) = 2m
, we have
2 +
T
i=1
d
i
2m
,
so
T
i=1
d
i
2(2 +)m
.
Note that there are at most log(1/) rounds in the second step. Thus, The total running time is bounded by
O(
(2+)mlog(1/)
). .
As an immediate consequence of Theorem 2, we have the following by taking to be the inverse of a
power of n.
Corollary 1. In a graph on n vertices, for any xed integer k, > 0 and any seed vector s, to approximate
pr
,s
the algorithm SharpApproximatePR(s, , ) computes approximate PageRank vector p = pr
,sr
such that the residual vector r satises [r(v)/d(v)[ n
k
for all v V . The running time of the algorithm
is O(
(2+)mlog n
).
3 The Green values for edges in a graph
Recall that graph G = (V, E, w) is a connected weighted undirected graph with n vertices, m edges. The
combinatorial Laplacian of G is dened by L = D A where A is the weighted adjacency matrix and D is
the diagonal matrix of weighted degrees. If we orient the edges of G in an arbitrary but xed way, we can
write its Laplacian as L = B
T
WB, where B
mn
is the signed edge-vertex incidence matrix, given by
B(e, v) =
_
_
_
1 if v is es head,
1 if v is es tail,
0 otherwise
and W
mm
is the diagonal matrix with W(e, e) = w(e). The normalized Laplacian of G is dened to be
L = D
1/2
LD
1/2
and we write L = S
T
WS where S
mn
= BD
1/2
. Since L is symmetric and we have
L =
n1
i=0
T
i
i
,
where
0
= 0 and 0 <
1
2
. . .
n1
2 are the nonzero eigenvalues of L and
0
, . . . ,
n1
are a
corresponding orthonormal basis of eigenvectors. Various properties concerning eigenvalues of the normalized
Laplacian can be found in [13].
5
Denote the -normalized Laplacian L
= S
T
W
where we dene S
and W
as follows:
S
=
_
I
S
_
(n+m)n
and W
=
_
I 0
0 W
_
(n+m)(n+m)
.
Here the index set for the columns of S
= I. Clearly, we have
(
=
n1
i=0
1
i
+
T
i
i
. (2)
We remark that the discrete Greens function is basically a symmetric form of the PageRank. Namely, it is
straightforward to check that
pr
,s
= sD
1/2
(
D
1/2
. (3)
For each edge e = u, v E, we dene the Green value g
(u, v) = (
u
v
)D
1/2
(
D
1/2
(
u
v
)
T
=
pr
,u
(u)
d(u)
pr
,u
(v)
d(v)
+
pr
,v
(v)
d(v)
pr
,v
(u)
d(u)
. (4)
From (2) and (4), we have the following fact which will be useful later.
Lemma 2. The Green value g
can be expressed by
g
(u, v) =
n1
i=0
i
+
_
i
(u)
_
d(u)
i
(v)
_
d(v)
_
2
.
Lemma 3. For two distinct vertices u, v V , we have
2 +
_
1
d(u)
+
1
d(v)
_
g
(u, v)
1
d(u)
+
1
d(v)
.
Proof. Since
i
2, we have
2 +
i
+
1.
From Lemma 2 we have
2 +
_
i
(u)
_
d(u)
i
(v)
_
d(v)
_
2
g
(u, v)
n1
i=0
_
i
(u)
_
d(u)
i
(v)
_
d(v)
_
2
.
Note that for two xed vertices u and v, the vector f
u
, dened by f
u
(i) =
i
(u), for i = 0, 1, . . . , n 1, is
orthogonal to f
v
. This implies
n1
i=0
i
(u)
i
(v)
_
d(u)d(v)
= 0
6
and
n1
i=0
_
i
(u)
_
d(u)
i
(v)
_
d(v)
_
2
=
n1
i=0
_
2
i
(u)
d(u)
+
2
i
(v)
d(v)
_
=
1
d(u)
+
1
d(v)
.
.
Since the Green values are relatively small (e.g., of order 1/n
c
, for some positive constant c), we need
very sharply approximate PageRank to be within a factor of 1 +O(n
c
) of the exact values in the analysis
of the performance bound for the graph sparsication algorithms that we will examine in Section 4.
For all the pairs (u, v), we dene the approximate Green value g
by
g
(u, v) =
pr
,uru
(u)
d(u)
pr
,uru
(v)
d(v)
+
pr
,vrv
(v)
d(v)
pr
,vrv
(u)
d(u)
where pr
,uru
and pr
,uru
are the approximate PageRank vectors as outputs of ApproximatePR for
pr
,u
and pr
,v
respectively, and r
u
and r
v
are the corresponding residual vectors such that |r
u
D
1
|
1
/4 and |r
v
D
1
|
1
/4. In the following, we will prove that
Lemma 4. For two distinct vertices u, v V , we have
[g
(u, v) g
(u, v)[ .
Proof. Let T
= D
1/2
(
D
1/2
, and we may express g
(u, v) = (
u
r
u
)T
D
1
T
u
(
u
r
u
)T
D
1
T
v
+
(
v
r
v
)T
D
1
T
v
(
v
r
v
)T
D
1
T
u
,
which implies that
g
(u, v) g
(u, v) = r
u
T
D
1
T
u
+r
v
T
D
1
T
u
+r
v
T
D
1
T
v
+r
u
T
D
1
T
v
.
Since 1Z = 1, we have 1DT
D
1
= 1 and |DT
D
1
|
1
= 1. Therefore
|r
u
T
D
1
|
1
= |r
u
D
1
DT
D
1
|
1
|r
u
D
1
|
1
|DT
D
1
|
1
= /4.
Thus,
[g
(u, v) g
(u, v)[ [r
u
T
D
1
T
u
[ +[r
v
T
D
1
T
u
[ +[r
v
T
D
1
T
v
[ +[r
u
T
D
1
T
v
[
/4 +/4 +/4 +/4 = .
The Lemma is proved. .
Here we will give a rough estimate for computing Greens values by directly using Theorem 1. Note that
by invoking Theorem similarly without using further techniques the running time does not seem to improve.
Theorem 3. Given any constant > 0 and any pair (u, v) V V , the approximate Green value g
(u, v)
can be computed in O(
2+
) time and
[g
(u, v) g
(u, v)[ .
In particular, after O(
(2+)n
(u, v) by
using a constant number of queries.
Proof. By Theorem 1, we can compute approximate PageRank vectors pr
,vrv
such that the residual
vector r
v
satises [r
v
(v)/d(v)[ /4 for all v V in O(
2+
(u, v) in
O(
(2+)
2
(u, v) g
(u, v)[ g
(u, v).
In particular, after O(
(2+)
2
n
(u, v)
by using a constant number of queries.
Proof. In Lemma 3, a lower bound for g
(u, v) is established: g
(u, v)
(2+)
. Thus, for the targeted error
bound , we set
=
(2+)
and apply Theorem 3 with parameter
instead of . .
We will improve both Theorem 3 and Corollary 2 in the Section 5 by using sharp approximate PageRank
algorithms and dimension reduction techniques.
4 Graph sparsication using Green values
To construct our sparsier, we use a method quite similar to the scheme by Spielman and Srivastava except
that PageRank is used here instead of eective resistance. We will give three graph sparsication algorithms,
two of which involve approximate Green values (which will be examined in section 5). In the section, we use
exact Green values for edges.
Recall that the graph G = (V, E, w) we consider here is an undirected weighted graph. For a subset S of
vertices in G, the edge boundary (S) of S consists of all edges with exactly one endpoint in S. The weight
of (S), denoted by w(S,
S), is the sum of all edge weights of edges in (S). The volume of S, denoted by
vol(S), is dened to be the sum of degrees d(v) over all v in S. When S = V , we write vol(S) = vol(G). The
Cheeger ratio (or conductance) h
G
(S) of S in G is dened by
h
G
(S) =
w(S,
S)
minvol(S), vol(
S)
.
The conductance h
G
of G is dened to be the minimum Cheeger ratio among all subsets S with vol(S)
vol(G).
The goal of sparsication is to approximate a given graph G by a sparse graph
G on the same set of
vertices while the sparse graph
G preserves the Cheeger ratios of every subset of vertices to within a factor
of 1 +. The sparse graph
G is called a sparsier of G if the Laplacian L and
L for G and
G satisfy (1) for
all x 0, 1
n
.
The main step in any sparsication algorithm [9, 3032, 4648] is to choose an appropriate probability
distribution for random sampling the edges in a way that Cheeger ratios of subsets change little. Our
sparsication algorithm is a sampling process using probabilities proportional to the Green values g
s as
follows:
G = SparsifyExactGreen(G, q, ):
For each e = (u, v) E, set probability p
e
w(e)g
G
(S) h
G
(S)[ .
2. (Sampling complexity) Algorithm SparsifyExactGreen can be performed by using O(
nlog n
2
) sampling.
4.1 Analyzing the sparsier
Our analysis follows the general scheme as that of [48]. In our analysis of sparsier, we consider the matrix
= W
1/2
S
T
W
1/2
. Note that
(e, e) =
1
_
W
(e, e)g
(e)
_
W
(e, e) =
1
w(e)g
.
Lemma 5. (i)
2
.
(ii) The dimension (or rank) of
, denoted by dim(
) is n.
(iii) The eigenvalues of
(e, e) = |
(, e)|
2
2
.
Proof. For (i), we note that
= (W
1/2
S
T
W
1/2
)(W
1/2
S
T
W
1/2
)
= W
1/2
(S
T
W
)(
S
T
W
1/2
= W
1/2
S
T
W
1/2
= W
1/2
S
T
W
1/2
.
For (ii), it is easy to verify that dim(W
1/2
S
T
W
1/2
) = dim(S
S
T
) = dim((
) = n.
For (iii), Since
2
, the eigenvalue of
is symmetric. .
Next, we introduce some notations and several lemmas that the theorems in later sections rely on. Let
w(e) be the edge weight of edge e in
G. Recall that q is a number and p
e
is the sampling probability for
e E. Denote I
=
_
I
nn
0
0 R
_
(n+m)(n+m)
,
where R is a mm nonnegative diagonal matrix denoted by
R(e, e) =
w(e)
w(e)
=
# of times e is sampled
qp
e
.
Lemma 6. Suppose I
|
2
. Then
x R
n
, [x
L
x
T
xL
x
T
[ xL
x
T
,
where L
= S
T
W
and
L
= S
T
W
1/2
W
1/2
.
9
Proof. The assumption of this lemma is equivalent to
sup
yR
m+n
,y=0
[y
(I
I)
y
T
[
yy
T
,
which implies
sup
y
T
im(W
1/2
),y=0
[y
(I
I)
y
T
[
yy
T
.
Here y im(M) means y = xM for some x. We have
sup
y
T
im(W
1/2
),y=0
[y
(I
I)
y
T
[
yy
T
= sup
xR
n
,W
1/2
x
T
=0
[xS
T
W
S
T
W
1/2
(I
I)W
1/2
S
T
W
x
T
[
xS
T
W
x
T
= sup
xR
n
,W
1/2
x
T
=0
[xL
S
T
W
1/2
(I
I)W
1/2
x
T
[
xS
T
W
x
T
= sup
xR
n
,W
1/2
x
T
=0
[xS
T
W
1/2
(I
I)W
1/2
x
T
[
xS
T
W
x
T
= sup
xR
n
,W
1/2
x
T
=0
[x
L
x
T
xL
x
T
[
xL
x
T
The lemma follows from the fact that dim(im(W
1/2
)) = n and W
1/2
x
T
= 0 if and only if x = 0. .
Lemma 7 ([44]). Let p be a probability distribution over R
d
such that sup
y
|y|
2
M and
E
p
|y
T
y|
2
1. Let y
1
. . . y
q
be independently samples drawn from p. Then for every 1 > > 0,
P
__
_
_
_
_
1
q
q
i=1
y
T
i
y
i
Ey
T
y
_
_
_
_
_
2
>
_
2 exp(
2
/a
2
),
where a = min
_
CM
_
log q
q
, 1
_
and C is an absolute constant.
4.2 Proof of Theorem 4
We rst prove the following theorem which leads to the proof of Theorem 4.
Theorem 5. Let L be the normalized Laplacian of G. Let
G be the output of the algorithm
SparsifyExactGreen(G, q, ), where q = 4C
2
nlog n/
2
, 1 > 0 and C is a absolute constant. Then with
probability at least 1/2, we have
x R
n
, [x
L
x
T
xL
x
T
[ xL
x
T
, (5)
where L
= I +L = S
T
W
and
L
= S
T
W
1/2
W
1/2
.
Remark 1. Let L
be the matrix D
1/2
D
1/2
L
D
1/2
D
1/2
where
D and
L are the diagonal matrices of weighted
degree and the normalized Laplacian of graph
G, respectively. We observe that the inequality in (5) is
equivalent to the following:
x R
n
, [xL
x
T
xLx
T
[ xL
x
T
. (6)
10
Remark 2. The probability 1/2 can be improved to high probability, i.e., 11/n
O(1)
. However, the number
of sampled edges will be increased from O(nlog n) to O(nlog
2
n).
Proof of Theorem 5: Before applying Lemma 7, we observe that
eE
R(e, e)
(e, )
T
(e, ) +
vV
(v, )
T
(v, )
eE
(e, )
T
(e, ) +
vV
(v, )
T
(v, ).
Thus we have
=
eE
(R(e, e) 1)
(e, )
T
(e, ).
Now, we consider
eE
R(e, e)
(e, )
T
(e, )
=
eE
# of times e is sampled
qp
e
(e, )
T
(e, )
=
1
q
eE
( # of times e is sampled)
(e, )
T
p
e
(e, )
p
e
=
1
q
q
i=1
y
T
i
y
i
where y
1
, . . . , y
q
are random vectors drawn independently with replacement from the distribution p dened
by setting y =
1
pe
eE
p
e
1
p
e
(e, )
T
(e, )|
2
= |
eE
(e, )
T
(e, )|
2
= sup
xR
m+n
,x=0
eE
x
(e, )
T
(e, )x
T
xx
T
11
Since for any x R
m+n
, we have
vV
x
(v, )
T
(v, )x
T
0. Thus
|E
p
y
T
y|
2
= sup
xR
m+n
,x=0
eE
x
(e, )
T
(e, )x
T
xx
T
sup
xR
m+n
,x=0
eE
x
(e, )
T
(e, )x
T
+
vV
x
(v, )
T
(v, )x
T
xx
T
= sup
xR
m+n
,x=0
x
_
eE
(e, )
T
(e, ) +
vV
(v, )
T
(v, )
_
x
T
xx
T
= sup
xR
m+n
,x=0
x
x
T
xx
T
= sup
xR
m+n
,x=0
x
x
T
xx
T
= |
|
2
= 1.
The last equality above follows from the fact that the eigenvalues of
p
e
|
(e, )|
2
=
1
p
e
_
(e, e) =
E
w(e
)g
(e
)
w
e
g
(e)
w
e
g
(e)
E
w(e
)g
(e
.
Note that
= W
1/2
S
T
W
1/2
_
(
1/2
(
S
T
W
1/2
1/2
W
1/2
S(
W
1/2
S(
S
T
W
1/2
_
(n+m)(n+m)
.
Thus,
eE
w
e
g
(e)
eE
(e, e) = Tr(W
1/2
S(
S
T
W
1/2
) = Tr(
) Tr((
).
Since Tr(
) = n, we have
1
pe
|
(e, )|
2
n. By setting q = 4C
2
nlog n/
2
, we have
min
_
CM
log q
q
, 1
_
C
2
nlog(4C
2
nlog n/
2
)
4C
2
nlog n
/2. (7)
By applying the Rudelson and Vershynins lemma in [44] (Lemma 7), we completes the proof of the theorem.
.
Before applying Theorem 5 to prove Theorem 4, we still need the following two lemmas. We will here
consider G as an unweighted graph rst, i.e w(e) = 1 for all edges, although this can be easily extended to
the general weighted graphs.
Lemma 8. Let
G be the output of algorithm SparsifyExactGreen(G, q, ), where q = 4C
2
n( + 2) log n/
2
.
Then, with probability 1 1/n, for all subsets S V , we have [vol
G
(S) vol
G
(S)[ vol
G
(S).
Proof. For a xed set S V , let E(S) be the set of edges with at least one endpoint in S. Consider the i.i.d.
random variables X
1
, X
2
, . . . , X
q
dened as follows:
X
i
=
_
1/p
e
with probability p
e
for e E(S)
0 otherwise.
12
We consider random variable V
S
=
1
q
q
i=1
X
i
. It is easy to verify that
V
S
=
eE(S)
# of times e is sampled
qp
e
=
eE(S)
w(e)
w(e)
= vol
G
(S).
Clearly,
E[X
i
] =
eE(S)
p
e
1
p
e
= vol
G
(S),
and
E[X
2
i
] =
eE(S)
p
e
1
p
2
e
=
eE(S)
1
p
e
=
eE(S)
E
w(e
)g
(e
)
w(e)g
(e)
.
Note that
eE
w
e
g
(e) Tr(
eE(S)
E
w(e
)g
(e
)
w(e)g
(e)
eE(S)
n
g
(e)
e=(u,v)E(S)
n( + 2) mind(u), d(v)
n( + 2)vol
2
(S).
By the Bernstein inequality, for i.i.d. random variables X
1
, X
2
, . . . , X
k
such that [X
i
[ M.
P
_
1
k
k
i=1
X
i
1
k
k
i=1
E[X
i
]
>
t
k
k
i=1
E[X
i
]
_
exp
_
t
2
(
k
i=1
E[X
i
])
2
k
i=1
E[X
2
i
] +Mt/3
_
.
In our case, t = , k = q,
q
i=1
E[X
i
] = qvol
G
(S), C > 1, and M = n( + 2)
S
where
S
is the maximum
degree of vertices in S, i.e.
S
= max
vS
d(v). Thus,
exp
_
t
2
(
k
i=1
E[X
i
])
2
k
i=1
E[X
2
i
] +Mt/3
_
exp
_
q
2
2
vol
2
(S)
qn( + 2)vol
2
(S) +n( + 2)
S
/3
_
exp
_
qt
2
2( + 2)n
_
= exp
_
2
2
C
2
log n
2
_
1/n
2
. (8)
Therefore, the probability of the event that there exists a vertex v such that [vol
G
(v) vol
G
(v)[ > is less
than 1/n and
[vol
G
(S) vol
G
(S)[
vV
[vol
G
(v) vol
G
(v)[
vV
vol
G
(v) vol
G
(S).
The proof of the lemma is complete .
Lemma 9. If sparse graph
G corresponding to graph G satises the following two conditions:
13
1. for all x R
n
, [x
L
x xL
x
T
[ xL
x
T
;
2. for all subsets S V , [vol
G
(S) vol
G
(S)[ vol
G
(S).
Then [h
G
(S) h
G
(S)[ 2h
G
(S) +.
Proof. Let L and
L denote the Laplacians of the graphs G and
G respectively. Let D and
D be the weighted
degree matrices of the graphs G and
G respectively. For > 0, we consider L
= D+L and
L
=
D+
L.
Let
S
denote the characteristic function on S V , satisfying
S
(v) = 1 if v S and
S
(v) = 0 otherwise.
From the denition of Cheeger ratio, we have
[h
G
(S) h
G
(S)[ =
T
S
S
D
T
S
S
D
T
S
S
D
T
S
S
L
T
S
S
D
T
S
T
S
S
D
T
S
S
(D
D)
T
S
S
D
T
S
S
L
T
S
S
D
T
S
S
L
T
S
S
D
T
S
.
Recall that L
= S
T
W
and
L
= S
T
W
1/2
W
1/2
S
L
T
S
S
D
T
S
=
S
D
1/2
L
D
1/2
T
S
S
D
T
S
and
S
L
T
S
S
D
T
S
=
S
D
1/2
L
D
1/2
T
S
S
D
T
S
.
Thus, the rst condition implies that
S
L
T
S
S
D
T
S
S
L
T
S
S
D
T
S
S
L
T
S
S
D
T
S
.
Now, by the second condition, we have
[h
G
(S) h
G
(S)[ (1 +)
S
L
T
S
S
D
T
S
S
(D
D)
T
S
S
D
T
S
S
L
T
S
S
D
T
S
(1 +)
S
L
T
S
S
D
T
S
S
L
T
S
S
D
T
S
S
L
T
S
S
D
T
S
S
L
T
S
S
D
T
S
+
= 2h
G
(S) +.
The lemma follows. .
Proof of Theorem 4: To prove Theorem 4, we combine Lemma 8, Lemma 9 and Theorem 5. For any 1 > > 0,
let
G be the output of the algorithm SparsifyExactGreen(G, q, ), where q = 256C
2
nlog n/
2
, = 1/2,
and C 1 is a constant. By Theorem 5 and Lemma 8, the conditions of Lemma 9 are satised with probability
at least 1/2. Note that we have chosen to be 1/2 and h
G
(S) 1, thus algorithm SparsifyExactGreen
can be applied by using O(
n log n
2
) sampling. Furthermore, for all S V , we have
[h
G
(S) h
G
(S)[ .
.
By choosing a dierent , namely, = /2, we have the following:
Theorem 6. Given any 1 > 0 and 1 > 0, let
G be the output of the algorithm SparsifyExactGreen(G, q, ),
where q = 256C
2
nlog n/
2
, = /2, C 1, and G is an unweighted graph with m edges. Then with proba-
bility at least 1/2, we have
1. (Performance guarantee) For all S V with h
G
(S) , [h
G
(S) h
G
(S)[ h
G
(S).
2. (Sampling complexity) Algorithm SparsifyExactGreen can be performed by using O(
nlog n
2
) sampling.
14
5 Sparsication using approximate PageRank vectors
In Corollary 2, we compute approximate Green values g
(u, v) g
(u, v) (1 +)g
(u, v)
in O(n) time, where is any absolute constant such that < 1 (e.g., = 0.01). Instead of exact Green val-
ues, we can use approximate Green values as we run the algorithm SparsifyExactGreen. The approximate
Green values g
,v
s. Here we choose the following
parameters for ApproximatePR :
pr
,v
= ApproximatePR(
v
, ,
(2 +)
).
It is not dicult to verify that all results in Section 4 will change at most by a constant factor if we
run the algorithm SparsifyExactGreen by using approximate Green values as above. The performance
guarantee and the number of sampled edges in the Theorems 4 dier by at most a constant factor, although
the computational complexity will increase to O(n). In order to further improve the running time, we use
several methods in the following subsections.
5.1 Graph sparsication by using sharply approximate Green values
In order to have better error estimate of approximate Green values, we need to improve the error estimate for
approximate PageRank vectors in Theorem 1. We will use the strengthened approximate PageRank algorithm
SharpApproximatePR and the dimension reduction technique in [48] to approximate the Green values
by using these sharply approximate PageRank vectors produced by SharpApproximatePR.
First, we notice that
g
(u, v) = (
u
v
)D
1/2
(
D
1/2
(
u
v
)
T
= (
u
v
)D
1/2
(
D
1/2
(
u
v
)
T
= |W
D
1/2
(
u
v
)
T
|
2
2
=
1
|W
D
1/2
[D
1/2
(
D
1/2
](
u
v
)
T
|
2
2
.
Therefore, g
vV
where
Z = W
D
1/2
[D
1/2
(
D
1/2
].
However, the dimension of the vectors in Z
T
v
vV
is m+n. In order to reduce the computational complexity
for computing these vectors, we project these vectors into a lower dimensional space while preserving their
pairwise distances by the following lemma.
Lemma 10 ([1]). Given vectors x
1
, . . . , x
n
R
d
and constants , > 0, let k
0
= c
log n/
2
where c
is
a constant depending on . For integer k k
0
, let R
kd
be a random matrix where R
ij
are independent
random variables with values 1/
, we have
(1 )[[x
i
x
j
[[
2
2
[[Rx
i
Rx
j
[[
2
2
(1 +)[[x
i
x
j
[[
2
2
.
Now, we are ready to state our algorithm to approximate the Green values.
15
ApproxiamteGreen(, , k):
1. Let R
k(n+m)
= [R
1
, R
2
] be a random matrix whose entries are independent random variables with
values 1/
k, where R
1
is an k n matrix and R
2
is an k m matrix.
2. Let Y = RW
1/2
D
1/2
and
Z = RZ.
3. For i = 1, . . . , k, do the following
(a) Let y
i
be the ith row of Y and z
i
be the ith row of
Z.
(b) Approximate z
i
by calling z
i
= SharpApproximatePR(y
i
, , /n
r
) where r is a constant.
4. Let
Z
, z
2
, . . . , z
k
.
For all (u, v) E, return g
(u, v) = |
Z
(
u
v
)
T
|
2
2
.
In order to analysis algorithm ApproxiamteGreen, we need to give a bound for y
i
s by the following
lemma.
Lemma 11. Given an integer k and random matrix whose entries are independent random variables with
values 1/
vV
d(v)
1/2
)
k (i.e., with
probability 1/2 to be 1/
l=1
X
l
+
1/2
_
d(v)X
0
.
By Chernos inequality, we have
P[
d(v)
l=1
X
l
[ >
ka < 2 exp
ka
2
2d(v)
,
which is equivalent to
P[
d(v)
l=1
X
l
[ > c
_
d(v) log n < 2 exp
c
2
log n
2
where c 4 is a constant. Notice that ? 1 and [X
0
[ = 1/
(u, v) satisfying [g
(u, v) g
(u, v)[ g
(u,v)
in O(
mlog
2
n
2
) time.
Proof. To bound the running time, note that step 1 can be completed in O(mlog n/) time since R is a
k (n +m) random matrix. In step 2, we set Y = RW
1/2
D
1/2
and it only takes O(mlog n/
2
) time since
S
is a diagonal matrix.
16
In step 3, Let y
i
be the ith row of Y and z
i
be the ith row of
Z, i.e.,
(Y [D
1/2
(
D
1/2
])
kn
.
Therefore, we have z
i
= y
i
[D
1/2
(
D
1/2
] and we can view z
i
as a scaled PageRank vector with seed
vector y
i
.
In Lemma 11, we have proved that with probability at least 1 1/n, |y
i
|
1
= O(
vV
_
d(v) log n) for
1 i k. Without loss of generality, we may assume O(
vV
_
d(v) log n) = O(m) otherwise the graph is
sucient sparse. Thus, z
i
can be approximated by using algorithm ApproximatePRB with arbitrary small
absolute error, say,
) time. By Lemma 3,
g
= /n
r
for some large enough but xed
constant r. Thus, each call of ApproximatePRB will actually take O(mlog n/) time. Since there are
k = O(log n/
2
) calls, the total running time of step 3 is O(mlog
2
n/
2
).
In step 4, since each column of
Z
has k = O(log n/
2
) entries and there are m edges, the running time
of step 4 is O(mlog n/
2
). The lemma then follows. .
G = SparsifyApprGreen(G, q, , ):
1. For all (u, v) E, compute approximate Green values by ApproxiamteGreen(, , k) where
= 1/2 and k = c log n/
2
for some absolute constant c.
2. Apply SparsifyApprGreen(G, q, ) with approximate Green values g
(e)s.
Theorem 8. Given any constant > 0, > 0 and an unweighted graph G with m edges, set q =
256C
2
nlog n/
2
, = /2, and let
G be the output of the algorithm SparsifyApprGreen(G, q, , ) where
we use algorithm ApproxiamteGreen as a subroutine to approximate the Green value g
G
(S) h
G
(S)[ h
G
(S).
2. (Complexity guarantee) Algorithm SparsifyApprGreen can be performed by using O(
mlog
2
n
) prepro-
cessing time, and O(
nlog n
2
) sampling.
The proof of Theorem 8 is quite similar to the analysis in Section 4 and will be omitted.
6 Graph partitioning using approximate PageRank vectors
In this section, we combine the graph sparsication and the partitioning algorithms using PageRank vectors
to derive an improved partitioning algorithm.
An application of our sparsication algorithm by PageRank is the balanced cut problem. For a given graph
G, we rst use our sparsication algorithm to preprocess the graph. Then we apply the local partitioning
algorithm using PageRank vectors [3, 4] on the sparsier. Since the local partitioning algorithm is a subroutine
for the balance cut problem, we obtain a balanced cut algorithm with an improved running time. Spielman
and Teng [46] gave a local partitioning algorithm which, for a xed value of , gives a cut with approximation
ratio O(
1/2
log
3/2
n) and of volume v
in O(mlog
c
(n)/
5/3
) time where v
G
(S) h
G
(S)[ h
G
(S) for all S with h
G
(S) . Then, we use
the algorithm PageRank-Partition [3] on graph
G instead of G for balanced cut problem. The algorithm
17
PageRank-Partition has two inputs including a parameter and a graph with m edges. As stated in [3],
the PageRank-Partition algorithm has expected running time O(mlog
4
m/
2
). Furthermore, with high
probability the PageRank-Partition algorithms was shown to be able to nd a set S, if exists, such that
vol(S) vol(C)/2 and h
G
(S) . This can be summarized as follows:
Theorem 9. Given an undirected, weighted graph G = (V, E, w) with n vertices and m edges, constant
> 0, and > 0. With probability 1/2, we can preprocess graph G in O(
mlog
2
n
2
) edges such that for all S V satisfying h
G
(S) ,
[h
G
(S) h
G
(S)[ h
G
(S).
Algorithm PageRank-Partition takes as inputs a parameter and a graph
G and has expected running
time O(nlog
6
m/(
2
2
)). If there exists a set C with h
G
(C) = O(
2
/log
2
n), then with high probability the
PageRank-Partition algorithm nds a set S such that vol(S) vol(C)/2 and h
G
(S) .
References
1. D. Achlioptas, Database-friendly random projections, PODS01, pages 274281.
2. N. Alon, A. Andoni, T. Kaufman, K. Matulef, R. Rubinfeld, N. Xie, Testing k-wise and almost k-wise indepen-
dence, STOC 07, 496505
3. R. Andersen, F. Chung, and K. Lang, Local graph partitioning using pagerank vectors, FOCS 06, 475486.
4. R. Andersen and Y. Peres, Finding sparse cuts locally using evolving sets, STOC 09, 235244.
5. D. Achlioptas and F. McSherry, Fast computation of low rank matrix approximations, STOC 01, 611618.
6. S. Arora, E. Hazan, and S. Kale, (