Random walk in a graph

guillefix 4th November 2016 at 2:43pm

A random walk is a path across a network created by taking repeated random steps. They are usually allowed to traverse edges more than once, and visit vertices more than once. If note it is a self-avoiding random walk.

We consider a random walk where at each vertex one will take a step (i.e. walker does not stay in vertex) along each of the edges connected to it, with uniform probability, i.e. with probability 1ki\frac{1}{k_i}, where kik_i is the degree. Thus, on an undirected network we have:

pi(t)=Aijkjpj(t1)p_i(t)=\sum\frac{A_{ij}}{k_j}p_j(t-1)

or p(t)=AD1p(t1)\mathbf{p}(t)=\mathbf{A}\mathbf{D}^{-1}\mathbf{p}(t-1)

Where pi(t)p_i(t) is the probability that the walker is at vertex ii at (discrete) time tt, and where D=diag(k1,...,kn)\mathbf{D}=diag(k_1,...,k_n). One can also write this relation in terms of the reduced adjacency matrix, D1/2AD1/2\mathbf{D}^{-1/2}\mathbf{A}\mathbf{D}^{-1/2}, and that can be useful sometimes.

We are interested in the limit as tt \rightarrow \infty where we expect the probability to approach a steady state p()p\mathbf{p}(\infty) \equiv \mathbf{p}:

p=AD1p\mathbf{p}=\mathbf{A}\mathbf{D}^{-1}\mathbf{p}, which can be rewritten as LD1p=0\mathbf{L}\mathbf{D}^{-1}\mathbf{p}=0, so D1p\mathbf{D}^{-1}\mathbf{p} is an eigenvector of the Graph laplacian (L\mathbf{L}) with eigenvalue 00, but we known (see Graph laplacian) that in a connected network only the vector 1=(1,1,1,...)\mathbf{1}=(1,1,1,...) has eigenvalue 00. Therefore pkip \propto k_i, so normalizing p=kiiki=ki2mp = \frac{k_i}{\sum_i k_i}=\frac{k_i}{2m} (see Degree of a vertex (Graph theory))

With a random walk, an interesting question is that of the mean first passage time, or the mean number of steps before reaching a certain node, when starting from a given node. To find this we consider an absorbing random walk, where a walk that arrives at a certain set of vertices (we will consider just one, call it vv) will stay there.

We can then consider the probability pv(t)p_v(t) of being at vertex vv at time tt. This is the same as the probability that the first passage time is equal to or less than tt, and thus the probability that it is exactly tt is pv(t)pv(t1)p_v(t)-p_v(t-1), and the mean first passage time is:

τ=tt[pv(t)pv(t1)]\tau =\sum_t^\infty t[p_v(t)-p_v(t-1)]

Note that we can't rearrange terms in this sum, because it is not absolutely convergent!

Following the manipulations shown in Newman's book (section 6.14), we get to:

τ=1DL1p(0)\tau = \mathbf{1} \cdot \mathbf{D'}\mathbf{L'}^{-1} \cdot \mathbf{p'}(0)

where the prime ' indicates that the vvth element, or the vvth row and columns have been removed. In particular L\mathbf{L'} is called the vvth reduced Laplacian. This can be re-expressed a bit further, following Newman's book, for computational convenience.

Resistor networks

Kirkoff's current law can be written as:

jAijVjViR+Ii=0\sum_j A_{ij} \frac{V_j-V_i}{R} +I_i=0

where IiI_i is an external current applied at some node in the network. This can be written in terms of the Graph laplacian as:

LV=RI\mathbf{L}\mathbf{V}=R\mathbf{I} \quad (\dagger)

where V\mathbf{V} is the vector of voltages. L\mathbf{L} is not invertible, but this cooresponds to the arbitrariness in the value of voltages ViV_i, which can be all shifted up and down and still satisfy the equation. This is equivalent to adding a multiple of the 1\mathbf{1} vector, which we know to have a 00 eigenvalue of the Graph laplacian. However, if we fix the voltage at some node (to be 00 say), then we can remove the corresponding columns and rows from the equation (\dagger), and the 00 eigenvalue is removed, and the reduced Laplacian is now invertible, so we can get the voltages, and thus the currents!

Some applications

Random walk sampling method for social networks

Random walk betweenness measure.