Large-scale structure of networks

guillefix 4th November 2016 at 2:43pm

Components

Networks often have the largest connected component covering most of the network (often more than 50% or 90%). This is sometimes called the "giant component" (however this is sloppy usage, as the term "giant component" means not precisely the same as "largest component" in network theory).

In directed networks, we can represent the largest strongly connected component, and its in and out components using a "bow tie" diagram

Shortest paths and the small-world effect

The small-world effect refers to the finding that the typical distance between nodes in many –perhaps most– networks is surprisingly small. The "typical distance" usually refers to the "mean geodesic distance". Networks that show this property are dubbed small-world networks.

The origin of the term comes from a series of experiments by social psychiatrist Stanley Milgram, the so called "small-world" studies, in the 60s.

Models of networks often show that this distance scales as logn\log{n}, where nn is the size of the network. This is often given as an upper limit for the growth of the distance with nn, so that the network is said to have the small-world proprety.

The diameter (the largest geodesic distance) is also found to scale similarly. For scale free networks, however, an interesting structure is often found with a core that contains most nodes, and is of lengthscale loglogn\log{\log{n}}, making the mean distance scale like that too, but there are a few nodes along "streamers" or "tendrils", around the core, whose lengthscale scales logn\log{n}, making the diameter scale like that too.

Another interesting effect that is observed, termed funneling, is that often it is found that the geodesic path (path(s) with shortest length) between vertex jj and ii passes through only a few particularly well-connected neighbours of ii for most choices of starting point jj. Thus if one follows shortest paths to try to reach ii, one is likely to be "funnelled" through those few or one particular neighbours of ii.

Degree distributions

The degree distribution pkp_k is the fraction of nodes in the network that have degree kk.

The same information can be given in a degree sequence, that is a sequence of the degrees of all the nodes in the network. One can easily see from simple examples, that this information doesn't, however, specify the network structure, in general.

For directed networks, we can define the joint in- and out- degree distribution pjkp_{jk}, the probability that a vertex has in-degree jj and out-degree kk. This has been currently rarely been measured in practice though.

Power laws and scale-free networks

Often (though definitely not always), real networks show a power law degree distribution:

pk=Ckαp_k=Ck^{-\alpha}

where α\alpha is the exponent. Values 2<α<32<\alpha<3 are typical. These are examples of right-skewed distributions. Typically, the power law is only obeyed for the tail of the distribution, but not for small values of kk. And typically it is also not obeyed in the high end, for example, due to some cut-off.

Networks with power-law degree distributions are sometimes called scale-free networks.

Distributions of other centrality measures

Distributions of the values for nodes for others centrality measures defined in Measures and metrics for networks.

Centralization

We can use the distribution of centrality measures to answer the question: "how are the centrality values spread?". High spreads indicate a good centrality measure (or very high centralization, I think), while low spreads indicate a low centrality measure (or descentralization, I think.).

A measure for it is:

C=i=1N[Cb(i)Cb(i)]i=1N[C~b(i)C~b(i)]\mathcal{C} = \frac{\sum_{i=1}^N [C_b(i^*)-C_b(i)]}{\sum_{i=1}^N [\tilde{C}_b(i^*)-\tilde{C}_b(i)]}

where in the denominator, C~b(i)\tilde{C}_b(i) is the betweeness centrality of node ii, and ii^* is a node that maximizes it, both for the graph that maximizes C~b(i)\tilde{C}_b(i^*) (a star graph for betweeness for example). The Cb(i)C_b (i) without the tilde is for the actual graph.

Dynamical importance (& eigenvalue elasticity)

  • measures changes in eigenvalues of AA due to some perturbations
  • suppose GG strongly connected.

Edge dynamical importance of (i,j)(i,j) is:

I(i,j)=Δλijλ1I(i,j)=-\frac{\Delta \lambda_{ij}}{\lambda_1}

where λ1\lambda_1 is the largest eigenvalue of A, and Δλij\Delta \lambda_{ij} is the change in λ1\lambda_1 from removing edge from jj to ii (i.e. removing AijA_{ij}).

The Node dynamical importance of (i)(i) is:

I(i)=Δλiλ1I(i)=-\frac{\Delta \lambda_{i}}{\lambda_1}

where Δλi\Delta \lambda_{i} is the change in λ1\lambda_1 from removing node ii (i.e. removing column and row ii).

One can show that:

I(i)viuivTuI(i)\approx \frac{v_i u_i}{v^T u}

I(i,j)Aijviujλ1vTuI(i,j)\approx \frac{A_{ij}v_i u_j}{\lambda_1 v^T u}

where the approximation is in only considering the changes in eigenvalue and eigenvector to 1st order. See problem sheet 4 answers for proof.

Structural things related (by spectrum, often) to dynamics Dynamics of removing nodes and edges he means?

Clustering coefficients

(see Transitivity (Graph theory))

Clustering coefficients, CC are often found to be larger than one would expect if edges where randomly chosen (for a fixed degree distribution, for example, see formula 8.24 in Newman's).

This is often the case for social networks. One mechanism that can lead to this is triadic closure (when an open triad is close, say because the common vertex introduces the other two, in case of social nets). This has indeed been found to happen in cases when time-resolved data for network formation is studied.

In, the Internet networks, however, CC is much smaller than the predicted value given by chance (eq 8.24 Newman), suggesting there are forces that shy away from creating triangles. However, different models to compare with (i.e. other random graph models), and other ways of measuring clustering coefficients give different results.

Other motifs apart from triangles are also measured sometimes and show interesting patterns (like in neural networks).

Local clustering coefficients often show an interesting anti-correlation with degree in real networks. An explanation to this phenomena can be given if the network has a community structure with nodes grouped together in groups of varying sizes. A hierarchical structure has also been proposed to explain this.

Assortative mixing

Assortative mixing is the tendency of nodes to connect to others that are like them in some way. The formula given in there is not very efficient to compute, because of the double sum going like n2n^2. There is however a more efficient one that goes like mm, the number of edges, which is often scales slower with nn (see eq. 8.27 in Newmann book).

Empirically, it is found that most social networks have positive assortativity while most others (technological, biological) have negative assortativity.

Part of the explanation for this seems to be that most networks are naturally dissasortative by degree because they are simple graphs (see Mathematics of networks) and so the number of connections between high degree nodes is limited, and so if there aren't many high degree nodes, they will have to connect mostly to lower degree nodes (I think this is gist of explanation).

Social networks, on the other hand, seem to overcome this due to their group structure (high clustering coefficient) so that in small groups people of low degree will be mostly connected to people with low degree (i.e. within the small group), and the larger groups will contribute to making people of high degree being mostly connected to people of high degree (i.e. within the large group).