Abstract
In large random networks, each eigenvector of the Laplacian matrix tends to localize on a subset of network nodes having similar numbers of edges, namely, the components of each Laplacian eigenvector take relatively large values only on a particular subset of nodes whose degrees are close. Although this localization property has significant consequences for dynamical processes on random networks, a clear theoretical explanation has not yet been established. Here we analyze the origin of localization of Laplacian eigenvectors on random networks by using a perturbation theory. We clarify how heterogeneity in the node degrees leads to the eigenvector localization and that there exists a clear degreeeigenvalue correspondence, that is, the characteristic degrees of the localized nodes essentially determine the eigenvalues. We show that this theory can account for the localization properties of Laplacian eigenvectors on several classes of random networks, and argue that this localization should occur generally in networks with degree heterogeneity.
Introduction
Localization of eigenmodes is a well known phenomenon in many fields of science, with the Anderson localization in disordered systems providing a prominent example^{1,2}. The localization properties of eigenvectors in random matrix models of disordered media have been studied intensively^{3}. Applications of the random matrix theory to crosscorrelation matrices of economic data have also been performed, and have revealed localized eigenvectors, which implies functional substructures in the data^{4}.
In this study, we focus on the Laplacian matrices of random networks, which describes diffusion processes in various models of networkorganized systems, such as random walks, epidemic spreading, information flow, coupled nonlinear oscillators, and activatorinhibitor systems^{5,6,7,8}. In many cases, the properties of the Laplacian eigenvectors play decisive roles in the network dynamics. Similar to the Fourier eigenmodes of the ordinary Laplacian operator in spatially extended systems, the Laplacian eigenvectors provide natural “coordinates” for describing the dynamics on networks.
A remarkable property of the Laplacian eigenvectors on random networks is their localization with respect to the node degrees. Namely, the components of each eigenvector take relatively large values on a particular subset of nodes, while taking small values otherwise. Moreover, the localized nodes have similar degrees, i.e., numbers of edges, and this characteristic degree corresponds closely with the Laplacian eigenvalue (We say that a pair of nodes have “similar degrees” when the difference in their degrees is sufficiently smaller than the range of the entire degree distribution; see the Results section for a discussion on the similarity in node degrees). This localization property has significant consequences for the dynamics of networkorganized systems. For example, in the pattern formation in networkorganized reactiondiffusion systems, the Laplacian eigenvectors determine the critical modes at the onset of instability and often dominate the developed patterns in the nonlinear regime^{9,10,11}. In coupled oscillators on networks, the Laplacian eigenvectors determine the synchronization dynamics of the oscillators^{12,13,14}.
The localization of the Laplacian eigenvectors was first reported by McGraw and Menzinger in their numerical analysis of coupled phase oscillators on several classes of random networks^{13}, and later utilized in the analysis of networkTuring patterns^{9,10,11}. Localized Laplacian eigenvectors have also been investigated in a few specific classes of networks^{15,16}. However, despite its apparent ubiquity and importance in dynamical processes on networks^{7,8,9,10,11,12,13,14,17}, a clear theoretical explanation of the localization mechanism has so far been lacking. Therefore, the class of networks in which this localization can be observed remains unclear.
In this study, we propose a simple theoretical approach to analyze the origin of the localization of Laplacian eigenvectors for a general class of networks. Based on the perturbation analysis of the Laplacian matrix, we argue that the localization should generally occur in networks with degree heterogeneity. For illustration, we analyze localization properties of Laplacian eigenvectors in several classes of random networks.
Results
Laplacian matrix and its eigenvectors
We consider a network consisting of N nodes. The network topology is specified by a N × N adjacency matrix A, whose element \({A}_{ij}\) takes a value of 1 if there is an edge between nodes i and j, and 0 otherwise (\(i,j=1,2,\cdots ,N\)). We assume that the network is connected (a path exists between arbitrary nodes), and that the connection is nondirected, i.e., \({A}_{ij}={A}_{ji}\). In this study, we define the Laplacian matrix \({\bf{L}}=\{{L}_{ij}\}\) of the network as \({L}_{ij}={A}_{ij}{k}_{i}{\delta }_{i,j}\), where \({k}_{i}={\sum }_{j=1}^{N}{A}_{ij}\) is the degree of the ith node, i.e., the number of edges, and \({\delta }_{i,j}\) is the Kronecker’s delta symbol^{18,19}. As we will show, it is convenient to sort the node indices {i} in decreasing order of the degree k _{ i }, so that inequalities \({k}_{1}\ge {k}_{2}\ge \cdots \ge {k}_{N}\) hold. We denote the average degree of the network as \(\langle k\rangle ={\sum }_{i=1}^{N}{k}_{i}/N\).
Diffusion processes on the network are described by the Laplacian matrix. Suppose that each network node is occupied by some substance X, and denote its concentration on node i as \({[X]}_{i}={x}_{i}\). The change in the concentration by the diffusive transportation of X is described as \(d{x}_{i}/dt={\sum }_{j=1}^{N}{L}_{ij}{x}_{j}={\sum }_{j=1}^{N}{A}_{ij}({x}_{j}{x}_{i})\), where the flux of the substance from node j to node i is proportional to the concentration difference x _{ j } − x _{ i } (Fick’s law).
The eigenvector \({\overrightarrow{\varphi }}^{(\alpha )}=({\varphi }_{1}^{(\alpha )},{\varphi }_{2}^{(\alpha )},\cdots ,{\varphi }_{N}^{(\alpha )})\) and the eigenvalue \({{\rm{\Lambda }}}^{(\alpha )}\) of the Laplacian matrix L satisfy the eigenvalue equation
where \(\alpha =\mathrm{1,}\,2,\cdots ,N\) is the index of the eigenvector. The eigenvectors can be orthonormalized as \({\sum }_{i=1}^{N}{\varphi }_{i}^{(\alpha )}{\varphi }_{i}^{(\beta )}={\delta }_{\alpha ,\beta }\) for \(\alpha ,\beta =\mathrm{1,}\,2,\cdots ,N\), because L is a real symmetric matrix. The Laplacian matrix L is negative semidefinite, i.e., \({\sum }_{i,j}{x}_{i}{L}_{ij}{x}_{j}={\sum }_{i,j}{A}_{ij}{({x}_{i}{x}_{j})}^{2}\le 0\) is satisfied for any vector \(\overrightarrow{x}=({x}_{1},\cdots ,{x}_{N})\) ^{6}. Therefore, all Laplacian eigenvalues are nonpositive, and only one of them, which corresponds to the uniform eigenvector \((1,\cdots ,1)/\sqrt{N}\), takes 0 because the network is connected. The eigenvector indices {α} are also sorted in increasing order of the Laplacian eigenvalues so that \({{\rm{\Lambda }}}^{\mathrm{(1)}}\le {{\rm{\Lambda }}}^{\mathrm{(2)}}\le \cdots \le {{\rm{\Lambda }}}^{(N)}=0\) hold.
Localization of Laplacian eigenvectors
First, let us illustrate the Laplacian eigenvectors for several classes of random networks. In Fig. 1, all eigenvectors (except the uniform eigenvector \({\overrightarrow{\varphi }}^{(N)}=(1,\cdots ,1)/\sqrt{N}\) with \({{\rm{\Lambda }}}^{(N)}=0\), which has exceptional characteristics and is excluded from the analysis) of the scalefree network generated by the BarabásiAlbert preferential attachment algorithm (BA)^{20}, classical ErdösRényi random network (ER)^{21}, and real neural network of C. elegans (CE)^{22,23} are displayed in the contour plot, where the horizontal axis is the node index and the vertical axis is the eigenvector index. We show the absolute value \({\varphi }_{i}^{(\alpha )}\) of the eigenvector components, because each component is statistically symmetric with respect to \({\varphi }_{i}^{(\alpha )}\iff {\varphi }_{i}^{(\alpha )}\). For the BA network, three typical eigenvectors with different α are also shown for illustration. For the CE network, we focus only on the connectivity and symmetrize the original network, which consists of 277 neurons with directed connections, by defining the adjacency matrix as \({A}_{ij}={A}_{ji}=1\) if there is a edge between nodes i and j.
Remarkably, clear diagonal structures are observed in all figures. Because the nodes are sorted by their degrees, this means that, in each eigenvector, only the nodes sharing similar degrees take large vector components, while other nodes have very small components. Indeed, for each eigenvector shown in Fig. 1, the mean difference in the degrees of the localized nodes where \({\varphi }_{i}^{(\alpha )} > 0.1\) is 1.33 for the BA network, 2.01 for the ER network, or 2.10 for the CE network. Each of these numbers is much smaller than the entire range of the degree distribution in each network, k _{1} − k _{ N }, which is 106 for the BA, 25 for the ER, or 75 for the CE.
Moreover, the visible diagonal structures indicate that the characteristic degree of each localized subset linearly correlates with the eigenvalue index, which is also sorted by their eigenvalues. Thus, clear degreeeigenvalue correlation exists in the Laplacian eigenvectors (See Fig. 2(c–f)). It is also notable that the patterns of the localization are qualitatively different among the networks. In the BA network [Fig. 1(a)], the localization is stronger near the hubs, i.e., the nodes with large degrees (e.g., α = 50 in panel (b)), while comparatively weak at the peripheries, i.e., the nodes with small degrees (e.g., α = 350 in panel (b)). In contrast to the BA network, in the ER [Fig. 1(c)] and CE networks [Fig. 1(d)], the localization is stronger both at hubs and peripheries, and weaker at the intermediate nodes. In McGraw and Menzinger^{13}, the level of localization has been quantified by using the inverse participation ratio, i.e., \({\sum }_{i}{({\varphi }_{i}^{(\alpha )})}^{4}/{\{{\sum }_{i}{({\varphi }_{i}^{(\alpha )})}^{2}\}}^{2}\), a standard quantity used in the analysis of Anderson localization.
Perturbation analysis of the Laplacian matrix
To analyze the origin of this intriguing localization property, we apply the perturbation theory^{24,25} to the Laplacian matrix. A similar perturbation approach was used by Kim and Motter to analyze the Laplacian eigenvalues of scalefree networks^{26}.
In the present problem, the Laplacian matrix L has two types of elements of distinct orders. The diagnoal elements \({k}_{i}{\delta }_{i,i}\) are order 〈k〉, while the nondiagional elements, which take 0 or 1, are \({\mathcal{O}}\mathrm{(1)}\). By introducing an expansion parameter \(\varepsilon ={\langle k\rangle }^{1}\), we can rewrite the Laplacian matrix as L = L _{0} + ε L _{1}, whose elements \({L}_{\mathrm{0,}ij}={k}_{i}{\delta }_{i,j}\) and \({L}_{\mathrm{1,}ij}=\langle k\rangle {A}_{ij}\) are of the same order, \({\mathcal{O}}(\langle k\rangle )\). When the network is sufficiently dense, i.e., \(\langle k\rangle \gg 1\), the expansion parameter ε is small, and it is expected that the perturbation theory yields reasonable approximation of the Laplacian eigenvectors.
For convenience, we employ the braket notation to denote the Laplacian eigenvector, i.e., \({\overrightarrow{\varphi }}^{(\alpha )}=\alpha \rangle \), and drop the summation symbol as \({\sum }_{j=1}^{N}{L}_{ij}{\varphi }_{j}^{(\alpha )}=L\alpha \rangle \). Expanding the Laplacian eigenvectors α〉 and eigenvalues \({{\rm{\Lambda }}}^{(\alpha )}\) in series of ε as \(\alpha \rangle ={\alpha \rangle }_{0}+\varepsilon {\alpha \rangle }_{1}+{\varepsilon }^{2}{\alpha \rangle }_{2}+\cdots \) and \({{\rm{\Lambda }}}^{(\alpha )}={{\rm{\Lambda }}}_{0}^{(\alpha )}+\varepsilon {{\rm{\Lambda }}}_{1}^{(\alpha )}+{\varepsilon }^{2}{{\rm{\Lambda }}}_{2}^{(\alpha )}+\cdots \), and substituting into the eigenvalue equation (1), the following set of equations is obtained up to \({\mathcal{O}}({\varepsilon }^{2})\):
Let us first consider the unperturbed system (2). One can easily find that the eigenvectors \({\alpha \rangle }_{0}\) and eigenvalues \({{\rm{\Lambda }}}_{0}^{(\alpha )}\) are given exactly as
for \(\alpha =\mathrm{1,}\,\mathrm{...,}\,N\). Each eigenvector is characterized by a single nonvanishing element at the network node \(i=\alpha \), and the corresponding eigenvalue \({{\rm{\Lambda }}}_{0}^{(\alpha )}\) is simply equal to the negative of the characteristic node degree \({k}_{\alpha }\). Thus, strictly localized eigenvectors are obtained at the zerothorder, where the network topology is completely ignored and the nondiagonal elements are assumed to be vanishingly small. Note that \({{\bf{L}}}_{0}\) is no longer a Laplacian matrix because its row sums do not vanish. Also, unlike \({{\rm{\Lambda }}}^{(N)}\), which always vanishes, the zerothorder eigenvalue \({{\rm{\Lambda }}}_{0}^{(N)}\) is not zero but equal to \({k}_{N}\), i.e., the smallest degree of the network.
In order to analyze the localization property of the Laplacian eigenvectors, we should consider the higherorder perturbation terms and, in particular, the fact that networks generally possess multiple nodes with the same degrees. From the zerothorder solution (5), this indicates that the zerothorder eigenvectors are degenerate. Therefore, we need to employ the degenerate perturbation theory^{24,25}. From Eqs (2–4), we can compute the approximate eigenvectors and eigenvalues by the first and secondorder degenerate perturbation theory, respectively. The complete derivation of the perturbation corrections for a general class of degenerate systems has been reported, e.g., in ref.^{25}. See Methods and Supplementary Information for details. Accuracy of the perturbation approximation is also discussed in the Methods.
Approximate eigenvectors
We now apply the perturbation theory to the networks used in Fig. 1, and demonstrate that it can predict the localization properties of random networks. The results are shown in Figs 2–6.
Figure 2(a–c) show the node degrees of the BA, ER, and CE networks as functions of the node index. Because the node indices are sorted in decreasing order of degrees, the curves monotonically decreases with the node index. Figure 2(d–f) show scatter plots of degreeeigenvalue pairs, \(({k}_{i},{{\rm{\Lambda }}}^{(\alpha =i)})\) \((i=1,\mathrm{...},N)\), for the three networks. We can observe that the data points approximately lie along the diagonal line in each figure, implying that the eigenvalues and node degrees are closely correlated in these networks. Such correlation between eigenvalues and degrees has also been reported in a preceding study^{27}.
Now, Fig. 2(g–i) show the zeroth and secondorder approximations of the Laplacian eigenvalues. We can see that the zerothorder result already provides a good approximation to the true Laplacian eigenvalues for the BA and CE networks. For the ER network, the zerothorder eigenvalues somewhat deviate from the true eigenvalues, but higherorder approximation gives closer values. Thus, the perturbation theory accounts for the Laplacian eigenvalues of these networks, which indicates that they are essentially determined by the node degrees.
Figure 3 displays the approximated eigenvectors as a function of the eigenvector index α and the node index i, similarly to Fig. 1. As can be seen, the firstorder predictions are in good qualitative agreement with the true Laplacian eigenvectors shown in Fig. 1. The diagonal structures indicating localization of eigenvectors are well reproduced for all networks. Moreover, the different patterns of localization among the networks are correctly reproduced. That is, the localization is stronger at hubs and weaker at peripheries in the BA network, while it is strong both at hubs and peripheries and weak at the intermediate nodes in the ER and CE networks.
The slightly broadened localization patterns along the diagonal line can be interpreted as follows. If some nodes in the network share the same degree, the corresponding zerothorder eigenvectors are degenerate. The eigenvectors in the degenerate subspace are mutually mixed, yielding blockdiagonal structures of various sizes in the contour plot. The size of each blockdiagonal component is equal to the number of the degenerate eigenvectors, i.e., the number of nodes sharing the same degree. Therefore, the degree distribution of the network determines the pattern of localization.
Indeed, in the BA network, the degrees obey a scalefree distribution, where only a small number of nodes have large degrees (hubs) and the majority of the nodes have small degrees (peripheries) [Fig. 3(d)]. Correspondingly, the degeneracy of the degrees, i.e., the zerothorder eigenvalues, is small for the hubs and large for the peripheries. Therefore, the localization is stronger at the hubs than that at the peripheries because less eigenvectors are involved.
In contrast, the ER network has a binomial distribution of the degrees [Fig. 3(e)]. The majority of the nodes belong to the intermediate degrees, and the hubs and peripheries are composed of relatively small numbers of nodes. This leads to stronger localization at both hubs and peripheries, and weaker localization at intermediate nodes, in contrast to the BA case. Similarly, in the CE network, the degrees obey a binomiallike distribution [Fig. 3(f)], so the localization pattern is also similar to that of the ER network.
In Fig. 4, the approximate eigenvectors of the BA network are compared with the true eigenvectors for several values of α. The eigenvectors shown in Figs (a) and (b), which correspond to the degeneracy types (A) and (B) (See Methods), show good agreement with the perturbation theory. However, the approximate eigenvectors in Figs (c) and (d), which are of the degeneracy type (C), are not in good quantitative agreement with the true vectors (although they exhibit qualitatively similar patterns). Thus, the degeneracy of the eigenvectors affects the performance of the perturbation approximation.
Indeed, quantitative nodewise comparison of the true and approximate vectors yields considerable discrepancy. The fact that the essential localization property of the vectors is qualitatively reproduced in Fig. 3 suggests that the true and approximate eigenvectors share similar characteristics when they are averaged over degenerate nodes and eigenvectors. In order to evaluate the performance of approximation while excluding the effect of degeneracy, we construct reduced degreewise vectors from the true nodewise Laplacian eigenvectors, where vector components of the original nodewise Laplacian vectors are averaged over degenerate nodes having the same degree and over the degenerate eigenmodes having the same zeroth order eigenvalues as follows:
Here, the degree index k runs from the minimum degree to the maximum degree of the network nodes, the reduced eigenvalue index β runs from the minimum to maximum of the zeroth order eigenvalues, \({N}_{k}\) is the number of degenerate nodes with degree k, and \({N}_{\beta }\) is the number of degenerate eigenmodes with the zeroth order eigenvalue \({k}_{\beta }\), respectively. We then calculate the correlation coefficient σ between the true and approximate eigenvectors, defined by
where \({\tilde{v}}_{k}^{(\beta )}\) is the reduced degreewise vector obtained from the approximate eigenvectors. For comparison, we also generate N independent random eigenvectors whose components are randomly drawn from a uniform distribution over [−1, 1], normalize the vectors so that their norms become equal to 1, and calculated their correlations similarly. We exclude the uniform eigenvector \({\overrightarrow{\varphi }}^{(N)}=(\mathrm{1,}\,1,\,\cdots ,1)/\sqrt{N}\) from the analysis, which is exceptional and cannot be predicted by the perturbation theory.
Figure 4(e–g) show the correlation coefficient σ between the true and approximate eigenvectors with respect to the reduced eigenvalue index β. As can be seen in the figures, the correlation coefficient σ takes large values for nondegenerate eigenmodes, which are much larger than the correlation coefficient for random vectors and thus indicate similarity between the true and approximate vectors. It can be seen that nondegenerate vectors show larger correlations than degenerate vectors. It can be also observed that σ is higher near hubs for the BA, and near hubs and peripheries for the ER and CE. For other eigenvectors, σ can be as small as those of random vectors, indicating that the perturbation theory does not predict some of the eigenvectors well.
We stress that, although the correlation coefficients can be small for some of the eigenvectors, that is, the perturbation approximation does not predict them quantitatively, essential qualitative properties of the true eigenvectors such as the localizing nodes and the degree of localization are still reproduced well. This is because such properties are mainly determined by the degree of the nodes in the same degenerate block, which share statistically similar connectivities to the rest of the network. See Methods for the discussion on the accuracy of the perturbation approximation.
Thus, the perturbation theory can account for the eigenvector localization and degreeeigenvalue correspondence reasonably well. It reveals how degree heterogeneity and degeneracy lead to the eigenvector localization and the degreeeigenvalue correspondence. Furthermore, it clarifies why the representation of the Laplacian eigenvectors in Fig. 1, with respect to the eigenvector index α and the node index i both sorted in decreasing order of the eigenvalues and degrees, yields the clearly visible localized structures.
Our perturbation analysis also explains how similar the degrees of the nodes should be in order that the Laplacian eigenvector localizes on these nodes. From Eqs (9–11), we observe that the difference in the zeroth order eigenvalues in the denominator, \({{\rm{\Lambda }}}_{0}^{(\alpha )}{{\rm{\Lambda }}}_{0}^{(\beta )}\), which is equal to the difference in the node degrees from Eq. (5), should be sufficiently smaller than the maximal range of the eigenvalues, \({{\rm{\Lambda }}}_{0}^{(N)}{{\rm{\Lambda }}}_{0}^{\mathrm{(1)}}\), which is equal to \({k}_{{\rm{\max }}}{k}_{{\rm{\min }}}\), in order to give a dominant contribution to the firstorder correction to the eigenvector. (If this is not the case, all nodes in the network will give similar contributions to the perturbation correction and localization will not be observed). Thus, the degrees of the nodes should be similar in the sense that their difference is much smaller than the range of the entire degree distribution, \({k}_{1}{k}_{N}\), as we mentioned in the introduction.
Weighted and directed networks
Although we have so far presented the results only for nondirected and nonweighted networks, our analysis can straightforwardly be extended to directed and weighted networks. For directed networks, the adjacency matrix A is generally asymmetric. The weight of the edge from node j to node i is specified by the element W _{ ij } of the weight matrix W. The Laplacian matrix of such a network is defined as
where \({k}_{i}^{{\rm{out}}}={\sum }_{j=1}^{N}{W}_{ji}{A}_{ji}\) is the outgoing degree of the ith node. The diagonal elements of this Laplacian matrix is of the order \({\mathcal{O}}(\langle {k}^{{\rm{out}}}\rangle )\), while nondiagonal elements are \({\mathcal{O}}({W}_{ij})\). Thus, if the network is sufficiently dense, \(\langle {k}^{{\rm{out}}}\rangle \gg {W}_{ij}\) holds generally.
By introducing an expansion parameter \(\varepsilon ^{\prime} =\langle W\rangle /\langle {k}^{{\rm{out}}}\rangle \), we can rewrite the Laplacian matrix as
whose elements
are of the same order, \({\mathcal{O}}(\langle {k}^{{\rm{out}}}\rangle )\). Thus, one can follow the perturbation analysis as described above with the generalized Laplacian matrix L.
As an illustrative example, Fig. 5 compares the true Laplacan eigenvectors and the result of the perturbation approximation of the real asymmetric neural network of C. elegans, which we used in Figs 1–4 after symmetrization. Note that the Laplacian matrix is now asymmetric and the elements of its eigenvectors can take complex values. We focus only on the localization pattern of the Laplacian eigenvectors and plot the absolute value of each vector component. As can be seen in the figure, the approximate eigenvectors can reproduce the localization pattern of the true eigenvectors qualitatively well.
Regular lattices
Our argument suggests that the blockdiagonal components representing the eigenvector localization can also be more or less observed even when the network is not random, but is formed by nodes with nonidentical degrees. In order to verify this statement, we here consider a regular lattice network as shown in Fig. 6(a), which is composed of three types of nodes with different degrees. Specifically, one third of the nodes have degree k = 6, another one third have degree k = 4, and the rest have degree k = 2.
The zerothorder unperturbed result is shown in Fig. 6(b). All eigenvectors degenerate into three classes at this stage, corresponding to the characteristic degrees k = 6, k = 4, and k = 2. Higherorder perturbations solve this degeneracy and mix the eigenvectors in each subset into three blocks, corresponding to k = 6, k = 4, and k = 2. Thus, at the firstorder perturbation, the Laplacian eigenvectors show blockdiagonal structures in the contour plot as shown in Fig. 6(c).
The above prediction is in good agreement with the true Laplacian eigenvectors obtained by direct numerical calculation, shown in Fig. 6(d). Thus, the degree heterogeneity generally leads to localized eigenvectors even in regular lattice networks. This result also suggests that the degree heterogeneity is the origin of the localization property of the Laplacian eigenvectors.
Discussion
The Laplacian eigenvectors of networks with degree heterogeneity generally exhibit localization on the subset of nodes with close degrees and the Laplacian eigenvalues show clear degreeeigenvalue correspondence. We have proposed a simple explanation for the localization property of Laplacian eigenvectors using the degenerate perturbation theory. It clarifies how degree heterogeneity and degeneracy lead to the eigenvector localization and degreeeigenvalue correspondence. We analyzed three kinds of random networks with different statistical properties, and confirmed that our approach can reasonably account for the true Laplacian eigenvectors. We have also shown that the analysis can straightforwardly be extended to directed and weighted networks.
Our results show that the node degrees in heterogeneous networks correspond to the wavenumbers in regular lattices. Therefore, the degree of the node can play an essential role as the “natural coordinate” in describing the dynamics or patterns on networks with heterogeneous degree distributions, because they determine the eigenvalue and the subset of nodes that participate in the eigenvector. We conjecture that this partly accounts for why various network dynamics, plotted with respect to node degrees, often exhibit ordered patterns and provide us with physical interpretations.
In this study, we used the C. elegans neuronal network to illustrate the generality of the eigenvector localization and did not analyze its particular functional structures. Though extraction of functional structures from the C. elegans neuronal network is beyond the scope of the present study, we can observe a possible sign of such structures from the eigenvalues of the C. elegans network shown in Fig. 2(f); that is, there exists a small cluster of eigenvalues (3 ≤ α ≤ 12) separated from other eigenvalues, and correspondingly a tiny block structure exists on top of the diagonal structure in the Laplacian eigenvectors in Fig. 1(d). Such detailed structures of the Laplacian eigenvectors could reflect some functional structure of the neuronal network of C. elegans. In analyzing such detailed structures, broad degree heterogeneity, which yields the diagonal localized structure, might be disturbing, and spectral clustering methods based on the normalized Laplacian matrix^{28}, which removes the effect of degree heterogeneity, could provide more useful information of the network.
Finally, we note that the localization property of eigenvectors is not restricted to network Laplacian matrices. We have recently formulated the advection equation for random networks^{29} and found that the eigenvectors of the advection matrix are also localized on a subset of nodes. This localization can also be accounted for by a similar perturbation analysis of the network, and the homogenization process on the network due to advection can be clearly visualized when plotted with respect to the node degrees. Further investigation of eigenvector localization on networks will provide us with insights into complex dynamics on networks.
Methods
Degenerate perturbation theory
Following a standard argument from quantum mechanics, we classify each eigenvector into the following three types according to its degeneracy: (A) nondegenerate, (B) degeneration that is solved at the first order, and (C) otherwise. We compute the approximate eigenvectors and eigenvalues by the first and secondorder perturbation theory, respectively. For each case, the perturbation corrections are given as follows (see Supplementary Information for the derivation).
For type (A), the first order correction is
The first order correction to the eigenvalue vanishes, i.e., \({{\rm{\Lambda }}}_{1}^{(\alpha )}={}_{0}\langle \alpha {L}_{1}{\alpha \rangle }_{0}=0\), because the diagonal elements of L _{1} are zero. The second order correction to the eigenvalue is given by \({{\rm{\Lambda }}}_{2}^{(\alpha )}={}_{0}\langle \alpha {L}_{1}{\alpha \rangle }_{1}\).
For type (B), we denote the degenerate eigenvectors corresponding to the same eigenvalue \({{\rm{\Lambda }}}_{0}^{(\alpha )}\) at the zeroth order as \({\alpha }_{1},\cdots ,{\alpha }_{m}\). We introduce new zerothorder eigenvectors so that the degeneration is solved at the firstorder perturbation as \({{\tilde{\alpha }}_{i}\rangle }_{0}={\sum }_{j=1}^{m}{b}_{i,j}{{\alpha }_{j}\rangle }_{0},\) where the mixing coefficients \({b}_{i,j}\) are the eigenvectors of the matrix V defined by \({V}_{ij}={}_{0}{\alpha }_{i}\rangle {L}_{1}{{\alpha }_{j}\rangle }_{0}\) (\(i,j=\mathrm{1,}\,\cdots ,m\)). Namely, they satisfy the secular eigenvalue equation \({\sum }_{j=1}^{m}{V}_{kj}{b}_{i,j}={{\rm{\Lambda }}}_{1}^{({\alpha }_{i})}{b}_{i,k},\) where \({{\rm{\Lambda }}}_{1}^{({\alpha }_{i})}\) gives the firstorder correction to the Laplacian eigenvalue. The firstorder correction to the eigenvector is
where the summation symbol with \(\beta \ne \alpha \) indicates that the index β runs over all eigenvectors except for the degenerate ones, i.e., \({\alpha }_{1},\cdots ,{\alpha }_{m}\). The second order correction to the eigenvalue is given by \({{\rm{\Lambda }}}_{2}^{({\alpha }_{i})}={}_{0}\langle {\tilde{\alpha }}_{i}{L}_{1}{{\tilde{\alpha }}_{i}\rangle }_{1}\).
For type (C), suppose that a subset of the eigenvalues \({\tilde{\alpha }}_{1},\cdots ,{\tilde{\alpha }}_{n}\) (\(n\le m\)) is still degenerate at the first order perturbation. In this case, the zerothorder eigenvector is further redefined as \({{\tilde{\tilde{\alpha }}}_{i}\rangle }_{0}={\sum }_{j\mathrm{=1}}^{n}{c}_{i,j}{{\tilde{\alpha }}_{j}\rangle }_{0},\) where \({c}_{i,j}\) are given by the eigenvectors of the matrix W defined as \({W}_{kj}={\sum }_{\beta \ne \alpha }({}_{0}{\tilde{\alpha }}_{k}\rangle {L}_{1}{\beta \rangle }_{0}{}_{0}\langle \beta {L}_{1}{{\tilde{\alpha }}_{j}\rangle }_{0})/({{\rm{\Lambda }}}_{0}^{(\alpha )}{{\rm{\Lambda }}}_{0}^{(\beta )})\). The firstorder correction to the eigenvector is given by
The first order correction \({{\rm{\Lambda }}}_{1}^{({\alpha }_{i})}\) to the eigenvalue is given as in (B), and the second order correction \({{\rm{\Lambda }}}_{2}^{({\alpha }_{i})}\) is determined from \({\sum }_{j\mathrm{=1}}^{m}{W}_{kj}{c}_{i,j}={{\rm{\Lambda }}}_{2}^{({\alpha }_{i})}{c}_{i,k}\).
If some eigenvectors are still degenerate even at the secondorder perturbation, one can further introduce new zerothorder eigenvectors so that the degeneracy is solved at the higherorder perturbation. However, for simplicity, we do not consider the higherorder perturbations. Therefore, the eigenvectors are not completely determined in this case.
Accuracy of the perturbation approximation
As is well known, it is generally difficult to prove the convergence of the perturbation series. Accuracy of the perturbation approximation may roughly be assessed by looking at the ratio \({}_{0}\langle \beta {L}_{1}{\alpha \rangle }_{0}/{{\rm{\Lambda }}}_{0}^{(\alpha )}{{\rm{\Lambda }}}_{0}^{(\beta )}\) for each pair of nondegenerate eigenmodes \(\alpha \) and \(\beta \). If this ratio is sufficiently small, the contribution of the zerothorder eigenvector \(\beta {\rangle }_{0}\) to the firstorder correction \(\alpha {\rangle }_{1}\) will be accurately evaluated.
In the present case, \({}_{0}\langle \beta {L}_{1}{\alpha \rangle }_{0}={A}_{\beta \alpha }\) is the element of the adjacency matrix and takes either 1 or 0, while \({{\rm{\Lambda }}}_{0}^{(\alpha )}{{\rm{\Lambda }}}_{0}^{(\beta )}={k}_{\alpha }{k}_{\beta }\) is the difference between the characteristic degrees of the corresponding eigenvectors and is greater than 1 (See Methods). Therefore, if \({A}_{\beta \alpha }\mathrm{=1}\) and k _{ β } is close to k _{ α }, the above ratio may not be small, namely, the contribution from \(\beta {\rangle }_{0}\) to the firstorder correction \(\alpha {\rangle }_{1}\) may be inaccurate at the nodes with degree k _{ β }.
We note, however, that the perturbation theory can still qualitatively account for the localization property in such cases. In Fig. 1, the eigenvector \(\alpha \rangle \) is localized at the diagonal nodes whose indices satisfy \(i\simeq \alpha \) and whose degrees are close to k _{ α }, because the node indices {i} are sorted so that \({k}_{1}\ge {k}_{2}\ge \cdots \ge {k}_{N}\). Our main aim is to explain that \(\alpha \rangle \) is almost vanishing at the nondiagonal nodes with \(i\alpha \gg 1\). The degree k _{ β } of such nondiagonal nodes are generally far from k _{ α } for networks with degree heterogeneity, so that the above ratio would generally be small and the firstorder correction \({\alpha \rangle }_{1}\) would reliably be obtained for such nodes. Thus, the perturbation theory can account for why the eigenvector takes tiny components at nondiagonal nodes even if they can give inaccurate results for diagonal nodes.
Indeed, as explained in the Results, the theory can reproduce the true eigenvalues very accurately and account for the localization property qualitatively well for all three networks shown in Fig. 1. Furthermore, it can even predict precise localizing patterns quantitatively well for some of the eigenvectors.
Change history
06 September 2017
A correction to this Article has been published and is linked from the HTML version of this paper. The error has not been fixed in the paper.
References
 1.
Anderson, P. W. Absence of Diffusion in Certain Random Lattices. Phys. Rev. 109, 1492–1505, doi:10.1103/PhysRev.109.1492 (1958).
 2.
Grebenkov, D. S. & Nguyen, B.T. Geometrical structure of Laplacian eigenfunctions. SIAM Review 55, 601–667, doi:10.1137/120880173 (2013).
 3.
Mirlin, A. D., Fyodorov, Y. V., Dittes, F.M., Quezada, J. & Seligman, T. H. Transition from localized to extended eigenstates in the ensemble of powerlaw random banded matrices. Phys. Rev. E 54, 3221–3230, doi:10.1103/PhysRevE.54.3221 (1996).
 4.
Plerou, V. et al. Random matrix approach to cross correlations in financial data. Phys. Rev. E 65, 066126, doi:10.1103/PhysRevE.65.066126 (2002).
 5.
Barrat, A., Barthélemy, M. & Vespignani, A. Dynamical Processes on Complex Networks (Cambridge, 2008).
 6.
Mohar, B. The Laplacian spectrum of graphs, in Graph Theory, Combinatorics, and Applications Vol. 2, ed. Y. Alavi et al., 871 (Wiley, 1991).
 7.
Jost, J. Dynamical networks, in Networks: from biology to theory, ed. Jianfeng, F., Jost, J. & Minping, Q. (Springer, 2007).
 8.
Newman, M. Networks: an introduction. (Oxford university press, 2010).
 9.
Nakao, H. & Mikhailov, A. S. Turing patterns in networkorganized activatorinhibitor systems. Nature Physics 6, 544–550, doi:10.1038/nphys1651 (2010).
 10.
Wolfrum, M. The Turing bifurcation in network systems: Collective patterns and single differentiated nodes. Physica D 241, 1351–1357, doi:10.1016/j.physd.2012.05.002 (2012).
 11.
Hata, S., Nakao, H. & Mikhailov, A. S. Dispersalinduced destabilization of metapopulations and oscillatory Turing patterns in ecological networks. Scientific Reports 4, 3585, doi:10.1038/srep03585 (2014).
 12.
Arenas, A., DíazGuilera, A. & PérezVicente, C. J. Synchronization Reveals Topological Scales in Complex Networks. Phys. Rev. Lett. 96, 114102, doi:10.1103/PhysRevLett.96.114102 (2006).
 13.
McGraw, P. N. & Menzinger, M. Laplacian spectra as a diagnostic tool for network structure and dynamics. Phys. Rev. E 77, 031102, doi:10.1103/PhysRevE.77.031102 (2008).
 14.
Sun, J., Bollt, E. M. & Nishikawa, T. Master stability functions for coupled nearly identical dynamical systems. Europhys. Lett. 85, 60011, doi:10.1209/02955075/85/60011 (2009).
 15.
Saito, N. & Woei, E. On the Phase Transition Phenomenon of Graph Laplacian Eigenfunctions on Trees (Recent development and scientific applications in wavelet analysis). RIMS kokyuroku 1743, 77 (2011).
 16.
Shuman, D. I., Ricaud, B. & Vandergheynst, P. Vertexfrequency analysis on graphs. Appl. Comput. Harmon. Anal. 40, 260–291, doi:10.1016/j.acha.2015.02.005 (2016).
 17.
Yan, G. et al. Spectrum of controlling and observing complex networks. Nature Physics 11, 779–786, doi:10.1038/nphys3422 (2015).
 18.
Asllani, S., Challenger, J. D., Pavone, F. S., Sacconi, L. & Fanelli, D. The theory of pattern formation on directed networks. Nature Communications 5, 4517, doi:10.1038/ncomms5517 (2014).
 19.
Kouvaris, N. E., Kori, H. & Mikhailov, A. S. Traveling and Pinned Fronts in Bistable ReactionDiffusion Systems on Networks. PLoS ONE 7(9), e45029, doi:10.1371/journal.pone.0045029 (2012).
 20.
Barabási, A.L. & Albert, R. Emergence of Scaling in Random Networks. Science 286, 509–512, doi:10.1126/science.286.5439.509 (1999).
 21.
Erdös, P. & Rényi, A. On random graphs. Publicationes Mathematicae 6, 290 (1959).
 22.
Choe, Y., McCormick, B. H. & Koh, W. Network connectivity analysis on the temporally augmented C. elegans web: A pilot study. Society of Neuroscience Abstracts 30, 912.9 (2004).
 23.
Kaiser, M. & Hilgetag, C. C. Nonoptimal component placement, but short processing paths, due to longdistance projections in neural systems. PLoS Computational Biology 2, e95, doi:10.1371/journal.pcbi.0020095 (2006).
 24.
Sakurai, J. J. & Napolitano, J. J. Modern Quantum Mechanics (Pearson Education Ltd., London, 2013).
 25.
Hirschfelder, J. O. & Certain, P. R. Degenerate RS perturbation theory. J. Chem. Phys. 60, 1118–1137, doi:10.1063/1.1681123 (1974).
 26.
Kim, D.H. & Motter, A. E. Ensemble Averageability in Network Spectra. Phys. Rev. Lett. 98, 248701, doi:10.1103/PhysRevLett.98.248701 (2007).
 27.
Zhan, C., Chen., G. & Yeung, L. F. On the distributions of Laplacian eigenvalues versus node degrees in complex networks. Physica A Statistical Mechanics & Its Applications 389(8), 1779–53 (2010).
 28.
von Luxburg, U. A tutorial on spectral clustering. Stat. Comput. 17, 395–416, doi:10.1007/s112220079033z (2007).
 29.
Hata, S., Nakao, H. & Mikhailov, A. S. Advection of passive particles over flow networks. Phys. Rev. E 89, 020801(R), doi:10.1103/PhysRevE.89.020801 (2014).
Acknowledgements
We acknowledge A.S. Mikhailov for useful discussions. We thank JSPS KAKENHI (15K12111 to S.H., 16K13847 and 16H01538 to H.N.) and JST CREST Kokubu Project for financial support.
Author information
Affiliations
Contributions
S.H. and H.N. designed the study, carried out the analysis, and contributed to writing the paper.
Corresponding author
Ethics declarations
Competing Interests
The authors declare that they have no competing interests.
Additional information
Change History: A correction to this Article has been published and is linked from the HTML version of this paper. The error has not been fixed in the paper.
Publisher’s note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
An erratum to this article is available at https://doi.org/10.1038/s41598017062986.
Electronic supplementary material
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Hata, S., Nakao, H. Localization of Laplacian eigenvectors on random networks. Sci Rep 7, 1121 (2017). https://doi.org/10.1038/s41598017010100
Received:
Accepted:
Published:
Further reading

Pattern invariance for reactiondiffusion systems on complex networks
Scientific Reports (2018)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.