Abstract
Graph layout algorithms used in network visualization represent the first and the most widely used tool to unveil the inner structure and the behavior of complex networks. Current network visualization software relies on the force-directed layout (FDL) algorithm, whose high computational complexity makes the visualization of large real networks computationally prohibitive and traps large graphs into high energy configurations, resulting in hard-to-interpret “hairball” layouts. Here we use Graph Neural Networks (GNN) to accelerate FDL, showing that deep learning can address both limitations of FDL: it offers a 10 to 100 fold improvement in speed while also yielding layouts which are more informative. We analytically derive the speedup offered by GNN, relating it to the number of outliers in the eigenspectrum of the adjacency matrix, predicting that GNNs are particularly effective for networks with communities and local regularities. Finally, we use GNN to generate a three-dimensional layout of the Internet, and introduce additional measures to assess the layout quality and its interpretability, exploring the algorithm’s ability to separate communities and the link-length distribution. The novel use of deep neural networks can help accelerate other network-based optimization problems as well, with applications from reaction-diffusion systems to epidemics.
Similar content being viewed by others
Introduction
The numerical and analytical toolset of network science has played a key role in the scientific community’s ability to explore large complex systems, helping to predict and manage the COVID pandemic1,2, identify drug repurposing opportunities3, quantify traffic patterns in cities4, or understand the spread of fake news5,6. The first step of network analysis requires us to visualize the network of interest, a process supported by multiple software packages. The two most popular visualization packages, Cytoscape7 and Gephi8, have been used in over 40,000 publications, documenting the wide and cross-disciplinary role of graph layouts from systems biology to ecology, social sciences, and even literature. Yet, most visualization efforts are limited to networks of hundreds, occasionally a few thousand nodes, constrained by the computational complexity of the existing algorithms.
Network visualization relies on different implementations of the force-directed layout (FDL)9,10,11,12, a graph layout algorithm that treats links as springs that pull connected nodes close to each other and relies on short-range repulsive forces to avoid node overlap. Inspired by energy minimization in computational chemistry13, the final layout is obtained by minimizing the total potential energy using gradient descent. While widely effective for hundreds of nodes, the O(N2) computational cost per iteration makes the algorithm prohibitively expensive for larger networks. Hence, our ability to explore large real systems, like the protein-protein interaction network of a human cell with 20, 000 proteins and 300, 000 links, or networks emerging in social media with millions of nodes, is hindered by computational complexity, placing fundamental limitations in our ability to unveil their architecture. Attempts to visualize the structure of such large systems often result in “hairballs,” i.e. high energy layouts that are difficult to interpret and offer only limited insights into the architecture of the network. For this reason, visualizations of very large networks are rarely seen in journals or in the media.
Results
Here we propose an unsupervised machine learning based process to accelerate FDL, demonstrating the potential of deep learning to dramatically speed up graph layout. The key to our approach are Graph Neural Networks (GNN)14,15, which we use to reparametrize the energy-based optimization problem behind FDL. The resulting NeuLay algorithm is one or two orders of magnitude faster than the existing layout methods, opening up the possibility to quickly and reliably visualize large graphs. Importantly, the algorithm often converges to lower energies than those accessible by FDL, identifying more optimal layouts with clearer and more informative structures. We analytically show that the superior performance of NeuLay is driven by the neural networks’ ability to take advantage of the large-scale architecture of the network, resulting in quantifiable and visually apparent differences in the quality of the layout.
Neural Networks for graph layout
Let X = (x1, . . . xN) be a N × d matrix that captures the location xi of node i in d-dimensional Euclidean space. FDL performs gradient descent (GD) to minimize the total energy using a loss function \({{{{{{{\mathcal{L}}}}}}}}(X)\) (see Methods A.), formally written as
where ε is the learning rate and L is the graph Laplacian. Computing the LX (elastic forces) term has time complexity O(N) for sparse graphs, and O(N2) for dense graphs. VNN is repulsive energy helping avoid node overlap, with complexity O(N2), which can be decreased to \(O(N\log N)\) by the Barnes-Hut algorithm16, hence on dense graphs the bottleneck remains the calculation of the elastic forces (Supplementary Information B Computational Complexity). The core idea of our approach is to represent the node positions X as the output of a neural network, relying on two architectures: (1) NodeMLP, that starts from a high dimensional random embedding of the nodes and finds a map to the target dimension d = 3 of the layout; (2) NeuLay, that exploits the graph structure via Graph Convolutional Networks (GCN)14 (Fig. 1b, c and Methods B). NeuLay is a flexible framework and allows for the use of different GNN architecture other than GCN, such as Graph Attention (GAT)17 or Graph Network (GN)18. Our experiments (Fig. S11) show similar performance when using GCN, GAT or GN, in terms of speedup and final energy. Hence here we focus on GCN in NeuLay due to its simplicity. In NeuLay-2, we apply two GCN layers and then concatenate the layer outputs to obtain a high dimensional node embedding, which is then projected down to d = 3 dimensions. In our method, unlike more familiar uses of deep neural networks, retraining of the model is required for each graph layout as the training process is the optimization of the FDL which needs to be performed for every new graph layout. As we show next, the proposed GNN-based method improves computational complexity by reducing the number of iterations required for convergence, rather than reducing the per-step time complexity.
NeuLay offers more optimal layouts faster
To assess performance, we rely on two figures of merit: speed and quality. For speed, we examine the running time (‘wall-clock’ time). As a proxy for the layout quality, we explore several measures. The most natural one is the potential energy (loss value) of the final layout which we find to strongly correlate with the quality of the layout. But we also explore two additional measures, such as cluster separation and link length distribution. We begin by comparing the performance of FDL with the three proposed neural network models, NodeMLP, NeuLay, and NeuLay-2 for a simple cubic lattice (Fig. 1d, and Fig. S1b). We find that while NodeMLP and NeuLay offer significant speedup in laying out this network with a known optimal layout (Fig. 1f), NeuLay-2 with two GCN layers has the fastest convergence of the energy, prompting us to focus on this architecture hereafter. Furthermore, we measured the speedup using GPU hardware (see in the Fig. S11), consistently observing results similar to that reported in Fig. 1g, h.
We compared the speedup for four networks constructed using various graph generation models, like the Erdős-Rényi (ER) random graph algorithm, Barabási-Albert19 (BA) model, Stochastic Block Model (SBM)20, and Random Geometric Graphs (RGG)21. While these networks span drastically different topologies, sizes, and link densities, in all cases NeuLay-2 reaches the final state one to two orders of magnitude faster than FDL (Fig. 1h, i). We find that the speedup increases with the number of nodes and links (Fig. 1h, i), and falls with increasing network density (Fig. 1j). The speedup is particularly remarkable for graphs with a strong community structure, such as networks generated by the stochastic block model (SBM), and grid-like graphs, like the random geometric graph (RGG) (red symbols in Fig. 1h–j), compared to graphs that lack local structure, like the ER and BA networks (blue and green symbols in Fig. 1h–j). Yet, we observe speedup for each of those networks for a fixed density, finding that the speedup scales as N0.8 for 2D RGG, N0.3 for BA networks, and N0.2 for ER random graphs (Fig. 1h).
NeuLay-2 is not only faster, but also identifies better layouts. Indeed, while for small and simpler networks, like the cubic lattice (Fig. 1d), FDL and NeuLay-2 converge to indistinguishable energies, for larger networks NeuLay-2 identifies a deeper energy minimum compared to FDL (Fig. 1g). To systematically quantify this difference, we measured the ratio between the final energy of FDL and NeuLay-2 (ΔE = EFDL/ENeuLay−2). We find this ratio to increase with the size of the network (Fig. 1g), indicating that for large networks FDL gets trapped into a local sub-optimal configuration, successfully avoided by NeuLay-2. This ratio is especially large and increasing with N for BA and ER graphs, indicating that while NeuLay-2 may not show as high speedup over FDL for these networks as it does for more structured architectures, like RGG, it offers a significant advantage in terms of energy. As we show later, this energy difference has a dramatic impact on the quality of the final layout.
Large structures and outlier Eigenvalues help accelerate the layout
The higher speedups observed for networks generated by SBM and RGG, characterized by communities (SBM) and spatial proximity (RGG), suggests that the speedup is related to the leading eigenvalues of the adjacency matrix. To test this hypothesis, we analytically derived the speedup, finding that: (i) Speedup of NeuLay-2 is expected to increase with of the number of outlier eigenvalues; (ii) As a falsifiable test, we predict that removing the outlier eigenvalues should significantly reduce the speedup of NeuLay-2; (iii) Keeping only the outlier eigenvalues should be sufficient to achieve a speedup comparable to using the full spectrum.
We tested predictions (i)–(iii) on networks generated by SBM, for which the number of outlier eigenvalues equals the number of blocks (communities), allowing for direct control of the spectrum. Figure 2a shows the evolution of FDL vs NeuLay-2 for four communities, indicating that in NeuLay-2 the communities converge to their final positions by step 100, much earlier than in FDL. We find the speedup for SBM to grow with the number of blocks (communities) as \(\sim {n}_{block}^{0.77}\) (Fig. 2b). Plotting the speedup vs number of outlier eigenvalues, (Fig. 2c), we find that for SBM (as well as for RGG), the speedup increases as \(\sim {n}_{out}^{0.96}\) with the number of outliers, validating prediction (i). Yet, it is not clear if the correlation between the speedup and the number of outliers is causal, or it is driven by some other uncontrolled features of the graphs. If the outlier eigenvalues are truly responsible for the speedup, replacing A with a similar matrix that lacks the outliers must reduce the speedup. We, therefore, used the spectral expansion, A = Atop + Abulk, to separate the outliers, (Atop, Fig. 2e, red part of the histograms) and the rest of the modes (Abulk, Fig. 2e, blue part of the histograms, also see Methods C.). We find that for SBM and RGG, networks with multiple outliers, using only the outlier eigenvalues Atop results in higher speedup than using the full spectrum A, in line with prediction (iii) (Fig. 2d, red bars). In contrast, removing the outliers of the RGG spectrum and using only Abulk in NeuLay-2 dramatically reduces the speedup (Fig. 2d, blue bars), supporting prediction (ii). Finally, in line with the prediction (i), we do not observe a difference in the speedup by using A, Atop, or Abulk in networks that lack outlier eigenvalues, like networks generated by the ER and the BA model (Fig. 2d).
Note that in most network visualization problems we do not know the relevant eigenvalues, nor the eigenvalue combination that offers the best optimization. Yet, NeuLay-2 automatically identifies structures useful for accelerating FDL, and offers a fast convergence whether the network is dominated by outliers (like SBM and RGG), or lacks multiple outliers (BA, ER), hence improving the layout of arbitrary networks.
The quality of large network layouts
To illustrate the practical value of NeuLay-2, we used it to lay out multiple large real networks in three dimensions, d = 3, like the word association graph (WAN)22 (N = 10,617, L = 63,781), the human protein-protein interaction network (PPI)23 (N = 18,448, L = 322,285), Facebook social network24 (N = 22,470, L = 171,002), and the Internet at the autonomous system level25 (N = 22,963, L = 48,436). For comparison, we laid out multiple smaller networks as well, like the flavor network26 (N = 182, L = 641), boards of directors (public companies in Norway)27 (N = 854, L = 2745), mouse vascular network28 (N = 1558, L = 2352), US power grid29 (N = 4941, L = 6594), and the road network in Oakland30 (N = 6105, L = 7,029) (Fig. 1f–h). In all cases, we find NeuLay-2 to be an order of a magnitude faster than FDL, resulting in a 14-fold improvement in speed for WAN and a 13-fold improvement for PPI (Fig. 1f–h). Even more important is the fact that for each real network NeuLay-2 converges to a deeper energy state, a difference that is particularly remarkable for large networks, like the PPI and WAN. We observe the most dramatic improvement in the case of the Internet, for which previous successful visualization efforts had to reduce the network to its backbone31. Indeed, we find that FDL becomes trapped in a suboptimal layout, whose energy is 12% larger than the one identified by NeuLay-2 (Fig. 3a, and Fig. 4a, c). To ensure that this sub-optimal configuration is not a result of an accidental trapping of FDL in some local minima, we have re-run both NeuLay-2 and FDL ten times, starting from different initial configurations, each time observing largely indistinguishable time and energy curves (Fig. 3a).
The lower energy identified by NeuLay-2 has a visually detectable impact on how informative the layout is: while the higher energy NeuLay-2 layout has an observable local community structure (Fig. 4a, b), the FDL layout appears to be largely random (Fig. 4c, d), reminiscent of an unstructured hairball. To better assess how well the two layouts capture the inherent structure of the network, we used the Louvain algorithm32 to identify 36 communities in the Internet graph, coloring 12 of them on Fig. 4 for visual clarity. As Fig. 4a, c indicate (see also the video https://vimeo.com/732791412), while in the NeuLay-2 layout nodes in the same community are spatially co-localized, the FDL distributes the community members throughout the layout, failing to co-localize them. To quantify this difference, we measured the link length distribution of each community’s internal links (Fig. 3c, and Fig. S10), finding that the distribution identified by NeuLay-2 is much narrower than the one identified by FDL, confirming better spatial localization. These local differences also impact the global link length distribution of the two layouts (Fig. 3b), indicating that the FDL layout generates more long links than the NeuLay-2 layout, which also explains its larger elastic energy. Additionally, we have introduced a spatial similarity metric measuring how well the clusters are separated in the final layout compared to the FDL layout, finding that NeuLay-2 not only discovers but also better separates the clusters in the final layout (Fig. S8).
The higher energy state to which FDL converges does not necessarily result in a random layout. To see this, we apply a geometric randomization, by randomly exchanging the nodes, while keeping the physical coordinates of the layout and the adjacency matrix unchanged. We find that the link length distribution in the FDL layout is shorter than expected under geometric randomization (Fig. 3b), indicating that FDL does converge to a non-random low energy layout. Yet, its higher energy compared to the layout identified by NeuLay-2 results in FDL’s failure to identify the network’s inherent local community structure.
Discussion
The proposed NeuLay algorithm, a Graph Neural Network (GNN) developed to parameterize node features, significantly improves both the speed and the quality of graph layouts, opening up the possibility to quickly and reliably visualize large networks. It offers a fast and easy-to-use tool for large network visualization. We find that, many large networks have informative large-scale structures that remain hidden if the layout algorithms do not extract their main structural characteristics and find a way to display them. As we have shown here, NeuLay excels at this task, producing a high-quality layout, with distinct clusters and a clear internal structure. It achieves this performance by speeding up the dynamics of slow modes. Indeed, the leading eigenvectors of the adjacency matrix, or Principal Components (PC) in machine learning, are the “slow modes” in the dynamics of FDL33,34. NeuLay projects the graph layout to the top few PC (Fig. 2a) from the first iteration, separating the large communities which slows the dynamics, and catalyzing a faster convergence.
The mechanism applied by NeuLay to accelerate convergence is not restricted to graph layouts, but can be applied to any energy minimization problem on graphs, or graph dynamical processes expressed as gradient descent. Indeed, FDL is a special case of general reaction-diffusion problems on graphs, where in (1)LX is the “diffusion” and FNN ≡ −∂VNN/∂x are the nonlinear “reaction” terms. As our theoretical results do not depend on the exact form of VNN, they apply to any problem in the reaction-diffusion class, independent if the node features xi, are densities (e.g. of material flowing on the graph), or probabilities (e.g. susceptible, or infected nodes in epidemic spreading). Hence the method can improve the finding of endemic state in epidemics35, help with interventions and mitigation36, improve the modeling of cascading failures37, and help find optimal graph layout in chip design38, as well as accelerate models capturing opinion dynamics in social media6,39.
Currently, the efficiency of NeuLay is limited only by the computational complexity of GNN, which, while considerably faster than FDL, can still be expensive on exceptionally large graphs. We foresee further improvement by exploiting symmetries or hierarchical structures40 present in networks, leading to more efficient message-passing in GNN. These ideas could result in more advanced GNN architectures similar to GraphSage41 and ClusterGCN42, which make the graph sparser and thus reduces the computational complexity of GNN. It would be equally valuable the development of GNN or other AI-based tools to accelerate the layout of physical networks whose links are not straight, but curve to avoid overlaps43,44, capturing network layouts observed in the brain connectome or metamaterials.
Methods
Force Directed Layout (FDL)
Consider an undirected network with N nodes and \(A\in {{\mathbb{R}}}^{N\times N}\) adjacency matrix, where Aij is the weight of the link connecting node, i and j, and denote with X = (x1, . . . xN) the N × d matrix that captures the location xi of node i in a d-dimensional Euclidean space. FDL brings connected nodes close by minimizing the total energy, \({{{{{{{\mathcal{L}}}}}}}}\), that also plays the role of the “loss function” in machine learning9,10,11,12:
where L = D − A is the graph Laplacian and Dij = δij∑kAik is the degree matrix. For the repulsive potential VNN we choose a short-range Gaussian repulsion \({V}_{NN}(X)={a}_{N}{\sum }_{ij}\exp ({\left|{x}_{i}-{x}_{j}\right|}^{2}/4{r}_{0}^{2})\)43, but any rapidly falling repulsive potential would work. FDL performs gradient descent (GD) to minimize the total energy (eq.(1)). Note that in FDL the repulsive potential is usually chosen to be “long-range”, e.g. VNN = aN∑ijr0/∥xi − xj∥. This results in an all-to-all repulsive force with complexity O(N2). The Barnes-Hut algorithm16 can be used to reduce this to \(O(N\log N)\). Despite the widespread use of long-range repulsive forces for the layout of large and sparse graphs, short-range repulsive forces are lower complexity (O(N)). We note that, while FDL with short-range forces failed to yield a good layout for the Internet graph, FDL using long-range forces does yield a good layout. However, because long-range forces can become intractable for large graphs, we implement short-range forces43. Our experiments on the Internet graph show that NeuLay does not require long-range forces to find good layouts for large graphs.
Reparametrizing node positions with deep neural networks
To reparametrize X, we introduce two architectures: NodeMLP and NeuLay, described in Figure 1 and in Supplementary Information A. NodeMLP starts from an N × h dimensional random Z embedding of the nodes. It then projects to the target dimension d = 3 of the layout by defining node positions as X = σ(ZW + b), where σ is a nonlinear function such as tanh and θ = {Z, W, b} are the set of trainable parameters of the neural network. NeuLay uses GNN, that starts from an N × h random embedding Z and applies a Graph Convolutional Networks (GCN)14 layer to obtain G1 = σ(f(A)ZW(1)), with \(f(A)={\tilde{D}}^{-1/2}\tilde{A}{\tilde{D}}^{-1/2}\), where \(\tilde{A}=A+I\) and \({\tilde{D}}_{ii}={\sum }_{j}{\tilde{A}}_{ij}\) is the degree matrix of \(\tilde{A}\). Here G1 is N × h1 and is a new embedding of nodes in h1 dimensions that incorporates the graph structure via f(A). In the two-layer NeuLay-2, we apply another GCN with output G2 = σ(f(A)G1W(2)) and dimensions N × h2. Then, we concatenate the layer outputs G1 and G2 with Z along the embedding dimensions to obtain the (h + h1 + h2) dimensional node embedding G = [Z∣G1∣G2]. Finally, we project G down to d dimension as X = σ(GW + b). The set of trainable parameters of NeuLay-2 are θ = {Z, W(1), W(2), W, b}.
To obtain a layout we input X(θ) into the FDL algorithm and using the loss function (2). We perform energy minimization using gradient descent. Instead of optimizing X directly, we optimize the neural network parameters θ. Using the chain rule we can rewrite the GD equation (1) in terms of θa,
Unlike the familiar uses of deep neural networks, where training is done only once, here we retrain each θ for each layout.
The role of outliers in the eigenspectrum
To understand the mechanism that drive the faster convergence of NeuLay, we study the spectral expansions \(A={\sum }_{i}{\lambda }_{i}{\psi }_{i}{\psi }_{i}^{T}\) and \(L=D-A={\sum }_{i}{l}_{i}{\phi }_{i}{\phi }_{i}^{T}\). While in general the eigenvectors ψi of A and ϕi of L differ, in RGG and SBM all node degrees are close to an average degree \( < k > \) and we have \(L\approx \langle k \rangle I-A\), yielding ϕi ≈ ψi and \({l}_{i}\approx \langle k \rangle -{\lambda }_{i}\). Therefore, the elastic energy \({V}_{el}={{{{{{{\rm{Tr}}}}}}}}[{X}^{T}LX]/2\) in (2) dominates in the early stages of the optimization. Using (1) and \({{{{{{{\mathcal{L}}}}}}}}\approx {V}_{el}\) to examine the early stage evolution of the overlap of X with ψi for graphs where \(L\approx \langle k \rangle I-A\), finding
Hence, for FDL on lattices, RGG, and SBM, the mode \({\psi }_{i}^{T}X\) for each i evolves almost independently of other modes j ≠ i. Eq. (4) predicts that in early iterations the magnitude of the mode \({\psi }_{i}^{T}X\) drops exponentially with a rate \(-\varepsilon ( \langle k \rangle -{\lambda }_{i})\), and that modes with the largest λi drop at the slowest rate during GD (as L is positive semi-definite, \({\lambda }_{i}\le \langle k \rangle\)). Importantly, if the spectrum of A contains “outlier” eigenvalues {λo}, with λo ≫ meani[λi], the corresponding modes \({\psi }_{o}^{T}X\) evolve the slowest. Specifically, define the set of outlier eigenvalues as the indices \(Out=\{j|{\tilde{\lambda }}_{j} > {{{{{{{{\rm{mean}}}}}}}}}_{i}[{\tilde{\lambda }}_{i}]+{\sigma }_{\tilde{\lambda }}\}\), with \({\sigma }_{\tilde{\lambda }}\) being the standard deviation of the eigenvalues. The projection onto top eigenvectors is defined as \({A}_{top}={\sum }_{i\in Out}{\tilde{\lambda }}_{i}{\psi }_{i}{\psi }_{i}^{T}\) and the rest is Abulk = A − Atop.
Both NodeMLP and NeuLay start from a random node embedding Z. The difference is that NodeMLP performs X = σ(ZW + b), while NeuLay applies G1 = σ(f(A)ZW(1)). In NeuLay we choose \(f(A)={\tilde{D}}^{-1/2}\tilde{A}{\tilde{D}}^{-1/2}\), which for lattices, RGG and SBM, again has approximately the same eigenvectors as A. Using the spectrum \(f(A)={\sum }_{i}{\tilde{\lambda }}_{i}{\psi }_{i}{\psi }^{T}\) to expand Z = ∑iziψi, we find \(f(A)Z={\sum }_{i}{\tilde{\lambda }}_{i}{z}_{i}{\psi }_{i}\). In graphs with many outliers significantly larger than the bulk of the eigenvalues, the outliers dominate the spectral expansion and \(f(A)Z\approx {\sum }_{i\in {{{{{Out}}}}}}{\tilde{\lambda }}_{i}{z}_{i}{\psi }_{i}\). Hence, when performing GD \(dZ/dt=-\varepsilon {\partial }_{Z}{{{{{{{\mathcal{L}}}}}}}}\), in the presence of a GCN layer the gradients for outliers are magnified by \({\tilde{\lambda }}_{i}\), supporting prediction (i), that the more outliers eigenvalues the graph has, the higher the speedup. We used Out, the set of outlier eigenvalues to build Atop and Abulk to separate the relevance of the outlier eigenvalues in predictions (ii) and (iii). The details for how the outliers eigenvectors result in a faster drop in loss \(d{{{{{{{\mathcal{L}}}}}}}}/dt\) refer to Supplementary Information C.
Data availability
All data that support the plots within this paper and other findings of this study are available at https://github.com/csabath95/NeuLay.gitand the listed public sources: the word association graph (WAN)22: [http://w3.usf.edu/FreeAssociation/], the human protein-protein interaction network (PPI)23, the Facebook social network data24: http://snap.stanford.edu/index.html, the Internet at the autonomous system level25: http://www-personal.umich.edu/~mejn/netdata/, the flavor network26, boards of directors (public companies in Norway)27: https://networks.skewed.de/net/board_directors, US power grid29: http://www-personal.umich.edu/~mejn/netdata/, and the road network in Oakland30: http://snap.stanford.edu/index.html.
Code availability
Code is available for this paper at https://github.com/csabath95/NeuLay.git. All other codes that support the plots within this paper and other findings of this study are available from the corresponding author upon request.
References
Chinazzi, M. et al. The effect of travel restrictions on the spread of the 2019 novel coronavirus (COVID-19) outbreak. Science 368, 395–400 (2020).
Schlosser, F. et al. COVID-19 lockdown induces disease-mitigating structural changes in mobility networks. Proc. Nat Acad. Sci. USA 117, 32883–32890 (2020).
Cheng, F. et al. Network-based approach to prediction and population-based validation of in silico drug repurposing. Nat. Commun. 9, 1–12 (2018).
Wang, P., Hunter, T., Bayen, A. M., Schechtner, K. & González, M. C. Understanding road usage patterns in urban areas. Sci. Rep. 2, 1–6 (2012).
Shao, C. et al. The spread of low-credibility content by social bots. Nat. Commun. 9, 1–9 (2018).
Shah, D. & Zaman, T. Rumors in a network: Who’s the culprit? IEEE Trans. Inf. Theory 57, 5163–5181 (2011).
Bastian, M., Heymann, S. & Jacomy, M. Gephi: an open source software for exploring and manipulating networks. Proc. Int. AAAI Conf. Weblogs Soc. Media 3, 361–362 (2009).
Shannon, P. et al. Cytoscape: a software environment for integrated models of biomolecular interaction networks. Genome Res. 13, 2498–2504 (2003).
Kamada, T. et al. An algorithm for drawing general undirected graphs. Inf. Process. Lett. 31, 7–15 (1989).
Fruchterman, T. M. & Reingold, E. M. Graph drawing by force-directed placement. Softw. Pract. Exp. 21, 1129–1164 (1991).
Hu, Y. Efficient, high-quality force-directed graph drawing. Math. J.10, 37–71 (2005).
Quinn, N. & Breuer, M. A forced directed component placement procedure for printed circuit boards. IEEE Trans. Circuits Syst. 26, 377–388 (1979).
Jensen, F. Introduction to Computational Chemistry (John Wiley & Sons, 2017).
Kipf, T. N. & Welling, M. Semi-supervised classification with graph convolutional networks. arXiv. https://doi.org/10.48550/arXiv.1609.02907 (2016).
Gilmer, J., Schoenholz, S. S., Riley, P. F., Vinyals, O. & Dahl, G. E. Neural message passing for quantum chemistry. Proc. 34th Int. Conf. Mach. Learn. 70, 1263–1272 (2017).
Barnes, J. & Hut, P. A hierarchical \({{{{{{{\mathcal{O}}}}}}}}(NlogN)\) force-calculation algorithm. Nature 324, 446–449 (1986).
Veličković, P. et al. Graph attention networks. arXiv. https://doi.org/10.48550/arXiv.1710.10903 (2017).
Battaglia, P. W. et al. Relational inductive biases, deep learning, and graph networks. arXiv. https://doi.org/10.48550/arXiv.1806.01261 (2018).
Barabási, A.-L. & Albert, R. Emergence of scaling in random networks. Science 286, 509–512 (1999).
Holland, P. W., Laskey, K. B. & Leinhardt, S. Stochastic blockmodels: first steps. Soc. Networks 5, 109–137 (1983).
Penrose, M. et al. Random Geometric Graphs (Oxford University Press, 2003).
Nelson, D. L., McEvoy, C. L. & Schreiber, T. A. The University of South Florida free association, rhyme, and word fragment norms. Behav. Res. Methods, Instrum. Comput. 36, 402- 407 (2004).
Morselli, G. D. et al. Network medicine framework for identifying drug-repurposing opportunities for COVID-19. Proc. Natl Acad. Sci. USA 118, e2025581118 (2021).
Rozemberczki, B., Allen, C. & Sarkar, R. Multi-scale attributed node embedding arXiv. https://doi.org/10.48550/arXiv.1909.13021 (2019).
Karrer, B., Newman, M. E. & Zdeborová, L. Percolation on sparse networks. Phys. Rev. Lett. 113, 208702 (2014).
Ahn, Y.-Y., Ahnert, S. E., Bagrow, J. P. & Barabási, A.-L. Flavor network and the principles of food pairing. Sci. Rep. 1, 1–7 (2011).
Seierstad, C. & Opsahl, T. For the few not the many? The effects of affirmative action on presence, prominence, and social capital of women directors in Norway. Scand. J. Manag. 27, 44–54 (2011).
Gagnon, L. et al. Quantifying the microvascular origin of BOLD-fMRI from first principles with two-photon microscopy and an oxygen-sensitive nanoprobe. J. Neurosci. 35, 3663–3675 (2015).
Watts, D. J. & Strogatz, S. H. Collective dynamics of ‘small-world’ networks. Nature 393, 440–442 (1998).
Leskovec, J., Lang, K. J., Dasgupta, A. & Mahoney, M. W. Community structure in large networks: natural cluster sizes and the absence of large well-defined clusters. Internet Math. 6, 29–123 (2009).
Munzner, T. Exploring large graphs in 3D hyperbolic space. IEEE Comput. Graph. Appli. 18, 18–23 (1998).
Blondel, V. D., Guillaume, J.-L., Lambiotte, R. & Lefebvre, E. Fast unfolding of communities in large networks. J. Stat. Mech. Theory Exp. 2008, P10008 (2008).
Pan, R. K. & Sinha, S. Modularity produces small-world networks with dynamical time-scale separation. EPL 85, 68006 (2009).
Kolchinsky, A., Gates, A. J. & Rocha, L. M. Modularity and the spread of perturbations in complex dynamical systems. Phys. Rev. E 92, 060801 (2015).
Pastor-Satorras, R. & Vespignani, A. Epidemic spreading in scale-free networks. Phys. Rev. Lett. 86, 3200 (2001).
Balcan, D. et al. Multiscale mobility networks and the spatial spreading of infectious diseases. Proc. Natl Acad. Sci. USA 106, 21484–21489 (2009).
Buldyrev, S. V., Parshani, R., Paul, G., Stanley, H. E. & Havlin, S. Catastrophic cascade of failures in interdependent networks. Nature 464, 1025–1028 (2010).
Mirhoseini, A. et al. A graph placement methodology for fast chip design. Nature 594, 207–212 (2021).
Vosoughi, S., Mohsenvand, M. N. & Roy, D. Rumor gauge: Predicting the veracity of rumors on Twitter. ACM Trans. Knowl. Discov. Data 11, 1–36 (2017).
Zafeiris, A. & Vicsek, T. Why We Live in Hierarchies? A Quantitative Treatise (Springer, 2017).
Hamilton, W., Ying, Z. & Leskovec, J. Inductive representation learning on large graphs. Adva. Neural Inf. Process. Syst. https://doi.org/10.48550/arXiv.1706.02216 (2017).
Chiang, W.-L. et al. Cluster-gcn: An efficient algorithm for training deep and large graph convolutional networks. In Proc. 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining 257–266 (ACM, 2019).
Dehmamy, N., Milanlouei, S. & Barabási, A.-L. A structural transition in physical networks. Nature 563, 676–680 (2018).
Liu, Y., Dehmamy, N. & Barabási, A.-L. Isotopy and energy of physical networks. Nat. Phys. 17, 216–222 (2021).
Acknowledgements
We thank A. Grishchenko for help with the 3D visualizations. A.-L.B. and C.B. were supported by ERC grant No. 810115-DYNASET, The Eric and Wendy Schmidt Fund for Strategic Innovation G-22-63228, NSF SES-2219575, and John Templeton Foundation #62452. R.Y. acknowledges support in part by the U.S. Department Of Energy, Office of Science, U. S. Army Research Office under Grant W911NF-20-1-0334, Google Faculty Award, Amazon Research Award, and NSF Grants #2134274, #2107256 and #2134178.
Author information
Authors and Affiliations
Contributions
C.B. developed, ran, and analyzed the simulations, generated figures, and contributed to writing the manuscript. N.D. performed the mathematical modeling and derivations and contributed to writing the manuscript. R.Y. have guided designing the machine learning models and wrote the manuscript. A.-L.B. contributed to the conceptual design of the study and was the lead writer of the manuscript.
Corresponding author
Ethics declarations
Competing interests
A.-L.B. is the founder of Naring and Scipher Medicine, companies exploring the role of networks in health. C.B., N.D., and R.Y. declare no competing interest.
Peer review
Peer review information
Nature Communications thanks Jianxi Gao, Petar Veličković, and the other, anonymous, reviewer(s) for their contribution to the peer review of this work. Peer review reports are available.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Both, C., Dehmamy, N., Yu, R. et al. Accelerating network layouts using graph neural networks. Nat Commun 14, 1560 (2023). https://doi.org/10.1038/s41467-023-37189-2
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41467-023-37189-2
This article is cited by
-
Towards equitable AI in oncology
Nature Reviews Clinical Oncology (2024)
-
Grounding force-directed network layouts with latent space models
Journal of Computational Social Science (2023)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.