Abstract
Decycling and dismantling of complex networks are underlying many important applications in network science. Recently these two closely related problems were tackled by several heuristic algorithms, simple and considerably suboptimal, on the one hand, and involved and accurate messagepassing ones that evaluate singlenode marginal probabilities, on the other hand. In this paper we propose a simple and extremely fast algorithm, CoreHD, which recursively removes nodes of the highest degree from the 2core of the network. CoreHD performs much better than all existing simple algorithms. When applied on realworld networks, it achieves equally good solutions as those obtained by the stateofart iterative messagepassing algorithms at greatly reduced computational cost, suggesting that CoreHD should be the algorithm of choice for many practical purposes.
Introduction
In decycling of a network we aim to remove as few nodes as possible such that after the removal the remaining network contains no loop. In network dismantling we aim to find the smallest set of nodes such that after their removal the network is broken into connected components of subextensive size. These are two fundamental networkoptimization problems with a wide range of applications, related to optimal vaccination and surveillance, information spreading, viral marketing, and identification of influential nodes. Considerable research efforts have been devoted to the network decycling and dismantling problems recently^{1,2,3,4,5,6,7,8}.
Both the decycling and the dismantling problems belong to the class of NPhard problems^{6,9}, meaning that it is rather hopeless to look for algorithms to solve them exactly in polynomial time. However, finding the best possible approximate solutions for as large classes of networks as possible is an open and actively investigated direction. Recent theoretic and algorithmic progress on both these problems^{1,2,3,5,6} came from the fact that, on random sparse networks with degree distributions having a finite second moment, methods from physics of spin glasses provide accurate algorithms for both decycling and dismantling. These sparse random networks are locally treelike and do not contain many short loops. On such networks the decycling is closely linked to dismantling and asymptotically almost the same set of nodes is needed to achieve both^{5,6,10}. Even on realworld networks that typically contain many small loops, best dismantling is currently achieved by first finding a decycling solution and then reinserting nodes that close short loops but do not increase too much the size of the largest component^{5,6}.
Both the algorithms of refs 5 and 6 achieve performance that is extremely close to the theoretically optimal values computed on random networks. However, both these algorithms are global, they need to iterate certain equations on the whole network in order to select the suitable candidate nodes. Although they are both scalable and can be run on networks with many millions of nodes, they are not completely straightforward to understand and require some experience with spin glass theory. The closetooptimal performance of these algorithms is theoretically justified only on random networks. Despite their good performance observed empirically on networks with many loops, there might still exist even better and analyzable strategies for realworld networks.
As usual in applied science, in many potential applications we are at first not even sure that optimal dismantling or optimal decycling is the best strategy to answer the question in hand (e.g., the problem of social influence maximization^{11,12,13,14}). Therefore it is extremely important to have a really very simple and fast decycling and dismantling strategy that can provide an accurate assessment of whether this approach is at all interesting for a given practical problem. However, existing simple strategies, such as removing adaptively high degree nodes^{15,16}, are very far from optimal performance and therefore not very useful. Recently the authors of ref. 4 claimed that a heuristics based on the socalled collective influence (CI) measure can be a perfect candidate for this purpose. This algorithm has attracted a lot of enthusiasm in the network science community. However, more systematic investigations performed in refs 5, 6 and 8 revealed that the CI algorithm is still considerably far from being optimal. The CI algorithm is also not particularly competitive in terms of computational time because a large neighborhood of a node needs to be considered in order to evaluate the CI measure.
In the present paper we introduce the CoreHD algorithm that is basically as simple and fast as the adaptive removal of high degree nodes, yet its performance is much closer to optimal than the CI algorithm or its extended versions, and comparably close as the best known messagepassing methods^{5,6} while several orders of magnitude faster. It hence provides simple and tractable solutions for networks with many billions of nodes. The method is simply based on adaptive removal of highestdegree nodes from the 2core of the network. Apart of its simplicity and speed the performance of the CoreHD algorithm is basically indistinguishable from the performance of the messagepassing algorithms on random graphs with longtailed degree distributions. On all realworld network instances we tested the result by CoreHD is within few nodes from the best one found by messagepassing and on some instances we found that it is even slightly better. On top of all that, the simple structure of CoreHD might be amenable to rigorous analysis providing guarantees for loopy networks that are not accessible for the messagepassing methods.
For all the above reasons we argue that in many applications of decycling and dismantling CoreHD should be the first choice. The simple algorithmic idea generalizes easily to the problem of destroying optimally the kcore of a network  one focuses on the current kcore and adaptively removes highest degree nodes.
Results
In this section we evaluate the CoreHD algorithm for both random and realworld networks, by comparing the minimum fraction of nodes we need to remove in order to break the network into a forest or components with size smaller than 0.01 N. We compare to the Belief Propagation guided Decimation (BPD)^{5} and Collective Influence method (CI)^{4} (CI_{4} results are obtained using the original code of ref. 4).
First, we notice that on some simple examples, e.g. regular random graphs with degree 3, the CoreHD algorithm reaches the exact optimal decycling fraction ρ = 0.25. This matches the performance of a greedy method of ref. 17 that for this particular case is provably optimal.
In Fig. 1 we compare the performance of the above algorithms on an ErdösRényi random network with N = 50000 nodes and average degree c = 3.5. In the left panel we plot the fraction of nodes in the largest connected component (LCC, denoted q) as a function of the fraction of removed nodes, denoted ρ. We see that compared to HD and CI the CoreHD algorithm works the best by a large margin, breaking the network into small component with size smaller than 0.01 N after removing fraction of only 0.1846 of nodes. While CI and HD need to remove fraction 0.2014, and 0.2225 of nodes respectively. This is compared to the closetooptimal performance of the iterative message passing BPD that needs to remove fraction 0.1780 of nodes, and to the theoretical prediction for the asymptotically optimal value 0.1753^{1,2,3,6,7}.
We also see from the figure that the fraction of nodes in the LCC obtained by CoreHD encounters a first order transition at ρ_{dec} = 0.1831, this is because at this point (just at the beginning of the discontinuity) the remaining network becomes a forest. The greedy treebreaking procedure then quickly breaks the forest into small components. While the other algorithms do not have this phenomenon, the size of the LCC goes to zero continuously. In the right panel of Fig. 1 we plot the fraction q of nodes in the 2core as a function of ρ. We can see that for CoreHD, q reaches zero at ρ = 0.1831 indicating that the remaining network contains no loop, thus is a forest. While for other algorithms the 2core remains extensive until the network is dismantled. On a larger ER random network with N = 10^{6}, c = 3.5, the difference between the sizes of decycling and the dismantling sets the CoreHD algorithm finds is not distinguishable within the precision of 4 valid digits and is 0.1830 for both. Note that this result is (slightly) better than yet another approach suggested recently in the literature^{8} that achieves 0.1838 with an algorithm still considerably more involved than CoreHD.
Besides performing much better than CI, the CoreHD is also much faster: the 2core of the network can be computed efficiently using a leafremoval process with O(N) operations. After deleting a node, one only needs to update the 2core, which requires on average O(1) operations in sparse networks, and is clearly much faster than updating the CI score. Actually, in sparse networks when the size of the 2core is much smaller than the size of the network, CoreHD is even faster than the HD algorithm which removes one by one nodes from the whole network.
The computational times for the CoreHG, CI and BPD algorithms as the system size grows are shown in Fig. 2 for ER network with mean degree c = 3. The BPD algorithm performs slightly better than the CoreHD algorithm but it is much slower. For example, for an ER network with c = 3 and N = 2 × 10^{8}, the solution obtained by CoreHD has relative dismantling/decycling set size ρ ≈ 0.1407 (computing time is 64 minutes), which is only slightly larger than the value of ρ ≈ 0.1357 obtained by BPD (computing time is 23.5 hours^{5}). We note that in these experiments, in each step of removal, BPD and CI_{4} remove 0.1% of nodes (e.g., 10000 nodes for N = 10^{7}), while CoreHD removes only 1 node per step. Even this way the computational time of CoreHD is shorter than the time used for reading the network from the data file (edgelist format). We note that in the current implementation we use ordering of lists of N elements. This could be improved further, but we saw little point in doing so since the bottleneck was loading the graph, rather than the algorithm for decycling and dismantling.
Figure 3 presents results for ErdösRényi random graphs, regular random graphs, and scalefree random networks of varying average degree. In all cases CoreHD works better than CI and worse than BPD, with the best performance obtained for scalefree networks. The good performance of CoreHD for the scalefree networks is of particular interest because almost all realworld networks have a heavytailed degreedistribution.
A set of experiments on realworld networks is presented in Table 1. We list the fraction of nodes we need to remove in order to remove all cycles, and in order to break the network into small components with size smaller than 0.01 N. For dismantling, in addition to Algorithm 6 we do a refinement by inserting back some deleted nodes that do not increase the largest component size beyond the 0.01 N. We can see that CoreHD works excellently for realworld network instances, giving decycling and dismantling sets very close to the stateofart BPD and much smaller than CI. It is also surprising to see that in some networks e.g. RoadEU, IntNet1 and RoadTX, CoreHD even outperforms BPD slightly. Table 1 clearly demonstrates the time superiority of CoreHD for realworld networks as compared with both CI and BPD.
Note that in recent work^{6} authors stressed that in random graphs there are many dismantling sets of size close to the optimum, and a given node can be included in some and not included in others. Such an observation is quite common to many optimization problems on networks where typically the ground state of a given optimization problem is degenerate and different lowenergy configurations may disagree on a considerable number of nodes. The CoreHD algorithm is also able to find different dismantling sets since nodes to be removed are chosen among the typically numerous ones in the core having the largest degree, by running the algorithm with different random number initialization for the ER network of average degree 3.5 we typically observed that two different dismantling sets obtained by the CoreHD algorithm agree in 74% of nodes. From this observation we conclude, in agreement with^{6}, that the concept of dismantling set is a result of highly correlated choice and not addition of a set of particularly important nodes (sometimes refer to as influential nodes or superspreaders^{4}.
Discussion
We have presented that iteratively removing nodes having the highest degree from the 2core of a network gives an ultrafast while very efficient algorithm for decycling and dismantling of networks. Our algorithm is so fast that its running time is shorter than the time of reading the network file.
It is still surprising to us that such a simple algorithm could work much better than more sophisticated algorithms: We have tried running CI (see SI), adjacency matrix centrality on the 2core of the network, and HD on 3core of the network, they are all slower but perform no better than CoreHD. Our experiments also show that CoreHD outperforms centrality measures using left and right eigenvector of the nonbacktracking matrix^{18}, an idea that originally inspired us to propose the CoreHD algorithm. More detailed understanding of why this is the best performing strategy is let for future work.
On the realworld networks which typically have many short loops and motifs, decycling is quite different from dismantling. A natural idea to generalize our CoreHD would be consider a factor graph treating short loops and motifs as factors, then do CoreHD on the 2core of the factor graph.
Finally, CoreHD can be generalized naturally to removal of the kcore, again running the adaptive HD heuristics on the kcore or the current graph. Comparison of this strategy to existing algorithms^{2,19} is in progress.
Methods
We now describe CoreHD as an extremely fast algorithm for decycling and dismantling of huge complex networks with closetooptimal outcomes. Let us begin with some motivating discussions.
Perhaps the simplest algorithms one can propose for decycling and dismantling is adaptive removal of highestdegree nodes. We call this method HD, it is indeed extremely fast, but empirically does not perform very well. One reason why HD does not work well is that some nodes of large degree, such as node i in Fig. 4, do not belong to any loop, and hence do not have to be removed for decycling. Due to the property that trees can always be dismantled by a vanishing fraction of nodes^{10}, nodes such as i of Fig. 4 are also not useful for dismantling. Note that the CI method of ref. 4 shares this problem, see the SI for the argument on why our algorithm outperforms CI.
The above observation motivates a very natural idea that dismantling and decycling algorithms should always focus only on the 2core of the network. The 2core is a subnetwork that is obtained after adaptive removal of all leaves (nodes with only a single attached edge). The simplest and fastest strategy is then to remove the highestdegree nodes from the remaining 2core. To our surprise this simple idea provides much better performance than other comparably simple approaches existing in the literature. We call the resulting algorithm CoreHD, it is detailed in Algorithm 1.
For the decycling problem, CoreHD simply removes highestdegrees nodes from the 2core in an adaptive way (updating node degree as the 2core shrinks), until the remaining network becomes a forest. For dismantling, after decycling, CoreHD also breaks the trees into small components, see SI that follows treebreaking strategy from refs 5 and 6. In case the original network has many small loops, a refined dismantling set is obtained after a reinsertion of nodes that do not increase (much) the size of the largest component, again as proposed recently in refs 5 and 6. For details on implementation of the reinsertion algorithm we refer to the SI.
Additional Information
How to cite this article: Zdeborová, L. et al. Fast and simple decycling and dismantling of networks. Sci. Rep. 6, 37954; doi: 10.1038/srep37954 (2016).
Publisher's note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
 1.
Zhou, H.J. Spin glass approach to the feedback vertex set problem. Eur. Phys. J. B 86, 455 (2013).
 2.
Altarelli, F., Braunstein, A., Dall’Asta, L. & Zecchina, R. Optimizing spread dynamics on graphs by message passing. Journal of Statistical Mechanics: Theory and Experiment 2013, P09011 (2013).
 3.
Guggiola, A. & Semerjian, G. Minimal contagious sets in random regular graphs. Journal of Statistical Physics 158, 300–358 (2015).
 4.
Morone, F. & Makse, H. A. Influence maximization in complex networks through optimal percolation. Nature 524, 65–68 (2015).
 5.
Mugisha, S. & Zhou, H.J. Identifying optimal targets of network attack by belief propagation. Phys. Rev. E 94, 012305 (2016).
 6.
Braunstein, A., Dall’Asta, L., Semerjian, G. & Zdeborová, L. Network dismantling. Proceedings of the National Academy of Sciences 113, 12368–12373 (2016).
 7.
Qin, S.M., Zeng, Y. & Zhou, H.J. Spin glass phase transitions in the random feedback vertex set problem. arXiv preprint arXiv:1603.09032 (2016).
 8.
Clusella, P., Grassberger, P., PérezReche, F. J. & Politi, A. Immunization and targeted destruction of networks using explosive percolation. arXiv:1604.00073 (2016).
 9.
Karp, R. M. Reducibility among combinatorial problems. In Complexity of computer computations, 85–103 (Springer, 1972).
 10.
Janson, S. & Thomason, A. Dismantling sparse random graphs. Combinatorics, Probability and Computing 17, 259–264 (2008).
 11.
Richardson, M. & Domingos, P. Mining knowledgesharing sites for viral marketing. In Proceedings of 8th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 61–70 (ACM, New York, NY, 2002).
 12.
Kempe, D., Kleinberg, J. & Tardos, E. Maximizing the spread of influence through a social network. Theory of Computing 11, 105–147 (2015).
 13.
Jung, K., Heo, W. & Chen, W. Irie: Scalable and robust influence maximization in social networks. In 2012 IEEE 12th International Conference on Data Mining, 918–923 (IEEE, 2012).
 14.
Borgs, C., Brautbar, M., Chayes, J. & Lucier, B. Maximizing social influence in nearly optimal time. In Proceedings of the TwentyFifth Annual ACMSIAM Symposium on Discrete Algorithms, 946–957 (Society for Industrial and Applied Mathematics, 2014).
 15.
Albert, R., Jeong, H. & Barabási, A.L. Error and attack tolerance of complex networks. Nature 406, 378–382 (2000).
 16.
Cohen, R., Erez, K., benAvraham, D. & Havlin, S. Breakdown of the internet under intentional attack. Physical review letters 86, 3682 (2001).
 17.
Bau, S., Wormald, N. C. & Zhou, S. Decycling numbers of random regular graphs. Random Structures & Algorithms 21, 397–413 (2002).
 18.
Zhang, P. Nonbacktracking operator for the ising model and its applications in systems with multiple states. Phys. Rev. E 91, 042120 (2015).
 19.
Pei, S., Teng, X., Shaman, J., Morone, F. & Makse, H. A. Collective influence maximization in threshold models of information cascading with firstorder transitions. arXiv preprint arXiv:1606.02739 (2016).
 20.
Goh, K.I., Kahng, B. & Kim, D. Universal behavior of load distribution in scalefree networks. Phys. Rev. Lett. 87, 278701 (2001).
 21.
Šubelj, L. & Bajec, M. Robust network community detection using balanced propagation. Eur. Phys. J. B 81, 353–362 (2011).
 22.
Bu, D. et al. Topological structure analysis of the proteinprotein interaction network in budding yeast. Nucleic Acids Res. 31, 2443–2450 (2003).
 23.
Watts, D. J. & Strogatz, S. H. Collective dynamics of ‘smallworld’ netowrks. Nature 393, 440–442 (1998).
 24.
Leskovec, J., Kleinberg, J. & Faloutsos, C. Graphs over time: densification laws, shrinking diameters and possible explanations. In Proceedings of the eleventh ACM SIGKDD international conference on Knowledge discovery in data mining, 177–187 (ACM, New York, 2005).
 25.
Leskovec, J., Kleinberg, J. & Faloutsos, C. Graph evolution: Densification and shrinking diameters. ACM Transactions on Knowledge Discovery from Data 1, 2 (2007).
 26.
Ripeanu, M., Foster, I. & Iamnitchi, A. Mapping the gnutella network: Properties of largescale peertopeer systems and implications for system design. IEEE Internet Comput. 6, 50–57 (2002).
 27.
Cho, E., Myers, S. A. & Leskovec, J. Friendship and mobility: User movement in localationbased social networks. In ACM SIGKDD International Conference o Knowledge Discovery and Data Mining, 1082–1090 (San Diego, CA, USA, 2011).
 28.
Leskovec, J., Lang, K. J., Dasgupta, A. & Mahoney, M. W. Community structure in large networks: Natural cluster sizes and the absence of large welldefined clusters. Internet Math. 6, 29–123 (2009).
Acknowledgements
H.J.Z was supported by the National Basic Research Program of China (grant number 2013CB932804), the National Natural Science Foundation of China (grant numbers 11121403 and 11225526), and the Knowledge Innovation Program of Chinese Academy of Sciences (No. KJCX2EWJ02).
Author information
Author notes
 Lenka Zdeborová
 , Pan Zhang
 & HaiJun Zhou
These authors contributed equally to this work.
Affiliations
Institut de Physique Thérique, CNRS, CEA and Université ParisSaclay, GifsurYvette, France
 Lenka Zdeborová
CAS Key Laboratory of Theoretical Physics, Institute of Theoretical Physics, Chinese Academy of Sciences, Beijing 100190, China
 Pan Zhang
 & HaiJun Zhou
Authors
Search for Lenka Zdeborová in:
Search for Pan Zhang in:
Search for HaiJun Zhou in:
Contributions
L.Z., P.Z., and H.Z. conducted the experiment(s), analyzed the results and reviewed the manuscript.
Competing interests
The authors declare no competing financial interests.
Corresponding author
Correspondence to Pan Zhang.
Supplementary information
PDF files
Rights and permissions
This work is licensed under a Creative Commons Attribution 4.0 International License. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in the credit line; if the material is not included under the Creative Commons license, users will need to obtain permission from the license holder to reproduce the material. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/
About this article
Further reading

1.
Nodal vulnerability to targeted attacks in power grids
Applied Network Science (2018)

2.
Optimal percolation on multiplex networks
Nature Communications (2017)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.