Introduction

Interconnected complex networks are ubiquitous in todays world. They control infrastructures of modern society (energy-communication-transportation), the financial system or even the human body1,2,3. Unfortunately, they are much more fragile than uncoupled networks as recently recognized through the finding that the robustness changes from a second order transition in uncoupled systems to first order in interdependent systems4,5,6. The obvious mitigation strategy consists in partially decoupling the networks by the creation of autonomous nodes7. Too much disconnection however risks endangering the functionality of the system. The question which we will address here is how to reduce fragility without losing functionality and we will in fact answer this question by developing an explicit algorithm based on betweenness that enables to avoid the abrupt collapse with a minimum number of autonomous nodes.

Buldyrev et al.4 proposed a percolation framework to study two coupled networks, A and B, where each A-node is coupled to a B-node, via bi-directional links, such that when one fails the other cannot function either. The removal of a fraction of A-nodes may trigger a domino effect where, not only their counterparts in B fail, but all nodes that become disconnected from the giant cluster of both networks also fail. This causes further cascading of failures, yielding an abrupt collapse of connectivity, characterized by a discontinuous (first order) percolation transition. Parshani et al.7 showed that damage can be mitigated by decreasing the degree of coupling, but only if a significant fraction (≈ 0.4) of nodes is decoupled, the transition changes from discontinuous to continuous. The coupling is reduced by randomly selecting a fraction of nodes to become autonomous and, therefore, independent on the other network. For the coupling between power stations and communication servers, for example, autonomous power stations have alternative communication systems which are used when the server fails and an autonomous server has its own energy power supply. We propose a method, based on degree and centrality, to identify these autonomous nodes that maximize the robustness of the system in terms of connectivity. We show that, with this scheme, the critical coupling increases, i.e., the fraction of nodes that needs to be decoupled to smoothen out the transition is much smaller (close to 0.1 compared to 0.4). Significant improvement is observed for different coupled networks including for Erdős-Rényi graphs (ER) where such improvement in the robustness was unexpected given their narrow degree distribution. To demonstrate the strength of our approach, in Fig. 1 we apply the proposed strategy to the real coupled system in Italy2 and show that by only protecting four servers the robustness is significantly improved (details in the figure caption).

Figure 1
figure 1

The herein proposed strategy improves significantly the resilience of the coupling between the communication system (39 stars) and the power grid (310 circles) in Italy.

The color scheme stands for the probability that the node is inactive after the random failure of 14 communication servers. In a) all communication servers are coupled while in b) four servers have been decoupled following the strategy proposed here. The coupling between the networks was established based on the geographical location of the nodes, such that each communication server is coupled with the closest power station2. The images were produced using the software Pajek.

We consider a pair of networks, A and B, where a fraction q (degree of coupling) of A-nodes are coupled with B-nodes. To be functional, nodes need to be connected to the giant cluster of their network. When an A-node fails, the corresponding one in B cannot function either. Consequently, all nodes bridged to the largest cluster through these nodes, together with their counterpart in the other network, become also deactivated. A cascade of failures occurs with drastic effects on the global connectivity (see Fig. 2)4,7. This process can also be treated as an epidemic spreading8. To study the resilience to failures, we follow the size of the largest connected cluster of active A-nodes, under a sequence of random irreversible attacks to network A. Notwithstanding the simplicity of solely considering random attacks, this model can be straightforwardly extended to targeted ones9. Recently, for single networks, it has been proposed10 to quantify the robustness R as

where Q is the number of node failures, S(Q) the size of the largest connected cluster in a network after Q failures and N is the total number of nodes in the network10,11. Here we extend this definition to coupled systems by performing the same measurement, given by Eq. (1), only on the network where the random failures occur, namely, network A. To follow the cascade triggered by the failure of a fraction 1 – p of A-nodes, similar to7, we solve the iterative equations,

with the initial condition α1 = p, where αn and βn are the fraction of A and B surviving nodes at iteration step n and Sxn) is the fraction of such nodes in the giant cluster. qχ,n is the fraction of dependent nodes in network χ fragmented from the largest cluster (see Methods for further details).

Figure 2
figure 2

Scheme of the cascade of node failures triggered by the initial failure of a node in network A (top network).

Two networks, A (top) and B (bottom), are considered. When a node initially fails in network A (a) all nodes connected to the largest component through it also fail (b) as well as the corresponding dependent nodes in network B (c). The failure of the dependent nodes in network B leads to further failures in both networks (d) and (e). For each iteration step, the degree of coupling qx and the size of the largest connected component Sx for each network x are listed in (f).

Results

To demonstrate our method of selecting autonomous nodes we consider two ER graphs with average degree 〈k〉 = 4 and 10% of autonomous nodes (q = 0.1). First we consider a method based on the degree of the node and later we compare with the method based on the betweenness. Under a sequence of random failures, the networks are catastrophically fragmented when close to 45% of the nodes fail, as seen in Fig. 3. For a single ER, with the same average degree, the global connectivity is only lost after the failure of 75% of the nodes. Figure 3 also shows ((green-)dotted-dashed line) the results for choosing as autonomous nodes in both networks the fraction 1 – q of the nodes with the highest degree and coupling the remaining ones at random. With this strategy, the robustness R can be improved and the corresponding increase of pc is about 40%, from close to 0.45 to close to 0.65. Also the order of the transition changes from first to second order. Further improvement can be achieved if additionally the coupled nodes are paired according to their position in the ranking of degree, since interconnecting similar nodes increases the global robustness12,13. In the inset of Fig. 3 we see the dependence on q of the relative robustness for the degree strategy compared to the random case R/Rrandom. For the entire range of q the proposed strategy is more efficient and a relative improvement of more than 15% is observed when still 85% of the nodes are coupled.

Figure 3
figure 3

Fraction of A-nodes in the largest connected cluster, s, as a function of the fraction of randomly removed nodes 1 – p from network A, for two coupled ER (average degree 〈k〉 = 4) with 90% of the nodes connected by inter-network links (q = 0.9).

It is seen that robustness can significantly be improved by properly selecting the autonomous nodes. We start with two fully interconnected ER and decouple 10% of their nodes according to three strategies: randomly ((black-)solid line), the ones with highest degree in network A ((red-)dotted line) and in network B ((blue-) dashed line). We also include the case where 10% autonomous nodes in both networks are chosen as the ones with highest degree and all the others are interconnected randomly ((green-)dotted-dashed line). The inset shows the dependence of the relative robustness of the degree strategy on the degree of coupling q compared with the random case. Results for the degree have been obtained with the formalism of generation functions (see Methods).

Two types of technological challenges are at hand: either a system has to be designed robust from scratch or it already exists, constrained to a certain topology, but requires improved robustness. In the former case, the best procedure is to choose as autonomous the nodes with highest degree in each network and couple the others based on their rank of degree. For the latter, rewiring is usually a time-consuming and expensive process and the creation of new autonomous nodes may be economically more feasible. The simplest procedure consists in choosing as autonomous both nodes connected by the same internetwork link. However, a high degree node in network A is not necessarily connected with a high degree node in network B. In Fig. 3 we compare between choosing the autonomous pairs based on the degree of the node in network A or in network B. When pairs of nodes are picked based on their rank in the network under the initial failure (network A), the robustness almost does not improve compared to choosing randomly. If, on the other hand, network B is considered, the robustness is significantly improved, revealing that this scheme is more efficient. This asymmetry between A and B network is due to the fact that we attack only nodes in network A, triggering the cascade, that initially shuts down the corresponding B-node. The degree of this B-node is related to the number of nodes which become disconnected from the main cluster and consequently affect back the network A. Therefore, the control of the degree of vulnerable B-nodes is a key mechanism to downsize the cascade. On the other hand, when a hub is protected in network A it can still be attacked since the initial attack does not distinguish between autonomous and non-autonomous nodes.

In Fig. 4(a) we compare four different criteria to select the autonomous nodes: betweenness, degree, k-shell and random choice, for two coupled ER networks. In the betweenness strategy, the selected autonomous are the ones with highest betweenness. The betweenness is defined as the number of shortest paths between all pairs of nodes passing through the node14. A k-shell is obtained by removing, iteratively, all nodes with degree smaller than k, until all remaining nodes have degree k or larger. In the k-shell strategy, the autonomous are chosen as the ones with highest k-shell in the k-shell decomposition15. The coupled nodes (not autonomous), for all cases, have been randomly inter-linked. Since ER networks are characterized by a small number of k-shells, this strategy is even less efficient than the random strategy for some values of q, while the improved robustness for degree and betweenness strategies is evident compared with the random selection. While in the random case, for , a significant decrease of the robustness with q is observed, in the degree and betweenness cases, the change is smoother and only significantly drops for higher values of q. A maximum in the ratio R/Rrandom occurs for q ≈ 0.85, where the relative improvement is above 12%. Since, in random networks, many metrics are strongly correlated14, the results for betweenness and degree are similar.

Figure 4
figure 4

Dependence of the robustness, R, on the degree of coupling, q, for two, interconnected, (a) ER (average degree 〈k〉 = 4) and (b) SF with degree exponent γ = 2.5.

Applying our proposed strategy is applied, the optimal fraction of autonomous nodes is relatively very small. Autonomous nodes are chosen in four different ways: randomly ((blue-) triangles), high degree ((black-)dots), high betweenness ((red-)stars) and high k-shell ((yellow-)rhombi). The insets show the relative improvement of the robustness, for the different strategies of autonomous selection compared with the random case. Results have been averaged over 102 configurations of two networks with 103 nodes each. For each configuration we averaged over 103 sequences of random attacks.

Many real-world systems are characterized by a degree distribution which is scale free with a degree exponent γ16,17. In Fig. 4(b) we plot R as a function of q for two coupled scale-free networks (SF) with 103 nodes each and γ = 2.5. Similar to the two coupled ER, this system is also significantly more resilient when the autonomous nodes are selected according to the highest degree or betweenness. For values of the robustness is similar to that of a single network (q = 0) since the most relevant nodes are decoupled. A peak in the relative robustness, R/Rrandom (see inset of Fig. 4b), occurs for q ≈ 0.95 where the improvement, compared to the random case, is almost 30%. Betweenness, degree and k-shell, have similar impact on the robustness since these three properties are strongly correlated for SF. From Fig. 4, we see that, for both SF and ER, the robustness is significantly improved by decoupling, based on the betweenness, less than 15% of the nodes. Studying the dependence of the robustness on the average degree of the nodes we conclude that for average degree larger than five, even 5% autonomous nodes are enough to achieve more than 50% of the maximum possible improvement.

For the cases discussed in Fig. 4, results obtained by selecting autonomous nodes based on the highest degree do not significantly differ from the ones based on the highest betweenness. This is due to the well known finding that for Erdős-Rényi and scale-free networks, the degree of a node is strongly correlated with its betweenness14. However, many real networks are modular, i.e., composed of several different modules interconnected by less links and then nodes with higher betweenness are not, necessarily, the ones with the largest degree18. Modularity can be found, for example, in metabolic systems, neural networks, social networks, or infrastructures19,20,21,22. In Fig. 5 we plot the robustness for two coupled modular networks. Each modular network was generated from a set of four Erdős-Rényi networks, of 500 nodes each and average degree five, where an additional link was randomly included between each pair of modules. For a modular network, the nodes with higher betweenness are not necessarily the high-degree nodes but the ones bridging the different modules. Figure 5 shows that the strategy based on the betweenness emerges as better compared to the high degree method.

Figure 5
figure 5

Dependence of the robustness, R, on the degree of coupling, q, for two, randomly interconnected modular networks with 2·103 nodes each.

The modular networks were obtained from four Erdős-Rényi networks, with 500 nodes each and average degree five, by randomly connecting each pair of modules with an additional link. Autonomous nodes are selected in three different ways: randomly (blue triangles), higher degree (black dots) and higher betweenness (red stars). In the inset we see the relative enhancement of the robustness, for the second and third schemes of autonomous selection compared with the random case. Results have been averaged over 102 configurations and 103 sequences of random attacks to each one.

Another example that shows that betweenness is superior to degree is when we study coupled random regular graphs. In random regular graphs all nodes have the same degree and are connected randomly. Figure 6 shows the dependence of the robustness on the degree of coupling, for two interconnected random regular graphs with degree 4. The autonomous nodes are selected randomly (since all degrees are the same) or following the betweenness strategy. Though all nodes have the same degree and the betweenness distribution is narrow, selecting autonomous nodes based on the betweenness is always more efficient than the random selection. Thus, the above two examples suggest that betweenness is a superior method to chose the autonomous nodes compared to degree.

Figure 6
figure 6

Dependence of the robustness, R, on the degree of coupling, q, for two, randomly interconnected random regular graphs with 8 · 103 nodes each, all with degree four.

Autonomous nodes are selected in two different ways: randomly (blue triangles) and higher betweenness (red stars). In the inset the relative enhancement of the robustness is shown for the betweenness compared to the random case. Results have been averaged over 102 configurations and 103 sequences of random attacks to each one.

The vulnerability is strongly related to the degree of coupling q. Parshani et al.7 have analytically and numerically shown that, for random coupling, at a critical coupling q = qt, the transition changes from continuous (for q < qt) to discontinuous (for q > qt). In Fig. 7 we see the two-parameter diagram (pc vs q) with the tricritical point and the transition lines (continuous and discontinuous) for the random (inset) and the degree (main) strategies. As seen in Fig. 7, when autonomous nodes are randomly selected, about 40% autonomous nodes are required to soften the transition and avoid catastrophic cascades, while following the strategy proposed here only a relatively small amount (q > 0.9) of autonomous nodes are needed to avoid a discontinuous collapse. Above the tricritical point, the jump increases with the degree of coupling, lending arguments to the paramount necessity of an efficient strategy for autonomous selection, given that the fraction of nodes which can be decoupled is typically limited. The dependence of qt on the average degree 〈k〉 is shown in Fig. 8. The ratio between the tricritical coupling for degree and random strategies increases with decreasing 〈k〉. For example, for 〈k〉 ≈ 2 the fraction of autonomous nodes needed to soften the transition with the random selection is six times the one for the degree strategy.

Figure 7
figure 7

Two-parameter diagram (blue curves) of two coupled ER (average degree 〈k〉 = 4) under random attack.

The horizontal axis is the degree of coupling q and the vertical one is p so that 1 – p is the fraction of initially removed nodes. The size of the jump in the fraction of A-nodes in the largest connected cluster is also included (red-dotted-dashed curve). The dashed curve stands for a discontinuous transition while the solid one is a critical line (continuous transition). The two lines meet at a tricritical point (TP). Autonomous nodes are selected based on the degree (main plot) and randomly (inset). Results have been obtained with the formalism of generating functions.

Figure 8
figure 8

Tricritical coupling qt dependence on the average degree 〈k〉 for two coupled ER, showing that the fraction of autonomous nodes to smoothen out the transition is significantly reduced with the proposed strategy when compared with the random case.

Autonomous nodes are selected following two different strategies: randomly (red squares) and high degree (black circles).

As in Ref. 23, following the theory of Riedel and Wegner24,25,26, we can characterize the tricritical point. Two relevant scaling fields are defined: one tangent (μp) and the other perpendicular (μq) to the critical curve at the tricritical point. In these coordinate axes the continuous line is described by , where the tricritical crossover exponent φt = 1.00 ± 0.05 for degree and random strategies. The tricritical order parameter exponent, βt, can be evaluated from,

giving βt = 0.5 ± 0.1 for both strategies. Since these two exponents are strategy independent (see Fig. 9), we conjecture that the tricritical point for degree and random selection are in the same universality class.

Figure 9
figure 9

Dependence of the fraction of A-nodes in the largest connected cluster on the scaling field μp along the direction perpendicular to the transition line at the tricritical point.

The slope is the tricritical exponent βt related with the order parameter. Autonomous nodes in the two coupled ER (〈k〉 = 4) have been selected randomly (red line) and following the ranking of degree (black line).

Discussion

Here, we propose a method to chose the autonomous nodes in order to optimize the robustness of coupled networks to failures. We find the betweenness and the degree to be the key parameters for the selection of such nodes and we disclose the former as the most effective for modular networks. Considering the real case of the Italian communication network coupled with the power grid, we show in Fig. 1 that protecting only the four communication servers with highest betweenness reduces the chances of catastrophic failures like that witnessed during the blackout in 2003. When this strategy is implemented the resilience to random failures or attacks is significantly improved and the fraction of autonomous nodes necessary to change the nature of the percolation transition, from discontinuous to continuous, is significantly reduced. We also show that, even for networks with a narrow distribution of node degree like Erdős-Rényi graphs, the robustness can be significantly improved by properly choosing a small fraction of nodes to be autonomous. As a follow-up it would be interesting to understand how correlation between nodes, as well as dynamic processes on the network, can influence the selection of autonomous nodes. Besides, the cascade phenomena and the mitigation of vulnerabilities on regular lattices and geographically embedded networks are still open questions. It is important to note that while we use here high betweenness and high degree as a criterion for autonomous nodes, it is possible that other metrics will be also useful. For example, the eigenvector component of the largest eigenvector of the adjacency matrix (even weighted) makes a very good candidate (see e.g. Ref. 28).

Methods

We consider two coupled networks, A and B, where a fraction of 1 – p A-nodes fails. The cascade of failures can be described by the iterative equations, Eqs. (2)4,7, where αn and βn are, respectively, the fraction of A and B surviving nodes at iteration step n (not necessarily in the largest component) and Sxn) (χ = α|β, x = A|B) is the fraction of nodes in the largest component in network x given that 1 – χ nodes have failed. This can be calculated for coupled networks in the thermodynamic limit (N → ∞) using generating functions.

Random Protection

As proposed by Parshani et al.7, when autonomous nodes are randomly selected and the degree of coupling is the same in A and B, the set of Eqs. (2) simplifies to

where q is the degree of coupling. The degree distribution of the networks does not change in the case of random failures and Sxn) can be calculated as 27, where is the generating function of the degree distribution of network x,

and ux satisfies the transcendental equation

The size of the largest component in network x is given by χSx(χ).

For ER networks , where 〈kx is the average number of links in network x and therefore

With the above equations one can calculate the size of the largest component in both networks at the end of the cascade process.

Recently, Son et al.8 proposed an equivalent scheme based on epidemic spreading to solve the random protection case.

High Degree Protection

When autonomous nodes are selected following the degree strategy, the fraction of dependent nodes qxn changes with the iteration step n and the set of Eqs. 2 no longer simplifies. We divide the discussion below into three different parts: the degree distribution, the largest component and the coupling (fraction of dependent nodes).

The Degree Distribution

The networks A and B are characterized by their degree distributions, PA(k) and PB(k), which are not necessarily the same. The developed formalism applies to any arbitrary degree distribution. We start by first splitting the degree distribution into two parts, the component corresponding to the low-degree dependent nodes, PxD(k) and the component corresponding to the high-degree autonomous ones, PxI(k). To accomplish this, one must determine two parameters, the maximum degree of dependent nodes, kxm and the fraction of nodes with degree kxm that are coupled with the other network, fxm. These two parameters can be obtained from the relations,

and

where qx is the initial degree of coupling. One can then write

and

In the model, a fraction of 1 – p A-nodes are randomly removed. If, at iteration step n, αn nodes survive (αnp), p(1 – qA) nodes are necessarily autonomous and the remaining ones, αnp(1 – qA), are dependent nodes. One can then show that the degree distribution of network A, under the failure of 1 – αn nodes, , is given by

while the fraction of surviving links is

All the B-nodes which do not survive are dependent and so the degree distribution at iteration n, , is given by

while the fraction of surviving links is

The Largest Component

With the degree distribution and the fraction of surviving links pxn one can calculate the size of the largest component as

where τxn satisfies the self consistent equation

The coupling

To calculate the fraction qα,n (and qβ,n) one must first calculate the degree distribution of the nodes in the largest component. This is given by

The fraction of nodes in the largest component that are autonomous is then given by

where the upper limit of the sum is the maximum degree in the network, which we consider to be infinity in the thermodynamic limit. The fraction of autonomous nodes from the original network remaining in the largest component is qxG,nχnSxn), while the total fraction of autonomous nodes is given by 1 – q. The fraction of nodes disconnected from the largest component that are autonomous is then given by

so that the fraction of dependent nodes which have fragmented from the largest component is

For simplicity, here we assume that kx,m and fxm are constant and do not change during the iterative process. In fact, this is an approximation as the degree of the autonomous nodes is expected to change when their neighbors fail. However, in spite of shifting the transition point, this consideration does not change the global picture described here.

Numerical simulations

Numerical results have been obtained with the efficient algorithm described in Ref. 29 for coupled networks.