Abstract
The heart of the scientific enterprise is a rational effort to understand the causes behind the phenomena we observe. In largescale complex dynamical systems such as the Earth system, real experiments are rarely feasible. However, a rapidly increasing amount of observational and simulated data opens up the use of novel datadriven causal methods beyond the commonly adopted correlation techniques. Here, we give an overview of causal inference frameworks and identify promising generic application cases common in Earth system sciences and beyond. We discuss challenges and initiate the benchmark platform causeme.net to close the gap between method users and developers.
Introduction
Since Galileo Galilei, insight into the causes behind the phenomena we observe has come from two strands of modern science: observational discoveries and carefully designed experiments that intervene in the system of interest under wellcontrolled conditions. In one of Galilei’s early experiments—albeit a thought experiment^{1}—, the law of falling bodies is discovered by dropping two cannonballs of different masses from the tower of Pisa and measuring the effect of mass on the rate of fall to the ground. Discovering physical laws this way is a challenging problem when studying largescale complex dynamical systems such as the Earth system, because replicated interventional experiments are either infeasible or ethically problematic^{2}. Surely, we should not conduct largescale experiments on the Earth’s atmosphere: anthropogenic climate change already represents a rather uncontrolled longterm experiment. While randomized controlled experiments are a standard approach in medicine and the social sciences^{3,4}, the main current alternative within most disciplines of Earth sciences are computer simulation experiments. However, these are very expensive, timeconsuming, and require substantial amounts of expert knowledge, which in turn may impose strong mechanistic assumptions on the system^{2}. Fortunately, recent decades have seen an explosion in the availability of largescale time series data, both from observations (satellite remote sensing^{5}, stationbased, or field site measurements^{6}), and from Earth system model outputs^{2}. Such data repositories, together with increasing computational power^{7}, open up novel ways to use datadriven methods for the alternative strand of modern science: observational causal discoveries.
In recent years, rapid progress has been made in computer science, physics, statistics, philosophy, and applied fields to infer and quantify potential causal dependencies from time series data without the need to intervene in systems. Although the truism that correlation does not imply causation holds, the key idea shared by several approaches follows Reichenbach’s common cause principle^{8}: if variables are dependent then they are either causal to each other (in either direction) or driven by a common driver. To estimate causal relationships among variables, different methods take different, partially strong, assumptions. Granger^{9} addressed this question quantitatively using prediction, while in the last decades a number of complementary concepts emerged, from nonlinear dynamics^{10,11} based on attractor reconstruction, to computer science exploiting statistical independence relations in the data^{4,12}. More recently, research in statistics and machine learning utilizes the framework of structural causal models (SCMs)^{13} for this purpose. Causal inference is growing to become a mature scientific approach^{14}.
In contrast to datadriven machine learning methods such as probabilistic modeling^{15}, kernel machines^{16}, or in particular deep learning^{17}, which mainly focus on prediction and classification, causal inference methods aim at discovering and quantifying the causal interdependencies of the underlying system. Although interpreting deep learning models is an active area of research^{18}, extracting the causes of particular phenomena, e.g., hurricanes, from a deep learning black box is usually not possible. Therefore, causal inference methods are crucial in complementing predictive machine learning to improve our theoretical understanding of the underlying system^{19}.
Unfortunately, many causal inference methods are still only known within a small community of methodological developers and rarely adopted in applied fields like Earth system sciences. Yet, databased inference of causation was already proposed in the early 20th century by the geneticist Wright^{20}, but it has not been widely adopted partly due to the fierce opposition of statisticians like Pearson^{14}. In Earth system sciences, besides simulation experiments, (Pearson) correlation and regression methods are still the most commonly used tools. However, causal inference methods do have the potential to substantially advance the stateoftheart—if the underlying assumptions and methodological challenges are taken into consideration.
With this Perspective, we aim to bridge the gap between potential users and developers of methods for causal inference. We discuss the potential of applying causal inference methods to four key generic problems that are also common in other fields: causal hypothesis testing, causal network analysis, exploratory causal driver detection, and causal evaluation of physical models. First, we provide examples where causal inference methods have already led to important insights in Earth system sciences before giving an overview of different methodological concepts. Next, we highlight key generic problems in Earth system sciences and outline new ways to tackle these within causal inference frameworks. These problems are translated into challenges from a methodological perspective. Finally, as a way forward, we give recommendations for further methodological research as well as new ways in which causal inference methods and traditional physical modeling can complement each other, in particular in the context of climate change research. This Perspective is accompanied by a website (causeme.net) hosting a causality benchmark platform to spur more focused methodological research and provide benchmarks useful not only in Earth system sciences, but also in related fields with similar methodological challenges.
Example applications of causal inference methods
As in many other fields, methods based on correlation and univariate regression are still the most common databased tools to analyze relationships in Earth system sciences. Such association approaches are useful in daily practice, but provide few insights into the causal mechanisms that underlie the dynamics of a system. Causal inference methods can overcome some of the key shortcomings of such approaches. In this section, we discuss application examples where causal inference methods have already led to important insights before providing a systematic overview.
Concurrently to Wright’s^{20} seminal works on causation in the 1920s, Walker was the first to introduce systematic correlation and regression analysis into climate science^{21}. He discovered the temperature and pressure relationships between the East and West Pacific giving rise to the Walker circulation, which has by now been established not only from observational studies, but also detailed physical simulation experiments^{22}. In Fig. 1a, we illustrate these relationships using different methods: classical correlation, standard bivariate Granger causality (GC), and PCMCI^{23,24} (described later) that is better suited to this problem. Whereas GC and standard correlation analysis results in unphysical links, the example demonstrates that with the correct application of an appropriate method the Walker circulation can be inferred from data alone.
Similary, Kretschmer et al.^{25} investigated possible Arctic mechanisms which could be pivotal to understand northern hemisphere midlatitude extreme winters in Eurasia and North America. Arctic teleconnection patterns are much less understood than tropical ones and datadriven causality analyses are especially important because different climate models partly give conflicting results^{26,27}. In Fig. 1b we highlight the Arctic teleconnection pathways of the stratospheric Polar vortex that were extracted from observational data alone: here causal inference methods have confirmed previous model simulation studies, finding that Arctic sea ice extent in autumn is an important driver of winter circulation in the midlatitudes^{28}.
Finally, Fig. 1c shows an example from ecology demonstrating that traditional regression analysis is unable to identify the complex nonlinear interactions among sardines, anchovy, and sea surface temperature in the California Current ecosystem. A nonlinear causal statespace reconstruction method^{11} here extracts the underlying ecologically plausible network of interactions, revealing that sea surface temperatures are a common driver of both sardine and anchovy abundances.
These examples demonstrate how causal inference methods can help in distinguishing direct from indirect links and common drivers from observational time series, while classical correlation methods are ambiguous to interpret and can lead to incorrect conclusions.
Next to Granger’s seminal works in economics^{9,29}, observational causal inference methods have mostly been applied in neuroscience^{30,31} and bioinformatics^{32,33} where observational causal inference can also be combined with interventional experiments. The challenges for causal inference on Earth system data, especially the spatiotemporal and nonlinear nature of the system, are more similar to those in neuroscience as further discussed in the application and challenges sections.
Overview of causal inference methods
Observational causal inference from time series has come a long way since Wiener’s^{34} and Granger’s^{9} seminal works in the 1950s and 1960s and a plethora of different methods have been developed since then. Importantly, in the past few decades the works of Pearl, Spirtes, Glymour, Scheines, and Rubin^{3,4,12,35} have grounded causal reasoning and inference as a rigorous mathematical framework, elucidating the conditions under which discovering causal graphical models, also called Bayesian networks^{36}, from purely observational data is at all possible. These are known as identifiability conditions in the field of statistics and causal inference. Many causal inference methods for time series are grounded on the assumptions of timeorder (causes precede effects), Causal Sufficiency, meaning that all direct common drivers are observed, and the Causal Markov Condition, stating that in a graphical model a variable Y is independent of every other variable (that is not affected by Y) conditional on Y’s direct causes, among other, more technical, assumptions^{12,24}. However, recent work shows that some of these assumptions can be relaxed. Peters et al.^{13} summarize recent progress of methods that utilize assumptions on the noise structure and dependency types in the framework of SCMs. Many causal inference methods are not restricted to time series to infer causal relations.
Granger causality
The concept of Granger causality^{9} was the first formalization of a practically quantifiable causality definition from time series. The original idea, based on work by Wiener^{34}, is to test whether omitting the past of a time series X in a time series model including Y’s own and other covariates’ past increases the prediction error of the next time step of Y (Fig. 2a). The concept of GC can be implemented with different time series models. Classically, the Granger causality test is based on linear autoregressive modeling (see Box 1), but nonlinear dependencies can be modeled with more complex time series models or even the informationtheoretic analog transfer entropy^{37}. While bivariate time series models do not explicitly account for indirect links or common drivers as shown in Fig. 1a, more variables can be included in multivariate extensions of GC. Nevertheless, as illustrated in Box 1 GC is limited to lagged causal dependencies and, furthermore, has known deficiencies in the presence of subsampled time series and other issues^{38}. GC has a long history of applications across a wide range of scientific domains, including Earth system science^{39,40,41}.
Nonlinear statespace methods
While GC and also the other frameworks discussed here view systems as having interactions that arise from an underlying stochastic process, convergent crossmapping^{11} (CCM) and related methods^{10,42} take a different dynamical systems perspective. These methods assume that interactions occur in an underlying dynamical system and attempt to uncover causal relationships based on Takens’ theorem and nonlinear statespace reconstruction. Thus, for these methods to apply it is necessary to demonstrate that a deterministic nonlinear attractor can be recovered from the data. In this sense it is thought to be complementary to the more statistical approaches discussed here. As illustrated in Fig. 2b, a causal relationship between two dynamical variables X and Y can be established if they belong to a common dynamical system, which can be reconstructed from timedelay embedding of each of the observed time series. More specifically, if variable X can be predicted using the reconstructed system based on the timedelay embedding of variable Y, then we know that X had a causal effect on Y. Nonlinear statespace methods have been applied to ecology^{11,43} as shown in Fig. 1c, as well as in climate science^{44}.
Causal network learning algorithms
For time series that are of a stochastic nature, CCM is less well suited. Multivariate extensions of GC fail if too many variables are considered or dependencies are contemporaneous due to timesampling^{24} and in other cases (see also the challenges section). Causal network learning algorithms of various types have been developed for the reconstruction of largescale causal graphical models. They can be classified by their search architecture, that is, whether they start with an empty or fully connected graph, and the statistical criterion for removing or adding an edge. The common feature of these algorithms is that they assume the Markov condition mentioned above together with the Faithfulness assumption, which requires that all observed conditional independencies arise from the causal structure^{12}. Taken together, these two conditions allow to infer information about causal interactions from testing which conditional independencies hold true for the observed data. For example, the PC algorithm^{45} (named after its inventors Peter and Clark) and related approaches^{23,24,46,47} start with a fully connected graph and test for the removal of a link between two variables iteratively based on conditioning sets of growing cardinality (Fig. 2c). In this way also causal directions for contemporaneous links can often be assessed. Greedy equivalence search^{48}, on the other hand, starts with an empty graph and iteratively adds edges. The statistical criterion for removing or adding an edge can either be a conditional independence test or a properly defined score function that quantifies the likelihood of a particular graph structure given the data. Conditional independencies can flexibly be tested with different types of tests: Linear conditional independence can be assessed with partial correlation, while a wealth of recent machine learning approaches on nonparametric tests addresses a wide range of independence and dependence types^{24,49,50}. Score functions can be based on Bayesian or informationtheoretic approaches. Sun et al.^{51}, for example, cast causal network learning as an informationtheoretic optimization problem. Causal network learning algorithms can incorporate timeorder as a constraint (causes precede effects) and utilize a set of causal orientation rules to identify causal directions. The PCbased method PCMCI^{23,24} applied in Fig. 1a addresses the particular challenges of autocorrelated highdimensional and nonlinear time series data based on a conditionselection step (PC), followed by the momentary conditional independence (MCI) test. As illustrated in Box 1, some network learning approaches, e.g., FCI^{12}, account for unobserved direct common drivers and can still partially identify which links must be causal. Causal network learning algorithms have started to be applied in Earth system sciences only recently, mainly focusing on climate science^{23,25,52,53}.
Structural causal model framework
GC requires a time delay between cause and effect to identify causal directionality. If causation occurs almost instantaneously, or at least faster than the observable sampling interval, then causal directions cannot be identified in general. Many causal network learning algorithms, on the other hand, are also applicable to contemporaneous dependencies, but they can only identify causal graphs up to a Markovequivalence class. For example, under the Faithfulness assumption, measuring that X is conditionally independent of Y given Z, while all other (conditional) relationships are dependent, gives rise to three different causal graphs that are Markovequivalent if no additional information about timeorder is available: X\(\leftarrow\)Z→Y, X→Z→Y, or X\(\leftarrow\)Z\(\leftarrow\)Y. As illustrated in Box 1, the simplest example of Markov equivalence are two contemporaneously dependent variables where the causal direction cannot be inferred with conditional independencebased methods. Structural causal models (SCMs) (Fig. 2d) can identify causal directions in such cases because they permit assumptions about the functional class of models (e.g., linear or nonlinear, additivity, noise distributions)^{54,55,56}. Other methods exploit heterogeneity in the data by searching for models that are invariant over space or time^{57,58,59,60,61}. For an overview see references^{13,38}. Most of these principles extend to settings with temporal dependence as further elaborated in the Way forward section. SCMs have not yet been applied in Earth system sciences except for one work in remote sensing^{62}.
Key generic problems in Earth system sciences
Causal hypothesis testing
We start by illustrating the challenges associated with a key causal hypothesis testing problem in climate research. Midlatitude weather (including extreme events) is largely determined by nonlinear dynamical interactions between jet streams, storm tracks, and lowfrequency teleconnections^{63}. These dynamical processes are partially not well represented in the latest climate models. Hence, understanding drivers and favorable boundary conditions of weatherdetermining circulation regimes is crucial to improve (sub)seasonal predictions, evaluate climate models, and reduce uncertainty in regional climate projections^{64}. Important questions (Fig. 3a) in this context include: what drives the strength, position, and shape of the jet stream? What is the relative importance of tropical and Arctic processes^{26,28,65}? Uncovering causal relations from the observational record here raises a number of challenges. To name just a few, first, time series representing the climatologically relevant subprocesses need to be extracted from typically gridded spatiotemporal datasets^{25,66}, as illustrated in Fig. 3a. This can, for example, be achieved by averaging over corresponding regions, defining an index describing the jet stream position, or a more datadriven approach using dimensionreduction methods^{66}. Secondly, reconstructing the causal relations between these extracted variables is challenging because different nonlinear processes can interact on vastly different time scales from fast synoptic and cloudradiative processes to multiyear variability driven by slow oceanic processes^{67}. Last, the distributions of climate variables, for example precipitation, are often nonGaussian. Similar data characteristics also occur in neuroscience where first different subprocesses of the brain need to be reconstructed, e.g., from spatiotemporal electroencephalography measurements, and time series reflect a multitude of processes operating on different frequencies^{30,68}.
Causal complex network analysis
Network analysis of complex systems is a rapidly growing field^{69} and the network perspective may help to identify aggregate and emergent properties of the human brain^{68} or the Earth system^{66}. For example, a phenomenon such as El Niño results from the complex interplay between multiple processes in the tropical Pacific^{70} and has a large effect on the global climate system. In standard approaches^{68,71}, nodes are defined as the time series at different grid locations and links are typically based on correlations between the grid point time series. A common network measure is the node degree, which quantifies the number of processes linked to a node. However, defined based on correlations, network measures^{69} do not allow for a causal interpretation such as the information flow within the system^{71}. Grounding network theory in causal networks allows to better interpret network measures^{66,72}: an example for linear measures is reproduced in Fig. 3b. Like for the other generic problems, the challenges lie in highdimensional nonlinear spatiotemporal data, and here also in a proper definition of network measures that takes into account causal interactions and accounts for the spatial definition of nodes. Causal network comparison metrics can then be utilized for a causal evaluation of physical models (see last paragraph in this section).
Exploratory detection of causes of extreme impacts
In the Earth system, as well as in many other complex systems, the most devastating impacts are often related to multiple, compound or synergistic drivers^{73}. For instance, devastating wildfires need dry and hot conditions, available fuel, and an ignition source. Many impacts are related to threshold behavior^{74}, and multiple drivers contribute to the tipping of the system^{75,76}. Consider the example shown in Fig. 3c where only the synergistic combination of extreme inland precipitation and extreme storm surge leads to coastal floods^{77}. Causal inference methods can be helpful in identifying the relevant drivers from a typically large number of potential drivers that may be correlated with impacts^{78}. Causal methods further allow us to identify regime shifts in functional relationships that are, e.g., triggered by extreme conditions. The challenges here include highdimensionality, synergistic effects, and the often small sample size of observed impacts, and are relevant also in other fields such as neuroscience^{68}.
Causal evaluation of physical models
In many disciplines of Earth system sciences, models of the system or subsystem play a fundamental role in understanding relevant processes. Models differ regarding which subprocesses are resolved and the type of parametrization used. Biogeochemical models, for instance, help to understand element cycles and are a crucial basis for carbonclimate feedbacks in the coupled Earth system. At a higher level, climate models^{2,79} simulate the interactions of the atmosphere, water bodies, land surface and the cryosphere. In all cases, and at all levels, models are based partly on differential equations representing known processes and partly on semiempirical relationships representing unknown processes or approximating known processes that cannot be resolved at the global scale due to numerical issues^{80}. Due to the nonlinear nature of the system, small differences in parameterization can potentially lead to large deviations in overall model characteristics. A key task is to evaluate which model better simulates the real system. Currently, such evaluations are based on simple descriptive statistics like mean and variance, climatologies, and spectral properties of model output and observations^{2,79}. However, even though a particular model might well fit descriptive statistics of the observational data, for example, the global distribution of gross primary production (GPP) (Fig. 3d), the model might not well simulate the physical mechanisms affecting GPP, given that multiple model formulations and parameterizations, even when wrong, can fit the observations equally well, a problem known as underdetermination or equifinality^{81}. As a complementary criterion we propose to compare reconstructed causal dependencies of models and observational data (Fig. 3d). The underlying premise is that causal dependencies are more directly linked to the physical processes and are, therefore, more robust against overfitting than simple statistics and, hence, models that are causally similar to observations will also yield more reliable future projections. As for the previous example, also here the challenges lie in extracting suitable causal variables from often noisy stationbased measurements or highdimensional spatiotemporal fields and also the fact that processes can interact nonlinearly involving different spatiotemporal scales. In addition, model output may not satisfy the conditions underlying some causal inference methods, e.g., if dependencies are purely deterministic. Finally, suitable evaluation and comparison statistics based on causal networks need to be defined (see paragraph on causal complex network analysis). In Earth system sciences, model evaluation can help to build more realistic models to improve projections of the future, which is highly relevant for policy making^{82}.
Challenges from a methodological perspective
Process challenges
At the process level, a number of challenges arise due to the timedependent nature of the processes giving rise to strong autocorrelation (Fig. 4, point 1) and time delays (Fig. 4, point 2). Next, ubiquitous nonlinearity (Fig. 4, point 3), also in the form of statedependence (Fig. 4, point 4) and synergy (see Fig. 3c), requires a careful selection of the estimation method (see nonlinear methods in method overview section). Note that sometimes variables from model output can be deterministically related via a set of equations, which poses a serious problem for many, but not all, causal methods^{12,24}. As mentioned in the jet stream example, a geoscientific time series will typically contain signals from different processes acting on vastly different time scales, e.g., oceanic and atmospheric ones, which may need to be disentangled to better interpret causal links (Fig. 4, point 5). A basic assumption in a number of statistical methods used in causal inference frameworks (e.g., linear regression) is the assumption that the noise distribution is Gaussian, which is violated by processes featuring heavy tails and extreme events (e.g., precipitation; Fig. 4, point 6). On the other hand, some methods turn nonGaussianity into an advantage^{54} (Fig. 2d).
Data challenges
At the data aggregation level, our generic examples demonstrate that a major challenge is to define and reconstruct the causally relevant variables that represent the subprocesses of interest (Fig. 4, point 7). These variables have to be extracted from typically highdimensional spatiotemporal gridded datasets (e.g., from satellite observations or model output) or station data measurements, which can be done by dimensionality reduction methods. Moreover, these extracted variables should be interpretable and represent physical subprocesses of the system.
Often, relevant drivers cannot be measured, which requires to consider the possibility of unobserved variables (Fig. 4, point 8) regarding a causal interpretation of the estimated graph, since they may render detected links spurious (see also Box 1). Arguably, identifying the absence of a causal link, implying that a physical mechanism is unlikely^{24}, is a more robust finding, which requires less strong assumptions (no Causal Sufficiency). Another aspect of Causal Sufficiency is that not taking into account important drivers, such as anthropogenic climate forcings, may render time series nonstationary. Time series pose a particular challenge regarding timesubsampling (Fig. 4, point 9), which can also be considered as a case of unobserved samples of a variable, and timeaggregation (Fig. 4, point 10) which can let causal dependencies appear contemporaneous or even cyclic. The standard GC cannot deal with contemporaneous links, which can be identified using network learning algorithms or SCMs (see also Box 1).
On the data quality side, satellites, as well as station instruments, are plagued by all kinds of measurement errors (Fig. 4, point 11) such as observational noise, systematic biases, and also missing values (notably cloud occlusions or sensor malfunctioning). These may also be causally related to the measured process, constituting a form of selection bias (Fig. 4, point 12).
While in Earth system sciences the data will often attain a continuous range of values (e.g., temperature), variables can also be of a discrete type (Fig. 4, point 13), either due to quantization, or as categorical data. For example, one may be interested in causal drivers of an index representing different weather regimes or a time series of rarely occurring extreme events, which additionally raises the challenge of class imbalance—many 0 and few 1. Causal inference problems with such data require a suitable choice of methods, for example, conditional independence tests adapted to mixed data types. For paleoclimate data, the assumption of a time order is challenged since the measurement time points typically are given only with uncertainty (Fig. 4, point 14).
Computational and statistical challenges
From a computational and statistical point of view, scalability is a crucial issue, both regarding sample size (Fig. 4, point 15) and high dimensionality (Fig. 4, point 16). While larger sample sizes (long time series) are typically always beneficial for more reliable causal inferences, the computational time of methods may scale unfavorably with sample size (e.g., cubically for some kernel methods^{16}). The more variables are taken into account for explaining a potentially spurious relationship, the more credible a causal discovery becomes. However, many variables together with large time lags to account for physical time delays (e.g., to identify atmospheric teleconnections), lead to high dimensionality which may strongly affect statistical reliability. This compromises statistical power, that is, the probability to detect a true causal link, and potentially also the control of false positives at a desired significance level^{23,24}. Lowstatistical power implies that, especially, weak causal effects with low signaltonoise ratio, which are sometimes of interest, are not well detected. Last, uncertainty estimation (Fig. 4, point 17) that also takes into account potentially available data uncertainties (measurement value as well as dating uncertainties, see points 11 and 14), poses a major challenge for causal inference methods.
Most of the challenges discussed in this section are the same for correlation or regression methods which are, in addition, ambiguous to interpret and often lead to incorrect conclusions as shown in the examples section. We therefore emphasize that there is no strong reason to avoid adoption and exploration of modern causal inference techniques. Each of the methods summarized in the method overview section addresses one or several of these challenges. In Table 1 we list key strengths and suggest future research directions further discussed in the next section.
Finally, a crucial challenge when interpreting the output of causal inference methods is that causal conclusions are based on the assumptions underlying the different methods^{12,13,24}. These assumptions should, but often cannot, be tested and it is important to make them transparent and discuss how different assumptions would alter conclusions for a particular application.
Way forward
Avenues of further methodological research
The preceding Earth system sciences challenges (Fig. 4) are rather generic for complex dynamical systems and apply to many other fields. The challenges point to a way forward to advance causal inference methods for such systems. In the short term, our example applications demonstrate that the existing methods already address some of the mentioned challenges. For example, PCMCI was developed to address highdimensional timelagged linear and nonlinear causal discovery and takes into account autocorrelation^{23,24} and CCM^{11} was specifically built to account for nonlinear statedependent relationships. The largest potential for shortterm methodological advancements lies in combining different conceptual approaches in order to address multiple challenges.
First, to give some examples, such as those listed in Table 1, causal network learning algorithms that deal well with highdimensional data are limited by their inability to identify causal directionality among Markov equivalence classes^{12}. This shortcoming can be alleviated by combining causal network learning algorithms with the SCM framework and making additional assumptions on (independence of) mechanisms^{4,13,57,83} that permits to identify causal directions in these cases. Secondly, novel methods can incorporate ideas from theory on causal discovery in the presence of unobserved variables and selection bias^{12,47}, timesubsampling^{84,85}, timeaggregation and cyclic feedbacks^{86}, and measurement error^{87}. Thirdly, filtering methods as preprocessing steps, e.g., based on wavelets^{88}, can help to disentangle causal relations on different time scales, in the simplest example by filtering out a confounder like the seasonal cycle.
In the midterm, it is worth exploring methods that have not been applied to Earth system data, but whose theoretical properties may render them suitable for the challenges at hand. For example, further methods that are based on the principle of independent mechanisms^{4,13,57,83} such as prediction invariance^{13,58,59,61} or causal discovery from nonstationary data^{60} can potentially make use of the ubiquitously present nonstationarity and external perturbations in Earth system data to infer causal structure. While the blackbox character of most machine learning algorithms and deep learning in particular does not lend itself directly to causal discovery, such tools can nevertheless be useful in many aspects of causal discovery. For example, Chalupka et al.^{89} use neural networks to reconstruct causal features from gridded time series datasets. Also conditional independence tests can be based on deep learning^{90} and causal inference can be phrased as a classification problem^{91}. And the other way around: causal knowledge, as argued by Pearl, should be incorporated into machine learning to yield more robust predictions and classifications, for example, in such unresolved problems as extrapolation and domain adaptation^{14}.
Validation and a benchmark platform
Method development and comparison require benchmark datasets with known causal ground truth for validation. Ideally, such ground truth comes from expert knowledge on real data or real experiments that can also be used for falsification of causal relationships predicted from observational causal inference methods. Unfortunately, in Earth system sciences such datasets currently exist only for expertlabeled causal relations among few variables (e.g., some bivariate examples in ref. ^{92}). To some extent, outofsample predictions can provide partial validation, but the main alternative in Earth system sciences is experiments from physical simulation models. Such experiments, however, are computationally expensive and carry the challenge how these have to be designed. A more tractable approach is to generate synthetic data with simple model systems that mimic properties and challenges of geoscientific data, but where the underlying ground truth is known. These can then be used to study the performance of causal inference methods for different challenges in realistic finite sample situations. From a practitioner’s perspective, it is important to find out which method is best suited for a particular task with particular challenges and for a particular set of assumptions. Synthetic data, adapted to the problem at hand, can be used to choose the right method including method parameters. As a first step to close the gap between method users and developers, we accompany this Perspective by a causality benchmark platform (causeme.net) with synthetic models mimicking real data challenges on which causal inference methods can be compared. Next to method comparison, the platform also calls for submissions of real and modeled data sets where the causal structure is known with high confidence. Insights from such benchmark studies are relevant also for many other fields.
Combining observational causal inference and physical modeling
In the long term, we envision that the two main approaches to understand the Earth system (observational data analysis and Earth system modeling) should become more and more integrated. On the one hand, the generic problem of model evaluation has outlined ways on how causal inference methods can be used to identify weaknesses of physical models and guide model improvement. Furthermore, the currently often heuristic parametrization schemes in physical models can be guided by causal analyses of the respective variables, similar to the proposal to utilize machine learning to systematically replace parametrization schemes^{19,93}. Causal discovery can also help to design computationally expensive physical model experiments more efficiently: causal relationships estimated from climate model control runs^{79} (long model runs with fixed preindustrial conditions) can provide guidance on where numerical experiments are useful and where causal effects are not to be expected.
On the other hand, physical constraints, either from theoretical knowledge or from experimental (modeling) results, can be used to regularize causal inference methods, for example, by defining variables, restricting functional classes, identifying expected noise distributions, time lags and time aggregation, or general data preprocessing. Even more integrated, novel causal inference methods can make combined use of observational as well as experimental data^{94,95} which has already led to fruitful insights in genetics. In Earth system sciences, also information from real experiments on subsystems can be incorporated, not on a large climatic scale^{2}, but for example from ecosystem^{96} and mesocosm experiments^{97} in ecological labs.
Detecting and attributing climate change
Detection and attribution approaches quantify the evidence for a causal link between external drivers of climate change and longterm changes in climatic variables^{2}. The goal is to first detect a change and then attribute this change to the contributions of multiple anthropogenic and natural forcings, and from internal variability^{2}. Importantly, the focus lies on the effects of longterm forcings on longterm climatic trends or also changes in, e.g., the frequency of extreme weather events. Such research questions require counterfactual worlds, which can only be constructed with climate models, that are then statistically analyzed. For example, the optimal fingerprinting method^{2} is based on attributing detected longterm responses to fingerprint patterns using multiple linear regression. Hannart et al.^{98} discuss the inclusion of Pearl’s^{4} causal counterfactual theory for a more rigorous foundation of detection and attribution studies.
Nevertheless, observational causal inference methods can help to improve climate models as discussed above and can also directly be used to analyze climate feedbacks in paleoclimate data^{44}, which is still challenging due to scarce available data and dating uncertainties (Fig. 4). Furthermore, the recent concept of emergent constraints attempts to identify an observable statistical relationship between a feature of interest and a future climate change signal. For example, climate sensitivity, i.e., the response of global mean temperature to greenhouse gas emissions, can be constrained this way^{82}. The underlying premise is, however, that today’s dependencies between the predictors and climate sensitivity represent actual physical processes that also hold under future climate change. Here causal discovery can give more robust insights by identifying causal predictors that are more likely to hold under future climate change scenarios.
Conclusions
The current stateoftheart in data analysis of the Earth system is still dominated by correlation and regression methods, despite the fact that these methods often lead to ambiguous and confounded results. Existing causality methods can already yield deeper insights from hypothesis testing to the causal evaluation of physical models—if the particular challenges of Earth system sciences are properly addressed. A major impediment to a much wider adoption of causal inference methods is the lack of a reliable benchmark database. We aim to fill this gap by the accompanying platform causeme.net which also includes links to accessible software packages. Applying and interpreting causal inference methods and integrating these with physical modeling, however, will also require more indepth training on methods in Earth system sciences. Moreover, datadriven causality analyses need to be designed carefully: They should be guided by expert knowledge of the system (requiring expertise from the relevant field) and interpreted based on the assumptions and limitations of the causality method used (requiring expertise from the causal inference method). Sensibly applied causal inference methods promise to substantially advance the stateoftheart in understanding complex dynamical systems from data also in many other fields with similar challenges as in Earth system sciences, if domain scientists and method developers closely work together—and join the ‘causal revolution’^{14}.
Data availability
This Perspective is accompanied by a website hosting a causality benchmark platform. causeme.net runs a fair use data policy by which data are made freely available to the public and the scientific community in the belief that their dissemination will lead to greater understanding and new scientific insights and that global scientific problems require international cooperation. Open access means that data are freely distributed without charge. Data download is unrestricted and requires only a free registration for web security reasons. The platform is intended as a system for causal inference method intercomparison in a consistent data environment.
References
 1.
Gendler, T. Galileo and the indispensability of scientific thought experiment. Br. J. Philos. Sci. 49, 397–424 (1998).
 2.
IPCC. Climate Change 2013: The Physical Science Basis. Contribution of Working Group I to the Fifth Assessment Report of the Intergovernmental Panel on Climate Change (eds. Stocker, T. F., Qin D., Plattner, G.K., Tignor, M., Allen, S.K., Boschung, J., Nauels, A., Xia, Y., Bex, V. & Midgley, P.M.). 1535 (Cambridge University Press, Cambridge, UK and New York, NY, USA, 2013).
 3.
Imbens, G. & Rubin, D. Causal Inference in Statistics, Social, and Biomedical Sciences. (Cambridge University Press, New York, NY, USA, 2015).
 4.
Pearl, J. Causality: Models, Reasoning, and Inference. (Cambridge University Press, New York, NY, USA, 2000).
 5.
Guo, H.D., Zhang, L. & Zhu, L.W. Earth observation big data for climate change research. Adv. Clim. Chang. Res. 6, 108–117 (2015).
 6.
Baldocchi, D., Chu, H. & Reichstein, M. Interannual variability of net and gross ecosystem carbon fluxes: a review. Agric. Meteorol. 249, 520–533 (2018).
 7.
Overpeck, JonathanT., Meehl, GeraldA., Bony, Sandrine & Easterling, DavidR. Climate data challenges in the 21st century. Science 331, 700–702 (2011).
 8.
Reichenbach, H. The Direction of Time. (University of California Press, Berkeley and Los Angeles, CA, USA, 1956).
 9.
Granger, C. W. J. Investigating causal relations by econometric models and crossspectral methods. Econometrica 37, 424–438 (1969).
 10.
Arnhold, J., Grassberger, P., Lehnertz, K. & Elger, C. E. A robust method for detecting interdependences: application to intracranially recorded EEG. Phys. D Nonlinear Phenom. 134, 419–430 (1999).
 11.
Sugihara, G. et al. Detecting causality in complex ecosystems. Science 338, 496–500 (2012).
 12.
Spirtes, P., Glymour, C. & Scheines, R. Causation, Prediction, and Search. (MIT Press, Cambridge, MA, USA, 2000).
 13.
Peters, J., Janzing, D. & Schölkopf, B. Elements of Causal Inference: Foundations and Learning Algorithms. (MIT Press, Cambridge, MA, USA, 2017).
 14.
Pearl, J. & Mackenzie, D. The Book of Why: The New Science of Cause and Effect. (Basic Books, New York, NY, USA, 2018).
 15.
Ghahramani, Z. Probabilistic machine learning and artificial intelligence. Nature 521, 452–459 (2015).
 16.
Schölkopf, B. & Smola, A. J. Learning with Kernels. (MIT Press, Cambridge, MA, USA, 2008).
 17.
Goodfellow, I., Bengio, Y. & Courville, A. Deep Learning. (MIT Press, Cambridge, MA, USA, 2016).
 18.
Montavon, G., Samek, W. & Müller, K. R. Methods for interpreting and understanding deep neural networks. Digit. Signal Process. 73, 1–15 (2018).
 19.
Reichstein, M. et al. Deep learning and process understanding for datadriven Earth system science. Nature 566, 195–204 (2019).
 20.
Wright, S. Correlation and causation. J. Agric. Res. 20, 557–585 (1921).
 21.
Walker, G. T. Correlation in seasonal variations of weather, VIII: A Preliminary Study of World. Weather. Mem. Indian Meteorol. Dep. 24, 75–131 (1923).
 22.
Lau, K.M. & Yang, S. Walker circulation. Encyclopedia of Atmospheric Sciences (eds. Holton, J. R., Curry, J. A. & Pyle, J. A.) pp. 2505–2510 (Academic Press, Cambridge, MA, USA 2003).
 23.
Runge, J., Nowack, P., Kretschmer, M., Flaxman, S. & Sejdinovic, D. Detecting causal associations in large nonlinear time series datasets. arXiv:1702.07007v2 [stat.ME] (2018).
 24.
Runge, J. Causal network reconstruction from time series: from theoretical assumptions to practical estimation. Chaos Interdiscip. J. Nonlinear Sci. 28, 075310 (2018).
 25.
Kretschmer, M., Coumou, D., Donges, J. F. & Runge, J. Using causal effect networks to analyze different arctic drivers of midlatitude winter circulation. J. Clim. 29, 4069–4081 (2016).
 26.
Screen, J. A. et al. Consistency and discrepancy in the atmospheric response to Arctic seaice loss across climate models. Nat. Geosci. 11, 155–163 (2018).
 27.
Shepherd, T. G. Climate change: effects of a warming Arctic. Science 353,, 989–990 (2016).
 28.
Kim, B.M. et al. Weakening of the stratospheric polar vortex by Arctic seaice loss. Nat. Commun. 5, 4646 (2014).
 29.
Hoover, K. D. Causality in economics and econometrics. In New Palgrave Dictionary of Economics. (eds Durlauf, S. N., & Blume, L. E.) 2nd ed. 2008 (Palgrave Macmillan, Basingstoke, UK, 2006).
 30.
Friston, K. J., Harrison, L. & Penny, W. Dynamic causal modelling. Neuroimage 19, 1273–1302 (2003).
 31.
Kaminski, M., Ding, M., Truccolo, W. A. & Bressler, S. L. Evaluating causal relations in neural systems: Granger causality, directed transfer function and statistical assessment of significance. Biol. Cybern. 85, 145–157 (2001).
 32.
Meinshausen, N. et al. Methods for causal inference from gene perturbation experiments and validation. Proc. Natl Acad. Sci. USA 113, 7361–7368 (2016).
 33.
Sachs, K., Perez, O., Pe’er, D., Lauffenburger, D. & Nolan, G. Causal ProteinSignaling Networks Derived from Multiparameter SingleCell Data. Science 308, 523–529 (2005).
 34.
N. Wiener. The Theory of Prediction. In Modern Mathematics for Engineers. (ed. Beckenbach, E.). (McGrawHill, New York, NY, 1956).
 35.
Rubin, D. B. Estimating causal effects of treatments in randomized and nonrandomized studies. J. Educ. Psychol. 66, 688–701 (1974).
 36.
Koller, D. & Friedman, N. Probabilistic Graphical Models: Principles and Techniques. (MIT Press, Cambridge, MA, 2009).
 37.
Schreiber, T. Measuring information transfer. Phys. Rev. Lett. 85, 461–464 (2000).
 38.
Spirtes, P. & Zhang, K. Causal discovery and inference: concepts and recent methodological advances. Appl. Inform. 3, 3 (2016).
 39.
Triacca, U. Is Granger causality analysis appropriate to investigate the relationship between atmospheric concentration of carbon dioxide and global surface air temperature? Theor. Appl. Climatol. 81, 133–135 (2005).
 40.
McGraw, M. C. & Barnes, E. A. Memory matters: a case for Granger causality in climate variability studies. J. Clim. 31, 3289–3300 (2018).
 41.
Papagiannopoulou, C. et al. A nonlinear Grangercausality framework to investigate climatevegetation dynamics. Geosci. Model Dev. 10, 1945–1960 (2017).
 42.
Hirata, Y. et al. Detecting causality by combined use of multiple methods: climate and brain examples. PLoS One 11, e0158572 (2016).
 43.
Ye, H., Deyle, E. R., Gilarranz, L. J. & Sugihara, G. Distinguishing timedelayed causal interactions using convergent cross mapping. Sci. Rep. 5, 14750 (2015).
 44.
Van Nes, E. H. et al. Causal feedbacks in climate change. Nat. Clim. Chang. 5, 445–448 (2015).
 45.
Spirtes, P. & Glymour, C. An algorithm for fast recovery of sparse causal graphs. Soc. Sci. Comput. Rev. 9, 62–72 (1991).
 46.
Verma, T. & Pearl, J. Causal networks: semantics and expressiveness. Mach. Intell. Pattern Recognit. 9, 69–76 (1990).
 47.
Zhang, J. On the completeness of orientation rules for causal discovery in the presence of latent confounders and selection bias. Artif. Intell. 172, 1873–1896 (2008).
 48.
Chickering, D. M. Learning equivalence classes of bayesiannetwork structures. J. Mach. Learn. Res. 2, 445–498 (2002).
 49.
Zhang, K., Peters, J., Janzing, D. & Schölkopf, B. Kernelbased conditional independence test and application in causal discovery. In Proceedings of the 27th Conference on Uncertainty in Artificial Intelligence (eds. Cozman, F. Pfeffer, A.) 804–813 (AUAI Press, Corvallis, Oregon, USA, 2011).
 50.
Runge, J. Conditional independence testing based on a niearestneighbor estimator of conditional mutual information. In Proceedings of the 21st International Conference on Artificial Intelligence and Statistics, (ed. Storkey, A. & PerezCruz, F.) pp. 938–947. (Playa Blanca, Lanzarote, Canary Islands: PMLR, 2018).
 51.
Sun, J., Taylor, D. & Bollt, E. M. Causal network inference by optimal causation entropy. SIAM J. Appl. Dyn. Syst. 14, 27 (2014).
 52.
EbertUphoff, I. & Deng, Y. Causal discovery for climate research using graphical models. J. Clim. 25, 5648–5665 (2012).
 53.
Runge, J., Petoukhov, V. & Kurths, J. Quantifying the strength and delay of climatic interactions: the ambiguities of cross correlation and a novel measure based on graphical models. J. Clim. 27, 720–739 (2014).
 54.
Shimizu, S., Hoyer, P. O., Hyvärinen, A. & Kerminen, A. A linear nongaussian acyclic model for causal discovery. J. Mach. Learn. Res. 7, 2003–2030 (2006).
 55.
Hoyer, P. O., Janzing, D., Mooij, J. M., Peters, J. & Schölkopf, B. Nonlinear causal discovery with additive noise models. In Proceedings of the 30th Conference on Advances in Neural Information Processing Systems (eds. Bengio, Y., Schuurmans, D., Lafferty, J. D., Williams, C. K. I. & Culotta, A.) 689–696 (Curran Associates, Red Hook, NY, USA, 2009).
 56.
Zhang, K. & Hyvärinen, A. On the identifiability of the postnonlinear causal model. In Proceedings of the TwentyFifth Conference on Uncertainty in Artificial Intelligence (eds Bilmes, J., & Ng, A.) 647–655 (AUAI Press, Corvallis, Oregon, USA, 2009).
 57.
Schölkopf, B. et al. On causal and anticausal learning. In Proceedings of the 29th International Conference on Machine Learning (eds Langford, J., & Pineau, J.) 459–466 (Omnipress, Madison, WI, USA, 2012).
 58.
Peters, J., Bühlmann, P. & Meinshausen, N. Causal inference by using invariant prediction: identification and confidence intervals. J. R. Stat. Soc. Ser. B 78, 947–1012 (2016).
 59.
Eaton, D. & Murphy, K. P. Exact Bayesian structure learning from uncertain interventions. In Proccedings of the 11th International Conference on Artificial Intelligence and Statistics (eds Meila, M. & Shen, X.) pp. 107–114, (PMLR, San Juan, Puerto Rico, 2007).
 60.
Zhang, K., Huang, B., Zhang, J., Glymour, C. & Schölkopf, B. Causal discovery from nonstationary/heterogeneous data: skeleton estimation and orientation determination. In Proceedings of the 26th International Joint Conference on Artificial Intelligence 1347–1353 (ed. Sierra, C.) (International Joint Conferences on Artificial Intelligence, Marina del Rey, CA, USA, 2017).
 61.
Pfister, N., Bühlmann, P. & Peters, J. Invariant causal prediction for sequential data. J. Am. Stat. Assoc. 6, 1–13 (2018).
 62.
PerezSuay, A. & CampsValls, G. Causal inference in geoscience and remote sensing from observational. Data. IEEE Trans. Geosci. Remote Sens. 3, 1502–1513 (2018).
 63.
Li, J., Swinbank, R., Grotjahn, R. & Volkert, H. Dynamics and Predictability of LargeScale, HighImpact Weather and Climate Events. (Cambridge University Press, Cambridge, UK, 2016).
 64.
Shepherd, T. G. Atmospheric circulation as a source of uncertainty in climate change projections. Nat. Geosci. 7, 703–708 (2014).
 65.
Feldstein, S. B. & Lee, S. Intraseasonal and interdecadal jet shifts in the Northern Hemisphere: the role of warm pool tropical convection and sea ice. J. Clim. 27, 6497–6518 (2014).
 66.
Runge, J. et al. Identifying causal gateways and mediators in complex spatiotemporal systems. Nat. Commun. 6, 8502 (2015).
 67.
Williams, P. D. et al. A census of atmospheric variability from seconds to decades. Geophys. Res. Lett. 44, 201–11,211 (2017).
 68.
Bullmore, E. & Sporns, O. Complex brain networks: graph theoretical analysis of structural and functional systems. Nat. Rev. Neurosci. 10, 186–198 (2009).
 69.
Boccaletti, S., Latora, V., Moreno, Y., Chavez, M. & Hwang, D. U. Complex networks: structure and dynamics. Phys. Rep. 424, 175–308 (2006).
 70.
Ludescher, J. et al. Very early warning of next El Niño. Proc. Natl Acad. Sci. 111, 201323058 (2014).
 71.
Tsonis, A. A., Swanson, K. L. & Wang, G. On the role of atmospheric teleconnections in climate. J. Clim. 21, 2990–3001 (2008).
 72.
Runge, J. Quantifying information transfer and mediation along causal pathways in complex systems. Phys. Rev. E 92, 062829 (2015).
 73.
Zscheischler, J. et al. Future climate risk from compound events. Nat. Clim. Chang. 8, 469–477 (2018).
 74.
Reichstein, M. et al. Climate extremes and the carbon cycle. Nature 500, 287–295 (2013).
 75.
Martius, O., Pfahl, S. & Chevalier, C. A global quantification of compound precipitation and wind extremes. Geophys. Res. Lett. 43, 7709–7717 (2016).
 76.
Zscheischler, J. et al. Carbon cycle extremes during the 21st century in CMIP5 models: future evolution and attribution to climatic drivers. Geophys. Res. Lett. 41, 8853–8861 (2014).
 77.
Wahl, T., Jain, S., Bender, J., Meyers, S. D. & Luther, M. E. Increasing risk of compound flooding from storm surge and rainfall for major US cities. Nat. Clim. Chang. 5, 1093–1097 (2015).
 78.
Kretschmer, M., Runge, J. & Coumou, D. Early prediction of weak stratospheric polar vortex states using causal precursors. Geophys. Res. Lett. 44, 8592–8600 (2017).
 79.
Eyring, V. et al. Overview of the Coupled Model Intercomparison Project Phase 6 (CMIP6) experimental design and organization. Geosci. Model Dev. 9, 1937–1958 (2016).
 80.
Reichstein, M. Inverse modeling of seasonal drought effects on canopy CO_{2} /H_{2}O exchange in three Mediterranean ecosystems. J. Geophys. Res. 108, 4726 (2003).
 81.
Beven, K. & Freer, J. Equifinality, data assimilation, and uncertainty estimation in mechanistic modelling of complex environmental systems using the GLUE methodology. J. Hydrol. 249, 11–29 (2001).
 82.
Eyring, V. et al. Taking climate model evaluation to the next level. Nat. Clim. Chang. 9, 102–110 (2019).
 83.
Hoover, K. D. The logic of causal inference: econometrics and the conditional analysis of causation. Econ. Philos. 6, 207 (1990).
 84.
Gong, M., Zhang, K., Schölkopf, B., Tao, D. & Geiger, P. Discovering temporal causal relations from subsampled data. In Proceedings of the 32nd International Conference on Machine Learning (eds. Bach, F., Blei, D.) 1898–1906 (PMLR, Lille, France, 2015).
 85.
Plis, S., Danks, D., Freeman, C. & Calhoun, V. Rateagnostic (causal) structure learning. In Proceedings of the 28th Conference on Advances in Neural Information Processing Systems (eds. Cortes, C., Lawrence, N.D., Lee. D.D., Sugiyama, M., Garnett, R.) 3303–3311 (Curran Associates, Red Hook, NY, USA, 2015).
 86.
Gong, M., Zhang, K., Schölkopf, B., Glymour, C. & Tao, D. Causal discovery from temporally aggregated time series. In Proceedings of the 33rd Conference on Uncertainty in Artificial Intelligence ID 269 (eds. Elidan, G., Kersting, K., & Ihler, A.T.) (AUAI Press, Corvallis, Oregon, USA, 2017).
 87.
Zhang, K. et al. Causal discovery with linear nonGaussian models under measurement error: Structural identifiability results. In Proceedings of the 34th Conference on Uncertainty in Artificial Intelligence (eds. Globerson, A., & Silva, Ricardo) 1063–1072 (AUAI Press, Corvallis, Oregon, USA, 2018).
 88.
Percival, D. & Walden, A. Wavelet Methods for Time Series Analysis. (Cambridge University Press, Cambridge, UK, 2006).
 89.
Chalupka, K., Bischoff, T., Perona, P. & Eberhardt, F. Unsupervised Discovery of El Nino Using Causal Feature Learning on Microlevel Climate Data. In Proceedings of the 32nd Conference on Uncertainty in Artificial Intelligence (eds Ihler, A., & Janzing, D.) (AUAI Press, Corvallis, Oregon, USA, 2016).
 90.
Sen, R., Suresh, A. T., Shanmugam, K., Dimakis, A. G. & Shakkottai, S. ModelPowered Conditional Independence Test. In Proceedings of the 30th Conference on Advances in Neural Information Processing Systems (eds Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H.M., Fergus, R., Vishwanathan, S., & Garnett, R.) 2955–2965 (Curran Associates, Red Hook, NY, USA, 2017).
 91.
LopezPaz, D., Muandet, K., Schölkopf, B. & Tolstikhin, I. Towards a Learning Theory of CauseEffect Inference. In Proceedings of the 32nd International Conference on Machine Learning (eds Bach, & F., Blei, D.) 1452–1461 (PMLR, Lille, France, 2015).
 92.
Mooij, J. M., Peters, J., Janzing, D., Zscheischler, J. & Schölkopf, B. Distinguishing cause from effect using observational data: methods and benchmarks. J. Mach. Learn. Res. 17, 1–102 (2016).
 93.
Schneider, T., Lan, S., Stuart, A. & Teixeira, J. Earth system modeling 2.0: a blueprint for models that learn from observations and targeted highresolution simulations. Geophys. Res. Lett. 44, 396–12,417 (2017).
 94.
Hauser, A. & Bühlmann, P. Characterization and greedy learning of interventional Markov equivalence classes of directed acyclic graphs. J. Mach. Learn. Res. 13, 2409–2464 (2012).
 95.
Wang, Y., Solus, L., Yang, K. D. & Uhler, C. Permutationbased causal inference algorithms with interventions. In Proceedings of the 30th Conference on Advances in Neural Information Processing Systems (eds Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., & Garnett, R.) 5824–5833 (Curran Associates, Red Hook, NY, USA, 2017).
 96.
Wu, Z., Dijkstra, P., Koch, G. W., Peñuelas, J. & Hungate, B. A. Responses of terrestrial ecosystems to temperature and precipitation change: a metaanalysis of experimental manipulation. Glob. Chang. Biol. 17, 927–942 (2011).
 97.
Moss, B. et al. Continentalscale patterns of nutrient and fish effects on shallow lakes: synthesis of a panEuropean mesocosm experiment. Freshw. Biol. 49, 1633–1649 (2004).
 98.
Hannart, A., Pearl, J., Otto, F. E. L., Naveau, P. & Ghil, M. Causal counterfactual theory for the attribution of weather and climaterelated events. Bull. Am. Meteorol. Soc. 97, 99–110 (2016).
 99.
Barnett, L. & Seth, A. K. Granger causality for state space models. Phys. Rev. E 91, 040101 (2015).
 100.
Tsamardinos, I., Brown, L. E. & Aliferis, C. F. The maxmin hillclimbing Bayesian network structure learning algorithm. Mach. Learn. 65, 31–78 (2006).
Acknowledgements
This work emerged partly from discussions at the workshop “Causality in Complex Systems” in June 27–30, 2017, in Soesterberg, The Netherlands. We thank Michael Eichler, Peter Grassberger, Ingrid van de Leemput, Diego Miralles, Joris Mooij, and Els Weinans for stimulating discussions on this workshop. We also thank Veronika Eyring for helpful comments on the manuscript. We gratefully acknowledge the Netherlands Earth System Science Centre (NESSC) for funding this workshop. Furthermore, this work was supported by the German Federal Ministry of Education and Research (BMBF), grant 01LN1304A (M.K. and D.C.) and the Netherlands Organisation for Scientific Research (NWO), grant 016.Vidi.171011 (D.C.). J.R. was funded by a Postdoctoral Fellowship by the James S. McDonnell foundation. G.C.V. was supported by the European Research Council (ERC) under the ERC Consolidator Grant ERCCoG2014 SEDAL (grant agreement 647423). J.S.’s work was funded in part by the Simons Foundation Grant No. 318812. E.M.B. was supported by the U.S. Office of Naval Research (N000141512093). J.S. and E.M.B. were supported by the U.S. Army Research Office (W911NF1610081). M.D.M. and M.R. acknowledge funding from the European Union Horizon 2020 project BACI (GA No. 640176). G.S. and E.D. acknowledge funding by the DoDStrategic Environmental Research and Development Program 15 RC2509, the Lenfest Ocean Program 00028335, and the National Science Foundation grants NSFDEB1655203 and NSFABIInnovation DBI1667584. G.S. thanks the McQuown Fund and the McQuown Chair in Natural Sciences, University of California, San Diego. K.Z. acknowledges the support by the United States Air Force under Contract No. FA865017C7715 and by National Science Foundation EAGER Grant No. IIS1829681. C.G. and K.Z. acknowledge funding from National Institutes of Health under Contract No. NIH1R01EB02285801, FAINR01EB022858, NIH1R01LM012087, NIH5U54HG00854002, and FAINU54HG008540. J.Z. acknowledges funding from the Swiss National Science Foundation Ambizione grant PZ00P2179876. J.P. was supported by a research grant (18968) from VILLUM FONDEN.
Author information
Affiliations
Contributions
The article is a result of a workshop organized by J.R. and S.B., invited by M.S. J.R. wrote the first draft of the paper and the revised versions with input from all other authors. For the introduction, outlook, and conclusions section input was contributed in particular by S.B., E.D., G.C.V., D.C., C.G., M.D.M., J.P., M.R., B.S., R.Q., M.S., G.S. and J.Z., for the examples section in particular by E.D., M.K. and D.C., for the method overview section in particular by E.B., E.H.N., G.C.V., C.G., J.P., B.S., J.S., P.S. and K.Z. and for the applications section in particular by D.C., M.K., M.D.M., M.R., K.Z. and J.Z. Box 1 was created by J.R. All figures were created by J.R. with input from M.K., E.D., M.D.M., J.P. and J.Z. All authors discussed the content and contributed to editing the manuscript. The website causeme.net was created by J.M.M., G.C.V. and J.R.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Peer review information: Nature Communications thanks Auroop Ganguly and other anonymous reviewers for their contribution to the peer review of this work.
Publisher’s note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Runge, J., Bathiany, S., Bollt, E. et al. Inferring causation from time series in Earth system sciences. Nat Commun 10, 2553 (2019). https://doi.org/10.1038/s41467019101053
Received:
Accepted:
Published:
Further reading

Fractional NeuroSequential ARFIMALSTM for Financial Market Forecasting
IEEE Access (2020)

Challenges in Applying Machine Learning Models for Hydrological Inference: A Case Study for Flooding Events Across Germany
Water Resources Research (2020)

Monitoring ocean biogeochemistry with autonomous platforms
Nature Reviews Earth & Environment (2020)

A typology of compound weather and climate events
Nature Reviews Earth & Environment (2020)

Validity and Limitations of the Detection Matrix to Determine Hidden Units and Network Size from Perceptible Dynamics
Physical Review Letters (2020)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.