Main

Changes in precipitation substantially affect societies and ecosystems1. Therefore, it is of the utmost importance to determine whether anthropogenic changes in precipitation are detectable. At the planetary scale, global-climate-model simulations show that globally averaged precipitation will increase by approximately 1–3% per degree of warming7,8,9. This change is not spatially homogenous. Wet regions are projected to have the largest future increase, which is sometimes referred to as the ‘wet-gets-wetter’10 or ‘wettest-gets-wetter’ response. Moreover, areas that will experience greater ocean warming are also projected to show a mean intensification of rainfall (‘warmer-gets-wetter’)11, which may further influence large-scale atmospheric circulation. In accord with the theory, the intensity of extreme daily precipitation events is projected to increase at the rate of about 7% K−1 following the Clausius–Clapeyron relation in many parts of the world2,12, whereas higher rates of increase have been observed regionally13. However, owing to the wide range of spatiotemporal scales of precipitation variability, an unequivocal fingerprint of human influence in precipitation has not yet been established from observational records7,14.

Previous detection and attribution (D&A) studies12,15,16 have identified anthropogenic influences on preprocessed precipitation statistics, such as the annual maxima of daily precipitation over land areas and the seasonal or zonal averages of global17,18 and Arctic precipitation19. Although using spatial/temporal averages is beneficial for detection because it lowers the uncertainty related to natural internal variability, it is uncertain to what extent detection results based on these smoothed fields can be applied to hydrometeorological weather events that affect our daily lives2,20.

Determining whether and to what degree greenhouse-gas-induced warming has altered daily precipitation in the observational record remains elusive for two reasons. First, daily precipitation amounts exhibit large internal variabilities associated with non-anthropogenic weather noise, which hinders climate-change-signal detection8,21,22,23. Second, conventional D&A methods assume a fixed spatial pattern of the climate-change signal (that is, fingerprint pattern)24,25,26, which may not be sufficient to capture changes in higher-moment statistics such as variance. Therefore, efforts to detect climate-change signals imprinted in daily precipitation have thus far been unsuccessful (Extended Data Fig. 1). In this study, we overcome these two issues by combining large-ensemble climate-model simulations with a deep-learning algorithm and show that deep learning can detect statistically significant climate-change signals in daily precipitation fields.

Detection results from a deep-learning model

Our deep-learning model for D&A is based on a CNN, which is a widely used deep-learning technique for pattern recognition5. The algorithm takes global maps of daily precipitation anomalies (deviations from the long-term daily climatology) as an input variable and outputs an annual global mean 2 m air temperature (AGMT), which is a key climate-change metric24 (see Methods and Extended Data Fig. 2 for the detailed model structure). To build a deep network that can detect the climate-change signal amidst large internal variability, we trained our deep-learning model with pairs of daily precipitation maps and the AGMT simulated by 80 members of the CESM2 Large Ensemble (LE)6, which was forced from 1850 to 2100 with estimates of historical forcings and the SSP3-7.0 greenhouse-gas emission scenario (Methods). Being applied as a detection algorithm, the deep-learning model will be referred to as the deep detection (DD) model.

The convolutional process embedded in the CNN is able to capture local features in the global domain5, making it suitable for detecting regional pattern changes associated with global warming. Also, with its translation-invariant feature, the DD model can extract common change patterns owing to the global warming in both the model simulations and the observations despite their systematic differences27. This feature contrasts with the existing D&A techniques, including the revised linear-regression-based approaches26,28 and the feedforward neural networks29, which detect climate-change signals based on a global stationary fingerprint pattern.

After being trained using the CESM2 LE data, the DD model was applied to satellite precipitation observations (Integrated Multi-satellite Retrievals for GPM (IMERG)30 and Global Precipitation Climatology Project (GPCP)31), the gauge-satellite-reanalysis merged data (Multi-Source Weighted-Ensemble Precipitation (MSWEP)32) and precipitation data from a modern reanalysis product (ERA5) that assimilated ground-based radar and satellite data33 (Methods). We used several datasets in our detection analysis to account for the uncertainties associated with indirect estimates of precipitation.

With the historical record of observed precipitation data as input, our predicted AGMT from the DD model reproduces the observations very well, with larger increases during recent decades (Fig. 1a), suggesting the possible influence of global warming on recent daily precipitation fields. The Pearson correlation between the annually averaged observed and predicted AGMT from 1980 to 2020 was 0.88 for both MSWEP and ERA5, with P values of less than 0.001. Slightly lower correlations (0.74–0.85) were found for the latest 20 years (2001–2020). By contrast, the corresponding correlation coefficients obtained using the ridge regression method24 were systematically lower (0.33 and 0.36 during 1980–2020 for MSWEP and ERA5, respectively) (Extended Data Fig. 3a). This demonstrates that the DD model recognizes the global-warming signal in spatiotemporal features in daily precipitation, which has—so far—not been possible with standard linear detection methods.

Fig. 1: Climate-change detection using deep learning.
figure 1

a, Time series of the observed AGMT anomaly from 1980 to 2020 (black line) and the annual average of the estimated AGMT obtained by using daily precipitation fields from the MSWEP (blue line), IMERG (purple line) and GPCP (orange line) observations and ERA5 reanalysis (red line) as inputs in the DD model, whose temporal correlation with the observed AGMT is 0.74, 0.80, 0.76 and 0.85 during 2001–2020, respectively. The corresponding coloured dots denote the daily estimated AGMT using the MSWEP and ERA5 precipitation data. The dashed black horizontal lines denote a 95% confidence range of internal variability of the AGMT estimates, defined as the 2.5th–97.5th percentile of the daily estimated AGMT obtained from historical CESM2 LE simulations during 1850–1950. Observed and modelled anomalies are relative to 1980–2010 climatology. b, Fractional number of EM days within a corresponding year from 1980 to 2020 for which the estimated AGMT is greater than the upper bound of the 95% confidence range. Dashed line denotes an upper bound of the 95% confidence range of internal variability of fractional EM days, which is 10.9%. c, Linear trend of the number of EM days during 1980–2020 in ERA5 and MSWEP and during 2001–2020 in ERA5, MSWEP, IMERG and GPCP. The dashed lines denote upper bounds of the 95% confidence level based on the bootstrap method estimated using the historical CESM2 LE simulations (Methods).

To measure the detectability of the observed AGMT variations associated with daily precipitation fields using the DD model, we defined the internal variability range of the daily estimated AGMT as the 2.5th–97.5th percentile values obtained from the CESM2 LE simulation during the historical period from 1850 to 1950 (dashed range in Fig. 1a). The detection results showed that, from the mid-2010s onward, the annual average of the DD-predicted AGMT exceeded the upper bound of internal variability, thereby indicating that greenhouse warming already altered daily precipitation fields.

The days with the estimated AGMT greater than the upper bound of internal variability, henceforth referred to as emergence (EM) days, increased continuously after 1980 (Fig. 1b). The EM days in recent years lie clearly above the 97.5th percentile (10.9%) of internally generated EM days estimated using the CESM2 LE (Methods). From the mid-2010s onward, the climate-change signal can be detected from daily precipitation maps in more than half of all days each year (that is, >50% of the fractional EM days), regardless of the input data type.

The strong positive linear trends of EM days were found for all precipitation datasets: 17.1% decade−1 and 16.3% decade−1 during 1980–2020 for MSWEP and ERA5, respectively, and 21.3% decade−1 and 16.5% decade−1 for IMERG and GPCP during 2001–2020, respectively (Fig. 1c). These trends also exceeded the internal variability ranges of the EM days trends (dashed line in Fig. 1c; Methods). The detection result remained largely unaffected by the choice of the climate-model simulations34 used in the training of the DD models (Extended Data Fig. 4), demonstrating a generalization capability of the deep-learning model for climate-change detection. Unlike our DD-based results, the ridge regression exhibited almost no trend in the EM days during recent decades (Extended Data Fig. 3b,c), indicating that the signal is barely detectable with the linear approach.

Precipitation timescales and hotspot regions

To identify the source of the climate-change signal, we repeatedly ran the DD model using the satellite and reanalysis precipitation products with each time using anomalies capturing a different timescale. For this task, the precipitation anomalies were decomposed into a linear trend and high-frequency (<10-day), submonthly (10–30-day), subseasonal (30–90-day), subyearly (90-day–1-year) and low-frequency (>1-year) variabilities using Lanczos filtering35. The detection results for these different timescales were then compared with those using precipitation anomalies retaining all timescales.

When the linear trend component of precipitation anomalies was given to the DD model, the estimated AGMT decreased in time for both the ERA5 and MSWEP datasets during 1980–2020 (Fig. 2a). During 2001–2020, the results with the IMERG and GPCP datasets disagree on the sign of temporal changes in the estimated AGMT (Fig. 2b), possibly because of the discrepancies in the trend between the precipitation datasets36. Clearly, the mean state changes in precipitation represented by the linear trend is not the primary source of the anthropogenic climate-change signal found in Fig. 1. The negative contribution of the linear trend component found in three of the four precipitation datasets may have been partly caused by the recent negative phase of the Interdecadal Pacific Oscillation (IPO) in the tropical Pacific during the early twenty-first century and its associated precipitation response37,38. The observed interdecadal trends in tropical rainfall were considerably different from the climate-projection results simulated by global climate models39 (Extended Data Fig. 5a,b). The weak contribution of the linear trend component to climate-change detection aligns with the outcomes obtained through the ridge regression method using daily precipitation, in which fingerprint pattern is coherent in its signs with the spatial pattern of the climatological precipitation change (Extended Data Fig. 5c).

Fig. 2: Critical role of high-frequency precipitation variations in climate-change detection.
figure 2

a,b, Linear trends of the estimated AGMT from the DD model for 1980–2020 using ERA5 reanalysis (red) or MSWEP (blue) (a) and 2001–2020 using IMERG (purple) or GPCP (orange) (b). Each case shows results from using unfiltered precipitation anomalies (denoted as ‘Total’), linear trends of the precipitation anomalies (‘Trend’), 10-day high-pass-filtered (‘10d HP’), 10–30-day band-pass-filtered (‘10–30d BP’), 30–90-day band-pass-filtered (‘30–90d BP’), 90-day–1-year band-pass-filtered (‘90d–1y BP’) and 1-year low-pass-filtered (‘1y LP’) precipitation. The dashed black horizontal lines in panels a and b denote the upper bound of a 95% confidence range of internal variability of the estimated AGMT linear trend, obtained from historical CESM2 LE simulations during 1850–1950. c, Linear trend of the AGMT occlusion sensitivity for 10-day high-pass-filtered ERA5 and MSWEP precipitation anomalies from 1980 to 2020. Black boxes in panel c denote hotspot regions in which a strong positive trend appears. The shaded area indicates that the linear trend value exceeds the 95% confidence level, as determined by a t-test. The map was generated using the Basemap Toolkit (version 1.2.0; https://matplotlib.org/basemap/).

Among all the timescales considered, the high-frequency precipitation anomalies with periods shorter than 10 days were mostly responsible for the positive trend in the AGMT (Fig. 2a,b), whereas the other temporal scales were found to exert negligible contributions. This clearly demonstrates that the emerging climate-change signal in the observed daily precipitation fields is mostly included in the high-frequency weather components rather than the low-frequency components or changes in the long-term mean states. The dominant role of high-frequency precipitation anomalies in yielding a positive AGMT trend was found regardless of the precipitation input dataset used, whereas the predicted global-warming trends differ slightly between precipitation products. This inter-dataset difference is presumably because of the uncertainties associated with the retrieval algorithms40 or the forecast model used in the production of the datasets41, especially over the ocean, for which direct observations of precipitation is lacking42.

Next, to identify the spatial locations at which the high-frequency precipitation anomalies showed notable changes in association with the climate-change signal, we used a machine-learning-explainable method called occlusion sensitivity43. This method quantifies the relative importance of the input fields in deriving the machine-learning prediction. The occlusion sensitivity of an input grid box was obtained as the difference between the DD-model-predicted AGMT obtained with the original input data and the corresponding value obtained after substituting the input data over the 7 × 7 grid boxes surrounding the target grid box with zero (Methods). For each grid box, the occlusion sensitivity was calculated for all days and its linear trend during 1980–2020 was obtained to measure its contribution to the global-warming-signal detection.

The linear trend of occlusion sensitivity (Fig. 2c) highlights several hotspots in which a strong positive trend appears: the northern tropical eastern Pacific, northern South America, north Pacific, north Atlantic and Southern Ocean. Therefore, our results suggest that the positive trend in the estimated AGMT from the DD model was mainly caused by changes in high-frequency precipitation anomalies over these hotspots. These hotspot regions appear distinctly when using a different patch size for occlusion sensitivity (Extended Data Fig. 6a) or using other explainable methods, such as Shapley additive explanations (SHAP)44 or the integrated gradients45 method (Extended Data Fig. 6b,c).

The same locations appear as hotspots even when unfiltered anomalies are used (Extended Data Fig. 7a). Also, note that the positive linear trend of the occlusion sensitivity over the hotspots is prominent for satellite precipitation products for a relatively short period (that is, 2001–2020) (Extended Data Fig. 7b,c), whereas those over the equatorial Atlantic and central Africa appear only with the ERA5 and MSWEP datasets.

Physical interpretations

When the occlusion sensitivity is obtained separately for each high-frequency precipitation percentile over the hotspot regions (boxed areas in Fig. 2c), it is highest for the top and bottom percentiles and lowest at around the 55th–60th percentiles that correspond to values around zero (green lines in Fig. 3a,b). This V-shaped pattern indicates that the DD model generates higher AGMT values for strong high-frequency precipitation anomalies with either a positive or a negative sign over the eastern Pacific Intertropical Convergence Zone (ITCZ), northern South America and mid-latitude storm-track regions. Note that this nonlinear response of the AGMT to high-frequency precipitation anomalies cannot be accounted for in the ridge regression method because of its linear nature (black lines in Fig. 3a,b). Also, the probability density function (PDF) of the high-frequency precipitation anomalies over the hotspots showed a systematic shift towards the extreme percentiles in recent decades; for the top (>90th) and bottom (<10th) percentiles, the ratio of the PDFs for each decade to the reference PDFs for the whole period is smallest in the 1980s and greatest in the 2010s (Fig. 3c,d).

Fig. 3: Dominant regional precipitation characteristic changes detected by the deep-learning model.
figure 3

a,b, The occlusion sensitivity with respect to the percentile of the high-frequency (that is, 10-day high-pass-filtered) precipitation anomalies in the DD model (green), as well as the ridge regression model (black) in the eastern Pacific ITCZ (a) and mid-latitude storm tracks (b). c,d, The ratios of the PDF of the high-frequency precipitation anomalies during the 1980s, 1990s, 2000s or 2010s to that during 1980–2020 in the eastern Pacific ITCZ (c) and mid-latitude storm tracks (d) (see Methods for the detailed procedure for calculating PDF values). The eastern Pacific ITCZ and mid-latitude storm tracks are defined as the boxed areas within 20° S–20° N and poleward of 30° S and 30° N, respectively (Fig. 2c).

In synthesizing the results presented in Fig. 3, the DD model produces a robust increase in the estimated AGMT over recent decades, with more frequent extreme swings of high-frequency precipitation events over the hotspots. In other words, the DD model underscores the observed amplification of the high-frequency precipitation variability over the eastern Pacific ITCZ and mid-latitude storm-track regions as the global-warming signal. Although past climate-projection results have shown an increase in precipitation variability with warming3,4,46, researchers have not assessed whether the projected changes can be detected in observations.

The substantial precipitation variability increases over the eastern Pacific ITCZ, northern South America and mid-latitude storm tracks were confirmed by the linear trend of the standard deviation of the observed high-frequency precipitation anomalies during 1980–2010 or 2001–2020 (Fig. 4a). In both satellite observations and the reanalysis products, the increases in the high-frequency precipitation variability in time were statistically significant, beyond the 95% range of internal variability estimated from historical CESM2 LE simulations during 1850–1950.

Fig. 4: Emergent precipitation variability amplification over the eastern Pacific ITCZ and mid-latitude storm-track regions caused by global warming.
figure 4

a, Linear trend of the standard deviation (s.d.) of normalized ERA5 (red) and MSWEP (blue) high-frequency precipitation anomalies during 1980–2020 (darker bars) or 2001–2020 (lighter bars) over the eastern Pacific ITCZ (boxed area within 20° S–20° N) and mid-latitude storm-track regions (boxed area poleward of 30° S and 30° N). b, The difference in the high-frequency precipitation s.d. between 2016–2020 and 2001–2005 averaged across datasets (MSWEP, ERA5, GPCP and IMERG). c, Ratio of the linear trend of high-frequency precipitation variability during 1980–2020 to the linear trend of precipitation climatology (each divided by the 1980–1984 mean) in the eastern Pacific ITCZ, mid-latitude storm tracks and subtropics using ERA5 (red) and MSWEP (blue). The error bars in panels a and c denote a 95% confidence interval of internal variability, obtained from historical CESM2 LE simulations during 1850–1950. d,e, Difference in the high-frequency precipitation s.d. between 2016–2020 and 1980–1984 (d) and between 2001–2020 and 1960–1979 (e) using National Oceanic and Atmospheric Administration Climate Prediction Center (CPC) daily rain-gauge data. The map was generated using the Basemap Toolkit (version 1.2.0; https://matplotlib.org/basemap/).

The recent robust increase in high-frequency variability over the eastern Pacific ITCZ and mid-latitude storm tracks, as represented by the shift of the high-frequency precipitation events from the moderate to the extreme percentiles, is confirmed by means of the spatial distribution of the difference in the high-frequency precipitation variability (Fig. 4b). The increase in the high-frequency precipitation variability during 2016–2020 relative to that during 2001–2005 is prominent over the eastern Pacific ITCZ and mid-latitude storm tracks. More notably, the spatial distribution of the increase in the high-frequency precipitation variability resembles that of the linear trend of the occlusion sensitivity (Fig. 2c). This clearly demonstrates that the global-warming signal was successfully detected from the daily precipitation through the increase in the extreme swings of the precipitation events on weather timescales.

The robust high-frequency precipitation variability increases over the eastern Pacific ITCZ and mid-latitude storm tracks can be physically understood using a simple moisture budget analysis4 (Methods). The historical mean precipitation and high-frequency variability are both prominent over the eastern Pacific ITCZ, northern South America and mid-latitude storm tracks (Extended Data Fig. 8), which supports the ‘wet-gets-more-variable’ and ‘variable-gets-more-variable’ paradigms, respectively4. Even though the observed long-term trend can be obscured by the recent negative IPO event37, the mean precipitation did increase slightly over the eastern Pacific ITCZ region (Extended Data Fig. 5a), at which the amplitude of the negative IPO-related tropical SST anomalies exhibited a local minimum47. Therefore, our conclusion does not invalidate ‘warmer-gets-wetter’ and its similar paradigm for the high-frequency variability (that is, so-called warmer-gets-more-variable model).

The degree of increase in high-frequency precipitation variability in the eastern Pacific ITCZ and mid-latitude storm tracks is much greater than the corresponding changes in climatological precipitation (Fig. 4c). The high-frequency variability trend ratio (that is, high-frequency precipitation variability trend divided by the variability during a reference period) is approximately three times greater than the climatology trend ratio (that is, mean precipitation trend divided by the climatological value during a reference period) over the eastern Pacific ITCZ and mid-latitude storm-track regions in MSWEP and ERA5. Therefore, our detection method enables one to overcome the limitations associated with linear methods, which have previously underestimated the detectable influence of global warming on precipitation data by focusing on the changes in the mean states and not the higher-order moments.

Our results are further evaluated using direct precipitation measurements. Although the hotspot regions identified in our study are mostly over the ocean, the one located in the Atlantic storm track covers the eastern USA, in which a relatively large number of stations provide daily rain-gauge data48. The results from the rain-gauge data are largely consistent with those from the satellite and reanalysis precipitation datasets, indicating a robust increase in the magnitude of high-frequency precipitation variability during recent decades in the eastern USA (Fig. 4d,e and Extended Data Fig. 9a,b). On the contrary, over the western USA, which is outside the Atlantic storm-track hotspot, the high-frequency precipitation variability does not show an organized trend pattern. Also, the change of high-frequency variability is greater than the mean precipitation change only over the eastern USA (Extended Data Fig. 9c,d). This rain-gauge-based analysis increases the robustness of our main findings.

Global warming has resulted in increased high-frequency precipitation variability over the tropical and mid-latitude regions, whereas the subtropical Atlantic and southeastern Pacific show a predominance of climatological drying instead (right bars in Fig. 4c); this is in accordance with the occlusion sensitivity over the corresponding regions, which indicated positive values in the bottom percentiles and negative values in the top percentiles (Extended Data Fig. 10). These results demonstrate that the unique convolutional process with the nonlinear response function in the deep-learning model allows for the detection of the dominant regional characteristic changes among various timescales. Note that the stronger positive AGMT response in the bottom extreme percentiles than the negative response in the top percentiles results in a net negative AGMT response to the decreased high-frequency variability over the subtropical Atlantic and southeastern Pacific (Fig. 4b). Consequently, this contributes to the negative occlusion sensitivity trend with high-frequency precipitation input (Fig. 2c).

Recent advances in deep learning have led to numerous innovative applications in climate science29,49,50. Deep learning is a useful method for revealing and categorizing patterns responsible for a target climate phenomenon at various spatiotemporal scales in an automated manner by compressing global information into an abstract level through nonparametric mapping. Through extracting robust regional fingerprints of global warming concealed in the complex probability distribution of precipitation, the deep-learning model has revealed that the observed increase in daily precipitation variability is an emergent anthropogenic signal despite a short period of precipitation datasets; however, the mean state changes remain virtually undetectable, as they are hindered by the large internal day-to-day variability. This confirms that the impact of global warming is ubiquitous and detectable, even in variables associated with high natural variance.

Methods

DD model for climate-change detection

The DD model, which refers to the CNN model for detecting climate-change signals embedded in daily precipitation anomalies, comprises an input layer, five convolution layers, two pooling layers, one fully connected layer and an output layer (Extended Data Fig. 2). The size of the convolution kernel, which extracts key features from the input to produce feature maps, is 3 × 3. Spatial pooling was performed after the first two convolution processes by using 2 × 2 max pooling with a stride of 2. L2 regularization was applied to minimize overfitting24,51.

The DD model accepts gridded data of normalized daily precipitation anomalies as an input variable. These anomalies were determined by subtracting the daily climatology data for 1980–2010 and then normalizing them by dividing the longitudinally averaged standard deviation of the daily precipitation anomalies at the corresponding latitude during the same time period. The input variable has dimensions of 160 × 55 (2.5° × 2.5° resolution over 0°–400° E, 62.5° S–76.5° N). To properly consider the precipitation pattern around 360°(0°) E, the data were longitudinally extended by concatenating 0°–360° E and 360°–400° E. Through five convolutional and two max-pooling processes, the horizontal dimension of the feature map is reduced to 40 × 14. As the last convolutional layer uses 16 convolutional filters, the size of the dimension of the final feature map is 8,960 (that is, 40 × 14 × 16). Then, each element of the final feature map is connected to the first dense layer with 32 neurons and, finally, the first dense layer is connected to the second dense layer with a single neuron to output a scalar value representing the AGMT anomaly of the corresponding year. The variability of the estimated annual mean AGMT anomaly was matched to the observed data to avoid the influence of systematic differences between the training and testing samples. Note that this post-processing did not affect the detection results, as both the test statistics (that is, internal variability of the estimated AGMT) and the detection metric (that is, AGMT on any specific day) were modified to the same degree.

We generated five ensemble members with different random initial weights and defined the ensemble-averaged AGMT as the final forecast. The Xavier initialization technique was applied to initialize weights and biases52. Tangent hyperbolic and sigmoid functions were used as the activation functions for the convolutional and fully connected layers, respectively. Adam optimization was applied as the gradient-descent method and the mean absolute error was applied as the loss function53.

Natural variability estimation

The natural variability ranges of the estimated AGMT and EM days are measured using a bootstrap method. First, the AGMT and the fractional EM days (that is, the number of EM days/365) are calculated for each year using the daily precipitation output of CESM2 historical ensemble simulations for 1850–1950. Then, the 97.5th percentile values of the 8,080 total cases (that is, 101 year × 80 ensemble members) are estimated, which corresponds to the upper bound of the 95% confidence range of the natural variability. The resulting 97.5th percentile values are 0.42 °C and 10.9% for AGMT and fractional EM days, respectively.

The natural variability range of the linear trends in the EM days (Fig. 1c), the estimated AGMT (Fig. 2a,b) and the precipitation variability (Fig. 4a,c) are also defined using a bootstrap method. We first sample 20-year segments from CESM2 LE simulations during 1850–1950 with a 10-year interval in the initial year of the segments. With nine values per ensemble member, a total of 720 (9 × 80) values of the 20-year segments are obtained. Similarly, a total of 960 samples of 41-year segments are constructed with a 5-year interval. Then, linear trends of the EM days, the estimated AGMT and the precipitation variability are calculated for each 20-year or 41-year segment. Finally, the upper and lower 2.5% percentile values are defined as the 95% two-tailed confidence interval of the natural variability.

Occlusion sensitivity

Occlusion sensitivity is used to quantify the relative importance of each grid point when deriving an output variable43. The occlusion sensitivity O(t, x, y) is a three-dimensional tensor incorporating time (t), longitude (x) and latitude (y) and is calculated using the following equation:

$$O\left(t,x,y\right)=\widehat{y}-D\left[P\left(t,x,y\right)\ast Z\left(7,7\right)\right].$$

Here ⁎ is the horizontal convolution operator, D and \(\hat{y}\) denote the DD model and the estimated AGMT with the original input data P(t, x, y), respectively, and Z(7, 7) denotes 7 × 7 grid points occluding a mask with zero filling. A different patch size of 5 × 5 grid points is used for a sensitivity test. The occlusion sensitivity is plotted at the centred grid point of the corresponding grid box. To maintain the original size of the input map, the edge of the map is filled with zeros (that is, zero padding).

Ridge regression method

The ridge regression method is used to estimate the coefficients of multiple regression models in which linearly independent variables are highly correlated. The loss function of ridge regression with i samples and j regression coefficients is defined by the following equation24:

$${\rm{Loss}}=\mathop{\sum }\limits_{i=1}^{M}{\left({y}_{i}-\mathop{\sum }\limits_{j=0}^{P}{w}_{j}{x}_{i,j}\right)}^{2}+\lambda \mathop{\sum }\limits_{j=0}^{P}{w}_{j}^{2},$$

in which xi,j and yi denote the input and label data, respectively. wj indicates the regression coefficient and \(\lambda {\sum }_{j=0}^{p}{w}_{j}^{2}\) is the regularization term based on the sum of the squared regression coefficients (that is, the L2 norm). P and M denote the number of samples and the number of regression coefficients, respectively. The regularization suppresses overfitting by preventing the regression coefficient from becoming excessively large. λ is a hyperparameter that determines the penalty intensity, which is set to 0.1 after several experiments to minimize the loss values for the validation dataset.

Moisture budget equation for precipitation variability change

For timescales longer than a day, the zeroth-order balance in the moisture budget is found between precipitation (P) and vertical moisture advection4:

$${P}_{f}\approx -\,{\left\langle \omega {\partial }_{p}q\right\rangle }_{f},$$
(1)

in which ω and q are the vertical pressure velocity and specific humidity, respectively. \(\left\langle \cdot \right\rangle =\frac{1}{g}{\int }_{{P}_{s}}^{{P}_{t}}\cdot {\rm{d}}p\) denotes the vertical integral throughout the troposphere. The subscript f denotes variations at a specific timescale derived from the time filtering. Zhang et al.4 suggested that the column-integrated vertical moisture advection can be reasonably approximated as the advection of the low-tropospheric mean moisture (\({\bar{q}}_{l}\)) by mid-tropospheric vertical velocity anomalies (\({\omega }_{{m}_{f}}\)), hence:

$${P}_{f}\approx -\frac{{\omega }_{{m}_{f}}{\bar{q}}_{l}}{g},$$
(2)

in which g is the gravitational acceleration.

Equation (2) can be modified to denote the variability of precipitation and its change owing to the global warming as follows:

$$\sigma \left[{P}_{f}\right]\approx \frac{\sigma \left[{\left({\omega }_{m}\right)}_{f}\right]{\bar{q}}_{l}}{g};$$
(3)
$$\Delta \sigma \left[{P}_{f}\right]=\frac{1}{g}\left[\Delta \sigma \left[{\left({\omega }_{m}\right)}_{f}\right]{\bar{q}}_{l0}+\sigma \left[{\left({\omega }_{m0}\right)}_{f}\right]\Delta {\bar{q}}_{l}\right],$$
(4)

in which σ denotes standard deviation and Δ denotes the difference between the historical and future warming periods. The subscript 0 indicates the values from the historical period.

According to equation (4), well-known models for global precipitation change are similarly applicable for the high-frequency precipitation variability changes. The historical moisture climatology term (that is, \({\bar{q}}_{l0}\)) on the right-hand side refers to the ‘wet-gets-more-variable’ paradigm and the historical precipitation variability term (that is, \(\sigma [{({\omega }_{m0})}_{f}]\)) refers to the ‘variable-gets-more-variable’ paradigm. Given the strong coupling between the low-level moisture and the sea-surface temperature, the climatological moisture change term (that is, \(\Delta {\bar{q}}_{l}\)) presumably implies the ‘warmer-gets-more-variable’ paradigm.

Satellite and reanalysis dataset

We analysed 21 years (2001–2020) of daily mean satellite-observed precipitation data from the IMERG version 6 (ref. 30) and the GPCP version 3.2 (ref. 31). Daily gauge-satellite-reanalysis merged precipitation was obtained from the MSWEP version 2.8 for the period from 1980 to 2020 (ref. 32). Daily reanalysis precipitation data obtained from ERA5, which spans 1980–2020, were also used33. Data were interpolated to a 2.5° × 2.5° horizontal grid. Domains over 0°–360° E and 61.25° S–76.25° N were used. Daily gauge-based precipitation at the horizontal resolution is 0.25° × 0.25° from National Oceanic and Atmospheric Administration (NOAA) CPC from 1960 to 2020 (ref. 48) was used for a domain over the continental USA (126.25°–67.25° W, 20°–49.5° N). The AGMT was obtained from HadCRUT5 data54.

The PDF of the daily precipitation anomalies with respect to its percentile is calculated for each decade. After arranging daily precipitation anomalies over certain regions during the whole period (that is, 1980–2020) by their magnitudes, the values of precipitation for every tenth percentile are defined. The PDF of precipitation anomalies for each decade were calculated in the same way and then compared with the reference PDF value estimated by using the whole period for each percentile (Fig. 3).

CESM2 LE simulations

To train the DD model, we used a climate model dataset from the CESM2 LE, which has state-of-the-art skills in simulating characteristics of the daily precipitation at various timescales6. All the ensemble members that provide daily precipitation output were used (that is, 80 ensemble members). With the aid of tens of realizations for historical and global-warming-scenario simulations, the total number of samples used in training our DD model is larger than what any other model simulation framework can provide, which is advantageous for training the deep-learning model. The simulations cover the period from 1850 to 2100, of which data from 1850 to 2014 were obtained from the historical simulations and the rest from the SSP3-7.0 scenario simulations. A domain over 0°–360° E and 61.25° S–76.25° N was used and the horizontal resolution was coarsened to 2.5° × 2.5°. The input data were prescribed in the form of a normalized anomaly; the modelled daily climatology from 1980 to 2010 was subtracted from the raw precipitation fields and then divided by longitudinally averaged standard deviation at the corresponding latitude during the same period.

Because the total number of samples was 7,329,200 days (80 members × 251 years × 365 days), which exceeded the limit of our computing resources, we subsampled the training and validation datasets by randomly selecting one year from each decade. Thus, the total number of days of training data was reduced to 730,000. For the validation dataset, we randomly selected a different year from each decade and then randomly selected 73 days from each selected year. The total number of days of validation data used was 146,000.