Abstract
We introduce the Precipitation Probability DISTribution (PPDIST) dataset, a collection of global high-resolution (0.1°) observation-based climatologies (1979–2018) of the occurrence and peak intensity of precipitation (P) at daily and 3-hourly time-scales. The climatologies were produced using neural networks trained with daily P observations from 93,138 gauges and hourly P observations (resampled to 3-hourly) from 11,881 gauges worldwide. Mean validation coefficient of determination (R2) values ranged from 0.76 to 0.80 for the daily P occurrence indices, and from 0.44 to 0.84 for the daily peak P intensity indices. The neural networks performed significantly better than current state-of-the-art reanalysis (ERA5) and satellite (IMERG) products for all P indices. Using a 0.1 mm 3 h−1 threshold, P was estimated to occur 12.2%, 7.4%, and 14.3% of the time, on average, over the global, land, and ocean domains, respectively. The highest P intensities were found over parts of Central America, India, and Southeast Asia, along the western equatorial coast of Africa, and in the intertropical convergence zone. The PPDIST dataset is available via www.gloh2o.org/ppdist.
Measurement(s) | hydrological precipitation process • precipitation occurrence • precipitation intensity |
Technology Type(s) | Neural networks models • Gauge or Meter Device • satellite imaging of a planet • reanalyses |
Factor Type(s) | daily and 3-hourly time-scales • year of data collection |
Sample Characteristic - Environment | climate system |
Sample Characteristic - Location | Earth (planet) |
Machine-accessible metadata file describing the reported data: https://doi.org/10.6084/m9.figshare.12815585
Background & Summary
Knowledge about the climatological probability distribution of precipitation (P) is essential for a wide range of scientific, engineering, and operational applications1,2,3. In the scientific realm, P distributional information is used, for example, to evaluate P outputs from climate models (e.g., Dai4 and Bosilovich et al.5) and to reduce systematic biases in gridded P datasets (e.g., Zhu and Luo6, Xie et al.7, and Karbalaee et al.8). In the engineering sector, information on the upper tail of the P distribution is crucial for planning, design, and management of infrastructure (e.g., Lumbroso et al.9 and Yan et al.10). In an operational context, reliable information on the P distribution can be used to improve flood and drought forecasting systems (e.g., Cloke and Pappenberger11, Hirpa et al.12, and Siegmund et al.13).
Past studies deriving (quasi-)global climatologies of P distributional characteristics typically relied on just one of the three main sources of P data (satellites, reanalyses, or gauges), each with strengths and weaknesses:
-
1.
Ricko et al.14 used the TMPA satellite P product15 while Trenberth and Zhang16 and Li et al.17 used the CMORPH satellite P product7,18 (both available from 2000 onwards) to estimate climatological P characteristics. Satellite P estimates have several advantages, including a high spatial (~0.1°) and temporal (≤3-hourly) resolution, and near-global coverage (typically <60°N/S)19,20. However, satellite P retrieval is confounded by complex terrain21,22,23 and surface snow and ice22,24, and there are major challenges associated with the detection of snowfall25,26.
-
2.
Courty et al.27 used P estimates between 1979 and 2018 from the ERA5 reanalysis28 to derive probability distribution parameters of annual P maxima for the entire globe. Reanalyses assimilate vast amounts of in situ and satellite observations into numerical weather prediction models29 to produce a temporally and spatially consistent continuous record of the state of the atmosphere, ocean, and land surface. However, reanalyses are affected by deficiencies in model structure and parameterization, uncertainties in observation error statistics, and historical changes in observing systems5,30,31.
-
3.
Sun et al.32 calculated P occurrence estimates directly from gauge observations, while Dietzsch et al.33 used the gauge-based interpolated GPCC dataset34 to estimate P characteristics for the entire land surface. Gauges provide accurate estimates of P at point scales and therefore have been widely used to evaluate satellite- and reanalysis-based P products (e.g., Hirpa et al.35, Zambrano-Bigiarini et al.36, and Beck et al.37). However, only 16% of the global land surface (excluding Antarctica) has a gauge located at <25-km distance38,39, gauge placement is topographically biased towards low elevations40,41, and there has been a significant decline in the number of active gauges over recent decades42. Additionally, interpolation of gauge observations generally leads to systematic overestimation of low intensity events and underestimation of high intensity events43,44.
Here, we introduce the Precipitation Probability DISTribution (PPDIST) dataset, a collection of high-resolution (0.1°) fully global climatologies (1979–2018) of eight P occurrence indices (using different thresholds to discriminate between P and no-P) and nine peak P intensity indices (using different return periods) at daily and 3-hourly time-scales (Figs. 1b and 2b). The climatologies were produced using neural network ensembles trained to estimate the P indices using quality-controlled P observations from an unprecedented database consisting of 93,138 daily and 11,881 hourly gauges worldwide (Fig. 1a and Supplement Fig. S2a). Ten global P-, climate-, and topography-related predictors derived from all three main P data sources — satellites, reanalyses, and gauges — were used as input to the neural networks to enhance the accuracy of the estimates (Table 1). Ten-fold cross-validation was used to obtain an ensemble of neural networks, evaluate the generalizability of the approach, and quantify the uncertainty. The performance was assessed using the coefficient of determination (R2), the mean absolute error (MAE), and the bias (B; Fig. 3). The neural networks yielded mean validation R2 values (calculated using the validation subsets of observations) of 0.78 and 0.86 for the daily and 3-hourly P occurrence indices, respectively, and 0.72 and 0.83 for the daily and 3-hourly peak P intensity indices, respectively (Fig. 3 and Supplement Fig. S4). The neural networks outperformed current state-of-the-art reanalysis (ERA528) and satellite (IMERG45,46) products for all P indices and performance metrics (Fig. 3). ERA5 consistently overestimated the P occurrence (Fig. 1c), while IMERG tended to overestimate the peak P intensity (Fig. 2d). The uncertainty in the neural network-based estimates of the P indices was quantified globally and was generally higher in sparsely gauged, mountainous, and tropical regions (Fig. 4). Overall, these results suggest that the neural networks provide a useful high-resolution, global-scale picture of the P probability distribution. It should be noted that although the climatologies are provided as gridded maps, the estimates represent the point scale rather than the grid-box scale, due to the use of gauge observations for the training43,47.
The >0.5 mm d−1 daily P occurrence according to (a) the gauge observations, (b) the PPDIST dataset, (c) the ERA5 reanalysis, and (d) the IMERG satellite product. Supplement Fig. S2 presents an equivalent figure for the >0.1 mm 3 h−1 3-hourly P occurrence. The other PPDIST P occurrence indices can be viewed by accessing the dataset.
The 15-year return-period daily P intensity according to (a) the gauge observations, (b) the PPDIST dataset, (c) the ERA5 reanalysis, and (d) the IMERG satellite product. IMERG has data gaps at high latitudes (>60°N/S) precluding the calculation of the 15-year return-period daily P intensity. Supplement Fig. S3 presents an equivalent figure for the 15-year return-period 3-hourly P intensity. The other PPDIST P intensity indices can be viewed by accessing the dataset.
Performance of the PPDIST dataset, the ERA5 reanalysis, and the IMERG satellite product in estimating the daily P occurrence and peak P intensity indices. For the PPDIST dataset, we calculated the mean of the ten validation scores (one for each cross-validation iteration). The training scores were not shown as they were nearly identical to the validation scores. All scores were calculated using square-root-transformed observed and estimated values. Supplement Fig. S4 presents an equivalent figure for the 3-hourly P indices.
PPDIST uncertainty estimates for (a) the >0.5 mm d−1 daily P occurrence and (b) the 15-year return-period daily P intensity. The uncertainty represents the spread of the ten cross-validation iterations. Supplement Fig. S5 presents an equivalent figure for the >0.1 mm 3 h−1 P occurrence and the 15-year return-period 3-hourly P intensity.
Since the eight PPDIST P occurrence climatologies were similar, we only present and discuss results for the >0.5 mm d−1 daily P occurrence climatology (Fig. 1b), although the others can be viewed by accessing the dataset. The P occurrence was particularly high (>0.5 mm d−1 on >50% of days) in windward coastal areas of Chile, Colombia, the British Isles, and Norway, along the Pacific coast of North America, and over parts of Southeast Asia (Fig. 1b)16,48,49,50,51. Conversely, the P occurrence was particularly low (>0.5 mm d−1 on <1% of days) in the Atacama and Sahara deserts. At a daily time-scale using a 0.5 mm d−1 threshold, P was found to occur 34.4%, 24.9%, and 38.9% of days, on average, over the global, land, and ocean domains, respectively (Fig. 1b). At a 3-hourly time-scale using a 0.1 mm 3 h−1 threshold, P was found to occur 12.2%, 7.4%, and 14.3% of the time, on average, over the global, land, and ocean domains, respectively (Supplement Fig. S2b). These estimates are in line with previous estimates derived from the satellite-based CMORPH product16 and from the satellite-, reanalysis-, and gauge-based MSWEP product52. The higher average P occurrence over ocean than over land reflects the unlimited water supply53. The PPDIST > 0.5 mm d−1 daily P occurrence climatology (Fig. 1b) exhibits reasonable visual agreement with similar climatologies from previous studies despite the use of different data sources, spatiotemporal resolutions, and thresholds to discriminate between P and no-P (Sun et al.32, their Fig. 1; Ellis et al.54, their Fig. 3; Trenberth and Zhang16, their Fig. 4; and Beck et al.52, their Fig. 8).
Since the nine PPDIST peak P intensity climatologies were similar, we only present and discuss results for the 15-year return-period daily P climatology (Fig. 2b). Peak P intensities were markedly higher at low latitudes (<30°N/S; Fig. 2b), due to the higher air temperatures which lead to more intense P in accordance with the Clausius-Clapeyron relationship55,56,57. The highest 15-year return-period P intensities (>200 mm d−1) over land were found in parts of Central America, India, and Southeast Asia, along the western equatorial coast of Africa, and over oceans in the intertropical convergence zone (Fig. 2b)17,48,58,59. Conversely, the lowest 15-year return-period P intensities (<50 mm d−1) were found in the desert regions of the world, at high latitudes where frontal P dominates, and over southeastern parts of the Atlantic and Pacific oceans where semi-permanent anticyclones prevail54,60. The PPDIST 15-year return-period daily P climatology (Fig. 2b) exhibits reasonable visual agreement with a maximum one-day P climatology derived by Dietzsch et al.33 (their Fig. 5d) from the gauge-based GPCC Full Data Daily dataset34, as well as with a 3.4-year return-period 3-hourly P climatology derived by Beck et al.52 (their Fig. 7) from the satellite-, reanalysis-, and gauge-based MSWEP dataset52.
Methods
Gauge observations and quality control
Our initial database of daily P gauge observations comprised 149,095 gauges worldwide and was compiled from (i) the Global Historical Climatology Network-Daily (GHCN-D) dataset (ftp.ncdc.noaa.gov/pub/data/ghcn/daily/)61 (102,954 gauges), (ii) the Global Summary Of the Day (GSOD) dataset (https://data.noaa.gov; 24,730 gauges), (iii) the Latin American Climate Assessment & Dataset (LACA&D) dataset (http://lacad.ciifen.org/; 231 gauges), (iv) the Chile Climate Data Library (www.climatedatalibrary.cl; 716 gauges), and (v) national datasets for Brazil (11,678 gauges; www.snirh.gov.br/hidroweb/apresentacao), Mexico (5398 gauges), Peru (255 gauges), and Iran (3133 gauges). Our initial database of hourly P gauge observations (resampled to 3-hourly) comprised 14,350 gauges from the Global Sub-Daily Rainfall (GSDR) dataset62 produced as part of the INTElligent use of climate models for adaptatioN to non-Stationary hydrological Extremes (INTENSE) project63.
Gauge data can have considerable measurement errors and therefore quality control is important64,65,66. To eliminate long series of erroneous zeros frequently present in daily GSOD records67,68, we calculated a central moving mean with a one-year window (assigning a value only if at least half a year of values were present), and discarded daily and 3-hourly P observations with a coincident moving mean of zero. To eliminate long series of P without zeros, indicative of biased reporting practices or quality control issues, we calculated a central moving minimum with a one-year window (assigning a value only if at least half a year of values were present), and discarded daily and 3-hourly P observations with a coincident moving minimum greater than zero. We only used observations between January 1st, 1979, and December 31st, 2018, and discarded gauges with (i) < 3 years of remaining data, (ii) the largest daily P > 2000 mm d−1 or 3-hourly P > 500 mm 3 h−1 (approximately the maximum recorded 24-h and 3-h P, respectively69), (iii) the highest P value < 2 mm d−1 or <1 mm 3 h−1, or (iv) the lowest P value > 1 mm d−1 or >0.5 mm 3 h−1. The resulting database comprised 93,138 gauges with daily data (Fig. 1a) and 11,881 gauges with 3-hourly data (Supplement Fig. S2a). See Supplement Fig. S1 for record lengths of the daily and 3-hourly gauge data.
Precipitation occurrence and peak intensity indices
We considered eight P occurrence indices (related to the lower tail of the P distribution) and nine peak P intensity indices (related to the upper tail of the P distribution), all calculated for both the daily and 3-hourly time-scale. The eight P occurrence indices measure the percentage of time with P using thresholds of 0.1, 0.15, 0.2, 0.25, 0.35, 0.5, 0.75, and 1 mm d−1 or mm 3 h−1 to discriminate between P and no-P. The P occurrence was calculated using the percentileofscore function of the scipy Python module70. We considered multiple thresholds because (i) different researchers and meteorological agencies adopt different definitions of “wet day” and “rain day”, (ii) the most suitable threshold depends on the application, and (iii) the estimation accuracy depends on the threshold16,71.
The nine peak P intensity indices measure the magnitude of daily or 3-hourly P events for return periods of 3 days, 7 days, 1 month, 3 months, 1 year, 3 years, 5 years, 10 years, and 15 years. The magnitudes were calculated using the percentile function of the numpy Python module72,73. Percentiles were calculated according to p = 100 − 100/(365.25 · R) for the daily time-scale and p = 100 − 100/(365.25 · 8 · R) for the 3-hourly time-scale, where p is the percentile (%) and R is the return period (years). Gauges with <3 years of data were discarded per the preceding subsection. However, we only calculated the three highest return periods if the record lengths were longer than 5, 10, and 15 years, respectively. We did not consider return periods >15 years because (i) the number of observations would significantly decrease, and (ii) the IMERG dataset, which was used as predictor, spans only 20 years.
Neural network ensembles
We used feed-forward artificial neural networks based on the multilayer perceptron74 to produce global maps of the P indices. Neural networks are models composed of interconnected neurons able to model complex non-linear relationships between inputs and outputs. Neural networks have been successfully applied in numerous studies to estimate different aspects of P (e.g., Coulibaly et al.75, Kim and Pachepsky76, and Nastos et al.77). We used a neural network with one hidden layer composed of 10 nodes with the logistic sigmoid activation function. The stochastic gradient descent method was used to train the weights. To make the training more efficient, the inputs (i.e., the predictors; Table 1) were standardized by subtracting the means and dividing by the standard deviations of the global predictor maps. Additionally, the outputs (i.e., the P indices) were square-root transformed to reduce the skewness of the distributions, and standardized using means and standard deviations of the observed P indices.
To obtain an ensemble of neural networks, evaluate the generalizability of the approach, and quantify the uncertainty, we trained neural networks for each P index using ten-fold cross-validation. For each cross-validation iteration, the observations were partitioned into subsets of 90% for training and 10% for validation. The partitioning was random, although each observation was used only once for validation. Since the validation subsets are completely independent and were not used to train the neural networks, they allow quantification of the performance at ungauged locations.
We subsequently applied the trained neural networks using global 0.1° maps of the predictors, yielding an ensemble of ten global maps for each P index (one for each cross-validation iteration). The values were destandardized by multiplying by the standard deviation and adding the mean of observed P indices, and back-transformed by squaring the values, after which we used the mean of the ensemble to provide a ‘best estimate’ and the standard deviation to provide an indication of the uncertainty. As a last step, we sorted the eight mean P occurrence estimates for each 0.1° grid cell to ensure that they continuously decrease as a function of the thresholds, and sorted the nine mean peak P intensity estimates to ensure that they continuously increase as a function of the percentiles. The uncertainty estimates were re-ordered according to the mean estimates.
Predictors
Table 1 presents the predictors used as input to the neural networks. The predictors were upscaled from their native resolution to 0.1° using average resampling. To highlight broad-scale topographic features, the elevation predictor (Elev) was smoothed using a Gaussian kernel with a 10-km radius following earlier studies78,79. The Elev predictor was square-root transformed to increase the importance of topographic features at low elevation80. The mean annual air temperature predictor (MAT) was included because higher air temperatures are generally associated with higher P intensities56,81.
Mean annual P predictors (MAP1 and MAP2) were included due to the high correlations found between mean and extreme values of P14,82. To account for the uncertainty in mean annual P estimates, we considered two state-of-the-art climatologies: CHPclim V183 and WorldClim V284. Since the use of correlated predictors can increase training time85, we did not directly use both climatologies as separate predictors. Instead, we used square-root transformed mean annual P from WorldClim as predictor MAP1, and the difference in square-root transformed mean annual P between CHPclim and WorldClim as predictor MAP2. The square-root transformation was used to increase the importance of smaller values.
The ERA5 and IMERG predictors represent maps of the P index subject of the estimation derived from ERA5 and IMERG P data, respectively. The ERA5 and IMERG predictors were added because they provide valuable independent information about the P distribution with (near-)global coverage. All predictors cover the entire land surface, with the exception of IMERG, which was often not available at latitudes higher than approximately 60°N/S46. This issue was addressed by replacing missing IMERG estimates of the P indices with equivalent estimates from ERA5, resulting in a duplication of ERA5 at high latitudes. P distributional characteristics associated with different durations often exhibit similar spatial patterns27,86. To allow the 3-hourly estimates to take advantage of the much better availability of daily P observations (Figs. 1a and 2a), we used the daily climatology of the P index subject of the estimation as additional predictor for producing the 3-hourly climatology.
Performance metrics
We assessed the performance of the PPDIST dataset, the ERA5 reanalysis, and the IMERG satellite product in estimating the P indices using the gauge observations as reference. Prior to the assessment, the P indices of the products and the reference were square-root transformed to reduce the skewness of the distributions. For the PPDIST dataset, we calculated the average of the ten scores calculated using the ten independent validation subsets of observations (one for each cross-validation iteration). For ERA5 and IMERG, we simply calculated the scores using all available observations. The following three performance metrics were used:
1. The coefficient of determination (R2), which measures the proportion of variance accounted for87, ranges from 0 to 1, and has its optimum at 1.
2. The mean absolute error (MAE), calculated as:
where \({X}_{i}^{e}\) and \({X}_{i}^{o}\) represent estimated and observed values of the transformed P indices, respectively, i = 1, …, n denotes the observations, and n is the sample size. The MAE ranges from 0 to ∞ and has its optimum at 0. We used the MAE, instead of the more common root mean square error (RMSE), because the errors are unlikely to follow a normal distribution88,89.
3. The bias (B), defined as:
B ranges from −1 to 1 and has its optimum at 0. Other widely used bias formulations, which contain the sum of only the observations in the denominator rather than the sum of both the observations and estimates90,91, were not used, because they yield unwieldy high B values when the observations tend to zero.
Data Records
The PPDIST dataset is available for download at figshare92 and www.gloh2o.org/ppdist. The data are provided as a zip file containing four netCDF-4 files (www.unidata.ucar.edu/software/netcdf) with gridded observations, mean estimates, and uncertainty estimates of (i) the daily P occurrence (daily_occurrence_point_scale.nc), (ii) the daily peak P intensity (daily_intensity_point_scale.nc), (iii) the 3-hourly P occurrence (3hourly_occurrence_point_scale.nc), and (iv) the 3-hourly peak P intensity (3hourly_intensity_point_scale.nc). The grids have a 0.1° spatial resolution and are referenced to the World Geodetic Reference System 1984 (WGS 84) ellipsoid. The observational grids were produced by averaging the observations if multiple observations were present in a single grid cell. The netCDF-4 files can be viewed, edited, and analyzed using most Geographic Information Systems (GIS) software packages, including ArcGIS, QGIS, and GRASS.
Technical Validation
Figure 3 presents the performance of the PPDIST dataset, the ERA5 reanalysis, and the IMERG satellite product in estimating the daily P indices (Supplement Fig. S4 presents the performance in estimating the 3-hourly P indices). The PPDIST scores represent averages calculated from the ten independent validation subsets of observations (one for each cross-validation iteration). The PPDIST R2 values ranged from 0.76 to 0.80 (mean 0.78) for the daily P occurrence indices, 0.85 to 0.88 (mean 0.86) for the 3-hourly P occurrence indices, 0.44 to 0.84 (mean 0.72) for the daily peak P intensity indices, and 0.72 to 0.88 (mean 0.83) for the 3-hourly peak P intensity indices (Fig. 3 and Supplement Fig. S4). The 3-hourly R2 values were thus higher than the daily R2 values, likely because the 3-hourly gauge observations tend to be of higher quality62. The markedly lower R2 for the 3-day return-period daily P intensity reflects the fact that the majority (58%) of the observations were 0 mm d−1. For the ≥3-year return-period daily and 3-hourly P indices, the lower performance reflects the uncertainty associated with estimating the magnitude of heavy P events93,94,95. The PPDIST scores were superior to the ERA5 and IMERG scores for every P index and performance metric by a substantial margin. The PPDIST B scores were close to 0 for all P indices, indicating that the estimates do not exhibit systematic biases, apart from those already present in the gauge data due to wind-induced under-catch64,96 and the underreporting of light P events65. Overall, these results suggest that the PPDIST dataset provides a reliable global-scale picture of the P probability distribution.
The ERA5 and IMERG products yielded mean R2 values of 0.59 and 0.58 for the daily P occurrence indices, 0.69 and 0.49 for the 3-hourly P occurrence indices, 0.65 and 0.56 for the daily P intensity indices, and 0.79 and 0.57 for the 3-hourly P intensity indices, respectively (Fig. 3 and Supplement Fig. S4). The global patterns of P occurrence and peak P intensity were thus generally better estimated by ERA5 than by IMERG. However, ERA5 consistently overestimated the P occurrence and the lowest return periods, in line with evaluations of other climate models30,37,97. Additionally, ERA5 generally showed lower values for the highest return periods (Fig. 3), which is at least partly due to the spatial scale mismatch between point-scale observations and grid-box averages43,47,98,99. Conversely, ERA5 strongly overestimated the P intensity over several grid cells in eastern Africa in particular (Fig. 2c), likely due to the presence of so-called “rain bombs” caused by numerical artifacts100.
The overly high P occurrence of IMERG in the tropics (Fig. 1d) may be at least partly attributable to the inclusion of relatively noisy P estimates from the SAPHIR sensor on board the tropical-orbiting Megha-Tropiques satellite46. Similar to the TMPA 3B42 and CMORPH satellite products, IMERG overestimated the P occurrence over several large rivers (notably the Amazon, the Lena, and the Ob), likely due to misinterpretation of the emissivity of small inland water bodies as P by the passive microwave P retrieval algorithms101. The too low P occurrence at high latitudes reflects the challenges associated with the retrieval of snowfall26,102 and light P103,104. Furthermore, IMERG tended to underestimate the P intensity over regions of complex topography (e.g., over Japan and the west coast of India; Fig. 2d), which is a known limitation of the product23,105. The too high P intensity in the south central US may stem from the way P estimates from the various passive microwave sensors are intercalibrated to the radar-based CORRA product in the IMERG algorithm46.
Usage Notes
The PPDIST dataset will be useful for numerous purposes, such as the evaluation of P from climate models (e.g., Dai4 and Bosilovich et al5.), the distributional correction of gridded P datasets (e.g., Zhu and Luo6, Xie et al.7, and Karbalaee et al.8), the design of infrastructure in poorly gauged regions (e.g., Lumbroso et al.9), and in hydrological modelling where P intensity is required (e.g., Donohue et al.106 and Liu et al.107). However, some caveats should be kept in mind:
-
1.
The uncertainty estimates included in the PPDIST dataset provide an indication of the reliability of the climatologies. The uncertainty is generally higher in sparsely gauged, mountainous, and tropical regions (Fig. 4). The uncertainty estimates may, however, be on the low end, because while we accounted for uncertainty in the mean annual P predictor, we did not account for uncertainty in the other predictors.
-
2.
The P occurrence estimates using low thresholds to discriminate between P and no-P (≤0.25 mm d−1 or mm 3 h−1) should be interpreted with caution due to the detection limits of gauges (~0.25 mm for the US National Weather Service [NWS] 8” standard gauge108) and satellite sensors (~0.8 mm h−1 for the TRMM microwave imager [TMI]109 and possibly higher for other sensors with coarser footprints) and the underreporting of light P events65.
-
3.
P observations from unshielded gauges (e.g., the US NWS 8” standard gauge) are subject to wind-induced gauge under-catch and therefore underestimate rainfall by 5–15% and snowfall by 20–80%64,96. This underestimation is reflected in the PPDIST estimates. The Precipitation Bias CORrection (PBCOR) dataset110 (www.gloh2o.org/pbcor) provides a means to ameliorate this bias to some degree.
-
4.
The PPDIST climatologies were derived from gauge observations and therefore represent the point scale. Compared to point-scale estimates, grid-box averages tend to exhibit more frequent P of lower intensity16,43,47,99. Point-scale P occurrence estimates can be transformed to grid-box averages using, for example, the equations of Osborn and Hulme111, whereas the peak P intensity estimates can be transformed to grid-box averages using, for example, the areal reduction factors collated by Pietersen et al.112.
-
5.
As mentioned before, ERA5 strongly overestimates the P intensity over several grid cells in eastern Africa in particular (Fig. 2c). Since ERA5 was used as predictor, these artifacts are also present, albeit less pronounced, in the PPDIST P intensity estimates (Fig. 2b).
-
6.
Some of the PPDIST climatologies (e.g., the 15-year return-period daily P climatology; Fig. 2b) exhibit a longitudinal discontinuity at 60°N/S, reflecting the spatial extent of the IMERG data (Table 1). Additionally, some of the climatologies exhibit discontinuities at coastlines (e.g., along the west coast of Patagonia), due to the use of different data sources for land and ocean for the MAP1, MAP2, MAT, and SnowFrac predictors.
Code availability
The neural networks used to produce the PPDIST dataset were implemented using the MLPRegressor class of the scikit-learn Python module113. The P occurrence for different thresholds was calculated using the percentileofscore function of the scipy Python module70, whereas the P magnitudes for different return periods were calculated using the percentile function of the numpy Python module72,73. The other codes are available upon request from the first author. The predictor, IMERG, and ERA5 data are available via the URLs listed in Table 1. Most of the gauge observations are available via the URLs provided in the “Gauge observations and quality control” subsection. Part of the GSDR database and some of the national databases are only available upon request.
References
Tapiador, F. J. et al. Global precipitation measurement: Methods, datasets and applications. Atmospheric Research 104–105, 70–97 (2012).
Kucera, P. A. et al. Precipitation from space: Advancing Earth system science. Bulletin of the American Meteorological Society 94, 365–375 (2013).
Kirschbaum, D. B. et al. NASA’s remotely sensed precipitation: A reservoir for applications users. Bulletin of the American Meteorological Society 98, 1169–1184 (2017).
Dai, A. Precipitation characteristics in eighteen coupled climate models. Journal of Climate 19, 4605–4630 (2006).
Bosilovich, M. G., Chen, J., Robertson, F. R. & Adler, R. F. Evaluation of global precipitation in reanalyses. Journal of Applied Meteorology and Climatology 47, 2279–2299 (2008).
Zhu, Y. & Luo, Y. Precipitation calibration based on the frequency-matching method. Weather and Forecasting 30, 1109–1124 (2015).
Xie, P. et al. Reprocessed, bias-corrected CMORPH global high-resolution precipitation estimates from 1998. Journal of Hydrometeorology 18, 1617–1641 (2017).
Karbalaee, N., Hsu, K., Sorooshian, S. & Braithwaite, D. Bias adjustment of infrared based rainfall estimation using passive microwave satellite rainfall data. Journal of Geophysical Research: Atmospheres 122, 3859–3876 (2017).
Lumbroso, D. M., Boyce, S., Bast, H. & Walmsley, N. The challenges of developing rainfall intensity-duration-frequency curves and national flood hazard maps for the Caribbean. Journal of Flood Risk Management 4, 42–52 (2011).
Yan, H. et al. Next-generation intensity-duration-frequency curves to reduce errors in peak flood design. Journal of Hydrologic Engineering 24, 04019020 (2019).
Cloke, H. L. & Pappenberger, F. Ensemble flood forecasting: A review. Journal of Hydrology 375, 613–626 (2009).
Hirpa, F. A. et al. The effect of reference climatology on global flood forecasting. Journal of Hydrometeorology 17, 1131–1145 (2016).
Siegmund, J., Bliefernicht, J., Laux, P. & Kunstmann, H. Toward a seasonal precipitation prediction system for West Africa: Performance of CFSv2 and high-resolution dynamical downscaling. Journal of Geophysical Research: Atmospheres 120, 7316–7339 (2015).
Ricko, M., Adler, R. F. & Huffman, G. J. Climatology and interannual variability of quasi-global intense precipitation using satellite observations. Journal of Climate 29, 5447–5468 (2016).
Huffman, G. J. et al. The TRMM Multisatellite Precipitation Analysis (TMPA): quasi-global, multiyear, combined-sensor precipitation estimates at fine scales. Journal of Hydrometeorology 8, 38–55 (2007).
Trenberth, K. E. & Zhang, Y. How often does it really rain? Bulletin of the American Meteorological Society 99, 289–298 (2018).
Li, X.-F. et al. Global distribution of the intensity and frequency of hourly precipitation and their responses to ENSO. Climate Dynamics 1–17 (2020).
Joyce, R. J., Janowiak, J. E., Arkin, P. A. & Xi, P. CMORPH: A method that produces global precipitation estimates from passive microwave and infrared data at high spatial and temporal resolution. Journal of Hydrometeorology 5, 487–503 (2004).
Stephens, G. L. & Kummerow, C. D. The remote sensing of clouds and precipitation from space: a review. Journal of the Atmospheric Sciences 64, 3742–3765 (2007).
Sun, Q. et al. A review of global precipitation datasets: data sources, estimation, and intercomparisons. Reviews of Geophysics 56, 79–107 (2018).
Prakash, S. et al. A preliminary assessment of GPM-based multi-satellite precipitation estimates over a monsoon dominated region. Journal of Hydrology 556, 865–876 (2018).
Cao, Q., Painter, T. H., Currier, W. R., Lundquist, J. D. & Lettenmaier, D. P. Estimation of precipitation over the OLYMPEX domain during winter 2015/16. Journal of Hydrometeorology 19, 143–160 (2018).
Beck, H. E. et al. Daily evaluation of 26 precipitation datasets using Stage-IV gauge-radar data for the CONUS. Hydrology and Earth System Sciences 23, 207–224 (2019).
Kidd, C. et al. Intercomparison of high-resolution precipitation products over northwest Europe. Journal of Hydrometeorology 13, 67–83 (2012).
Levizzani, V., Laviola, S. & Cattani, E. Detection and measurement of snowfall from space. Remote Sensing 3, 145–166 (2011).
Skofronick-Jackson, G. et al. Global Precipitation Measurement Cold Season Precipitation Experiment (GCPEX): for measurement’s sake, let it snow. Bulletin of the American Meteorological Society 96, 1719–1741 (2015).
Courty, L. G., Wilby, R. L., Hillier, J. K. & Slater, L. J. Intensity-duration-frequency curves at the global scale. Environmental Research Letters 14, 084045 (2019).
Hersbach, H. et al. The ERA5 global reanalysis. Quarterly Journal of the Royal Meteorological Society (2020).
Bauer, P., Thorpe, A. & Brunet, G. The quiet revolution of numerical weather prediction. Nature 525, 47–55 (2015).
Stephens, G. L. et al. Dreary state of precipitation in global models. Journal of Geophysical Research: Atmospheres 115 (2010).
Kang, S. & Ahn, J.-B. Global energy and water balances in the latest reanalyses. Asia-Pacific Journal of Atmospheric Sciences 51, 293–302 (2015).
Sun, Y., Solomon, S., Dai, A. & Portmann, R. W. How often does it rain? Journal of Climate 19, 916–934 (2006).
Dietzsch, F. et al. A global ETCCDI-based precipitation climatology from satellite and rain gauge measurements. Climate 5 (2017).
Schamm, K. et al. Global gridded precipitation over land: a description of the new GPCC First Guess Daily product. Earth System Science Data 6, 49–60 (2014).
Hirpa, F. A., Gebremichael, M. & Hopson, T. Evaluation of high-resolution satellite precipitation products over very complex terrain in Ethiopia. Journal of Applied Meteorology and Climatology 49, 1044–1051 (2010).
Zambrano-Bigiarini, M., Nauditt, A., Birkel, C., Verbist, K. & Ribbe, L. Temporal and spatial evaluation of satellite-based rainfall estimates across the complex topographical and climatic gradients of Chile. Hydrology and Earth System Sciences 21, 1295–1320 (2017).
Beck, H. E. et al. Global-scale evaluation of 22 precipitation datasets using gauge observations and hydrological modeling. Hydrology and Earth System Sciences 21, 6201–6217 (2017).
Beck, H. E. et al. MSWEP: 3-hourly 0.25° global gridded precipitation (1979–2015) by merging gauge, satellite, and reanalysis data. Hydrology and Earth System Sciences 21, 589–615 (2017).
Kidd, C. et al. So, how much of the Earth’s surface is covered by rain gauges? Bulletin of the American Meteorological Society 98, 69–78 (2017).
Briggs, P. R. & Cogley, J. G. Topographic bias in mesoscale precipitation networks. Journal of Climate 9, 205–218 (1996).
Schneider, U. et al. GPCC’s new land surface precipitation climatology based on quality-controlled in situ data and its role in quantifying the global water cycle. Theoretical and Applied Climatology 115, 15–40 (2014).
Mishra, A. K. & Coulibaly, P. Developments in hydrometric network design: A review. Reviews of Geophysics 47 (2009).
Ensor, L. A. & Robeson, S. M. Statistical characteristics of daily precipitation: comparisons of gridded and point datasets. Journal of Applied Meteorology and Climatology 47, 2468–2476 (2008).
Hofstra, N., Haylock, M., New, M. & Jones, P. D. Testing E-OBS European high-resolution gridded data set of daily precipitation and surface temperature. Journal of Geophysical Research: Atmospheres 114 (2009).
Huffman, G. J. et al. NASA global Precipitation Measurement (GPM) Integrated Multi-satellitE Retrievals for GPM (IMERG). Algorithm Theoretical Basis Document (ATBD), NASA/GSFC, Greenbelt, MD 20771, USA (2014).
Huffman, G. J., Bolvin, D. T. & Nelkin, E. J. Integrated Multi-satellitE Retrievals for GPM (IMERG) technical documentation. Tech. Rep., NASA/GSFC, Greenbelt, MD 20771, USA (2018).
Olsson, J., Berg, P. & Kawamura, A. Impact of RCM spatial resolution on the reproduction of local, subdaily precipitation. Journal of Hydrometeorology 16, 534–547 (2015).
Dai, A. Global precipitation and thunderstorm frequencies. Part I: Seasonal and interannual variations. Journal of Climate 14, 1092–1111 (2001).
Qian, J.-H. Why precipitation is mostly concentrated over islands in the Maritime Continent. Journal of the Atmospheric Sciences 65, 1428–1441 (2008).
Ogino, S.-Y., Yamanaka, M. D., Mori, S. & Matsumoto, J. How much is the precipitation amount over the tropical coastal region? Journal of Climate 29, 1231–1236 (2016).
Curtis, S. Means and long-term trends of global coastal zone precipitation. Scientific Reports 9, 5401 (2019).
Beck, H. E. et al. MSWEP V2 global 3-hourly 0.1° precipitation: methodology and quantitative assessment. Bulletin of the American Meteorological Society 100, 473–500 (2019).
Schlosser, C. A. & Houser, P. R. Assessing a satellite-era perspective of the global water cycle. Journal of Climate 20, 1316–1338 (2007).
Ellis, T. D., L’Ecuyer, T., Haynes, J. M. & Stephens, G. L. How often does it rain over the global oceans? The perspective from CloudSat. Geophysical Research Letters 36 (2009).
Hardwick Jones, R., Westra, S. & Sharma, A. Observed relationships between extreme sub-daily precipitation, surface temperature, and relative humidity. Geophysical Research Letters 37 (2010).
Peleg, N. et al. Intensification of convective rain cells at warmer temperatures observed from high-resolution weather radar data. Journal of Hydrometeorology 19, 715–726 (2018).
Allan, R. P. et al. Advances in understanding large-scale responses of the water cycle to climate change. Annals of the New York Academy of Sciences (2020).
Zipser, E. J., Cecil, D. J., Liu, C., Nesbitt, S. W. & Yorty, D. P. Where are the most intense thunderstorms on earth? Bulletin of the American Meteorological Society 87, 1057–1072 (2006).
Liu, C. & Zipser, E. J. The global distribution of largest, deepest, and most intense precipitation systems. Geophysical Research Letters 42, 3591–3595 (2015).
Behrangi, A., Tian, Y., Lambrigtsen, B. H. & Stephens, G. L. What does CloudSat reveal about global land precipitation detection by other spaceborne sensors? Water Resources Research 50, 4893–4905 (2014).
Menne, M. J., Durre, I., Vose, R. S., Gleason, B. E. & Houston, T. G. An overview of the Global Historical Climatology Network-Daily database. Journal of Atmospheric and Oceanic Technology 29, 897–910 (2012).
Lewis, E. et al. GSDR: a global sub-daily rainfall dataset. Journal of Climate 32, 4715–4729 (2019).
Blenkinsop, S. et al. The INTENSE project: using observations and models to understand the past, present and future of sub-daily rainfall extremes. Advances in Science and Research 15, 117–126 (2018).
Goodison, B. E., Louie, P. Y. T. & Yang, D. WMO solid precipitation intercomparison. Tech. Rep. WMO/TD-872, World Meteorological Organization, Geneva (1998).
Daly, C., Gibson, W. P., Taylor, G. H., Doggett, M. K. & Smith, J. I. Observer bias in daily precipitation measurements at United States cooperative network stations. Bulletin of the American Meteorological Society 88, 899–912 (2007).
Sevruk, B., Ondrás, M. & Chvíla, B. The WMO precipitation measurement intercomparisons. Atmospheric Research 92, 376–380 (2009).
Durre, I., Menne, M. J., Gleason, B. E., Houston, T. G. & Vose, R. S. Comprehensive automated quality assurance of daily surface observations. Journal of Applied Meteorology and Climatology 49, 1615–1633 (2010).
Funk, C. et al. The climate hazards infrared precipitation with stations—a new environmental record for monitoring extremes. Scientific Data 2, 150066 (2015).
Meteorological Organization (WMO), W. Guide to hydrological practices, volume II: Management of water resources and applications of hydrological practices, http://www.wmo.int/pages/prog/hwrp/publications/guide/english/168_Vol_II_en.pdf (WMO, Geneva, Switzerland, 2009).
Virtanen, P. et al. SciPy 1.0: Fundamental algorithms for scientific computing in Python. Nature Methods 17, 261–272 (2020).
Haylock, M. R. et al. A European daily high-resolution gridded data set of surface temperature and precipitation for 1950–2006. Journal of Geophysical Research: Atmospheres 113 (2008).
Oliphant, T. E. NumPy: A guide to NumPy. USA: Trelgol Publishing. www.numpy.org (2006).
van der Walt, S., Colbert, S. C. & Varoquaux, G. The Numpy array: A structure for efficient numerical computation. Computing in Science Engineering 13, 22–30 (2011).
Bishop, C. M. Neural networks for pattern recognition. (Clarendon Press, Oxford, UK, 1995).
Coulibaly, P., Dibike, Y. B. & Anctil, F. Downscaling precipitation and temperature with temporal neural networks. Journal of Hydrometeorology 6, 483–496 (2005).
Kim, J.-W. & Pachepsky, Y. A. Reconstructing missing daily precipitation data using regression trees and artificial neural networks for SWAT streamflow simulation. Journal of Hydrology 394, 305–314 (2010).
Nastos, P., Paliatsos, A., Koukouletsos, K., Larissi, I. & Moustris, K. Artificial neural networks modeling for forecasting the maximum daily total precipitation at Athens, Greece. Atmospheric Research 144, 141–150 (2014).
Hutchinson, M. F. Interpolation of rainfall data with thin plate smoothing splines — part I: Two dimensional smoothing of data with short range correlation. Journal of Geographic Information and Decision Analysis 2, 168–185 (1998).
Smith, R. B. et al. Orographic precipitation and air mass transformation: An Alpine example. Quarterly Journal of the Royal Meteorological Society 129, 433–454 (2003).
Roe, G. H. Orographic precipitation. Annual Review of Earth and Planetary Sciences 33, 645–671 (2005).
Molnar, P., Fatichi, S., Gaál, L., Szolgay, J. & Burlando, P. Storm type effects on super Clausius-Clapeyron scaling of intense rainstorm properties with air temperature. Hydrology and Earth System Sciences 19, 1753–1766 (2015).
Benestad, R., Nychka, D. & Mearns, L. Spatially and temporally consistent prediction of heavy precipitation from mean values. Nature Climate Change 2, 544–547 (2012).
Funk, C. et al. A global satellite assisted precipitation climatology. Earth System Science Data 7, 275–287 (2015).
Fick, S. E. & Hijmans, R. J. WorldClim 2: new 1-km spatial resolution climate surfaces for global land areas. International Journal of Climatology 37, 4302–4315 (2017).
Halkjær, S. & Winther, O. The effect of correlated input data on the dynamics of learning. In NIPS, 169–175, http://papers.nips.cc/paper/1254-the-effect-of-correlated-input-data-on-the-dynamics-of-learning (1996).
Overeem, A., Buishand, A. & Holleman, I. Rainfall depth-duration-frequency curves and their uncertainties. Journal of Hydrology 348, 124–134 (2008).
Rao, C. R. Linear statistical inference and its applications (2 edn, John Wiley and Sons, New York, 1973).
Chai, T. & Draxler, R. R. Root mean square error (RMSE) or mean absolute error (MAE)? – Arguments against avoiding RMSE in the literature. Geoscientific Model Development 7, 1247–1250 (2014).
Willmott, C. J., Robeson, S. M. & Matsuura, K. Climate and other models may be more accurate than reported. Eos 98 (2017).
Sharifi, E., Steinacker, R. & Saghafian, B. Assessment of GPM-IMERG and other precipitation products against gauge data under different topographic and climatic conditions in Iran: preliminary results. Remote Sensing 8 (2016).
Gupta, H. V., Kling, H., Yilmaz, K. K. & Martinez, G. F. Decomposition of the mean squared error and NSE performance criteria: Implications for improving hydrological modelling. Journal of Hydrology 370, 80–91 (2009).
Beck, H. E. et al. PPDIST: global 0.1° daily and 3-hourly precipitation probability distribution climatologies for 1979–2018. figshare https://doi.org/10.6084/m9.figshare.12317219 (2020).
Zolina, O., Kapala, A., Simmer, C. & Gulev, S. K. Analysis of extreme precipitation over Europe from different reanalyses: a comparative assessment. Global and Planetary Change 44, 129–161 (2004).
Mehran, A., AghaKouchak, A. & Capabilities, A. of satellite precipitation datasets to estimate heavy precipitation rates at different temporal accumulations. Hydrological Processes 28, 2262–2270 (2014).
Herold, N., Behrangi, A. & Alexander, L. V. Large uncertainties in observed daily precipitation extremes over land. Journal of Geophysical Research: Atmospheres 122, 668–681 (2017).
Legates, D. R. A climatology of global precipitation. Ph.D. thesis, University of Delaware (1988).
Herold, N., Alexander, L. V., Donat, M. G., Contractor, S. & Becker, A. How much does it rain over land? Geophysical Research Letters 43, 341–348 (2016).
Tustison, B., Harris, D. & Foufoula-Georgiou, E. Scale issues in verification of precipitation forecasts. Journal of Geophysical Research: Atmospheres 106, 11775–11784 (2001).
Chen, C.-T. & Knutson, T. On the verification and comparison of extreme rainfall indices from climate models. Journal of Climate 21, 1605–1621 (2008).
Harrigan, S. et al. GloFAS-ERA5 operational global river discharge reanalysis 1979–present. Earth System Science Data Discussions 2020, 1–23 (2020).
Tian, Y. & Peters-Lidard, C. D. Systematic anomalies over inland water bodies in satellite-based precipitation estimates. Geophysical Research Letters 34 (2007).
You, Y., Wang, N.-Y., Ferraro, R. & Rudlosky, S. Quantifying the snowfall detection performance of the GPM microwave imager channels over land. Journal of Hydrometeorology 18, 729–751 (2017).
Kubota, T. et al. Verification of high-resolution satellite-based rainfall estimates around Japan using a gauge-calibrated ground-radar dataset. Journal of the Meteorological Society of Japan. Ser. II 87A, 203–222 (2009).
Tian, Y. et al. Component analysis of errors in satellite-based precipitation estimates. Journal of Geophysical Research: Atmospheres 114 (2009).
Derin, Y. et al. Evaluation of GPM-era global satellite precipitation products over multiple complex terrain regions. Remote Sensing 11 (2019).
Donohue, R. J., Roderick, M. L. & McVicar, T. R. Roots, storms and soil pores: Incorporating key ecohydrological processes into Budyko’s hydrological model. Journal of Hydrology 436–437, 35–50 (2012).
Liu, Q. et al. The hydrological effects of varying vegetation characteristics in a temperate water-limited basin: Development of the dynamic Budyko-Choudhury-Porporato (dBCP) model. Journal of Hydrology 543, 595–611 (2016).
Kuligowski, R. J. An overview of National Weather Service quantitative precipitation estimates. United States, National Weather Service, Techniques Development Laboratory. https://repository.library.noaa.gov/view/noaa/6879 (1997).
Wolff, D. B. & Fisher, B. L. Comparisons of instantaneous TRMM ground validation and satellite rain-rate estimates at different spatial scales. Journal of Applied Meteorology and Climatology 47, 2215–2237 (2008).
Beck, H. E. et al. Bias correction of global precipitation climatologies using discharge observations from 9372 catchments. Journal of Climate 33, 1299–1315 (2020).
Osborn, T. J. & Hulme, M. Development of a relationship between station and grid-box rainday frequencies for climate model evaluation. Journal of Climate 10, 1885–1908 (1997).
Pietersen, J. P. J., Gericke, O. J., Smithers, J. C. & Woyessa, Y. E. Review of current methods for estimating areal reduction factors applied to South African design point rainfall and preliminary identification of new methods. Journal of the South African Institution of Civil Engineering 57, 16–30 (2015).
Pedregosa, F. et al. Scikit-learn: machine learning in Python. J. Mach. Learn. Res. 12, 2825–2830 (2011).
Wilson, A. M. & Jetz, W. Remotely sensed high-resolution global cloud dynamics for predicting ecosystem and biodiversity distributions. Plos Biology 14, 1–20 (2016).
Yamazaki, D. et al. A high-accuracy map of global terrain elevations. Geophysical Research Letters 44, 5844–5853 (2017).
Legates, D. R. & Bogart, T. A. Estimating the proportion of monthly precipitation that falls in solid form. Journal of Hydrometeorology 10, 1299–1306 (2009).
Acknowledgements
The Water Center for Arid and Semi-Arid Zones in Latin America and the Caribbean (CAZALAC) and the Centro de Ciencia del Clima y la Resiliencia (CR) 2 (FONDAP 15110009) are thanked for sharing the Mexican and Chilean gauge data, respectively. We acknowledge the gauge data providers in the Latin American Climate Assessment and Dataset (LACA&D) project: IDEAM (Colombia), INAMEH (Venezuela), INAMHI (Ecuador), SENAMHI (Peru), SENAMHI (Bolivia), and DMC (Chile). We further wish to thank Ali Alijanian and Piyush Jain for providing additional gauge data. We gratefully acknowledge the developers of the predictor datasets for producing and making available their datasets. We thank the editor and two anonymous reviewers for their constructive comments which helped us to improve the manuscript. Hylke E. Beck was supported in part by the U.S. Army Corps of Engineers’ International Center for Integrated Water Resources Management (ICIWaRM). Hayley J. Fowler is funded by the European Research Council Grant, INTENSE (ERC-2013-CoG-617329) and the Wolfson Foundation and the Royal Society as a Royal Society Wolfson Research Merit Award holder (grant WM140025).
Author information
Authors and Affiliations
Contributions
H.E.B. conceived, designed, and performed the analysis and took the lead in writing the paper. E.F.W. was responsible for funding acquisition. All co-authors provided critical feedback and contributed to the writing.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
The Creative Commons Public Domain Dedication waiver http://creativecommons.org/publicdomain/zero/1.0/ applies to the metadata files associated with this article.
About this article
Cite this article
Beck, H.E., Westra, S., Tan, J. et al. PPDIST, global 0.1° daily and 3-hourly precipitation probability distribution climatologies for 1979–2018. Sci Data 7, 302 (2020). https://doi.org/10.1038/s41597-020-00631-x
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41597-020-00631-x
This article is cited by
-
Positive correlation between wet-day frequency and intensity linked to universal precipitation drivers
Nature Geoscience (2023)
-
Deep learning-based multi-source precipitation merging for the Tibetan Plateau
Science China Earth Sciences (2023)