Abstract
The quality of digital elevation models (DEMs), as well as their spatial resolution, are important issues in geomorphic studies. However, their influence on landslide susceptibility mapping (LSM) remains poorly constrained. This work determined the scale dependency of DEMderived geomorphometric factors in LSM using a 5 m LiDAR DEM, LiDAR resampled 30 m DEM, and a 30 m ASTER DEM. To verify the validity of our approach, we first compiled an inventory map comprising of 267 landslides for Sihjhong watershed, Taiwan, from 2004 to 2014. Twelve landslide causative factors were then generated from the DEMs and ancillary data. Afterward, popular statistical and machine learning techniques, namely, logistic regression (LR), random forest (RF), and support vector machine (SVM) were implemented to produce the LSM. The accuracies of models were evaluated by overall accuracy, kappa index and the receiver operating characteristic curve indicators. The highest accuracy was attained from the resampled 30 m LiDAR DEM derivatives, indicating a fineresolution topographic data does not necessarily achieve the best performance. Additionally, RF attained superior performance between the three presented models. Our findings could contribute to opt for an appropriate DEM resolution for mapping landslide hazard in vulnerable areas.
Similar content being viewed by others
Introduction
Globally, landslides are one of the most devastating of geohazards that impose serious threats to human life and economic conditions by the neverending socioeconomic burdens^{1}. With the recent changes associated with the unplanned urban expansions and severe climatic extremes, landslides are expected to increase dramatically due to the heavy rainfall and the serious infrastructure constructions in mountainous areas^{2,3}. To reduce the economic burden and human losses, it is helpful to delineate and identify potential landslideprone areas. In this regard, landslide susceptibility mapping (LSM) is regarded as a useful tool in disaster management and mitigation^{4,5,6,7}. The susceptibility zonation maps can be the first step towards a complete risk assessment that assist authorities and decisionmakers for initiating appropriate mitigation measures. Over the past few decades, LSM techniques were extensively developed and implemented by several researchers, and each technique has proven to have its own sole merits and demerits^{6,8,9}.
Landslide susceptibility expresses the likelihood of a landslide event occurring in a given area based on local terrain conditions. LSM partitions the geographical surface into zones of varying grades of stability based on evaluating the consequence of the probability toward landsliding generated using the causative factors estimated to induce the instability^{10,11,12,13}. Landslide susceptibility mapping (LSM) plays a significant role in risk mitigation, especially in introducing countermeasures aimed at decreasing the risks associated with landslides^{14,15}. Moreover, LSM could be engaged to depict locations of unknown landslides in the future, support with emergencydecisions, and mitigate future hazardous events.
For landslide susceptibility evaluation, GIS has proven to be a powerful tool due to its capability of handling a variety of spatial data, their processing capability, and easiness in the decisionmaking process^{9,16,17}. Over the years, landslide susceptibility assessment topic has gained significant attention of many scholars owing to its direct consequences on the people life. Numerous studies have been carried out for landslide susceptibility assessment around the world in the current decade using a wide variety of techniques^{9,18}. Several of them employed the relationship between the landslide causative factors and landslide occurrence through the spatial data analysis^{8,13}. Such relationships can be categorized in terms of rankings or weight. In this context, such datadriven methods can be classified into two distinct categories: qualitative and quantitative^{13,19}. The former are rather subjective^{20}. On the other hand, the latter are based on statistics, and with the development of computer systems and GIS tools, these models have become more prevalent than the qualitative methods^{21}. Quantitative methods such as logistic regression, fuzzy logic, certainty factor, and information value approaches are useful for problemsolving and have been successfully used in different scientific fields, such as engineering, and hazard evaluation applications. Very recently, numerous machinelearning (ML) techniques have been applied for different fields, owing to their robustness in handling large complicated data^{8,22,23,24,25}. These methods include Artificial Neural Networks (ANN), Decision Trees (DT), Support Vector Machine (SVM) and Random Forests (RF) were comparatively new in the field of landslide research. Despite different statistics involved, their terminologies, and computation capability, all of the aforementioned methodologies are largely based upon the following assumptions^{20}; (i) past is the key to the future; (ii) factors involving the landsliding are spatially linked and therefore could be used in predictive functions; (iii) future events will likely happens in similar conditions.
Apart from the statistical and computational aspects, the accuracies of the susceptibility model depend upon the quality of spatial data and the choice of relevant causative factors. Diverse intrinsic and extrinsic factors are castoff to analyze LSM. The typical factors that can be derived from a DEM and other sources which influence the landslides are known through several past research. For example, the review of ‘statisticallybased landslide susceptibility models’ by Reichenbach et al.^{20} grouped the influencing factors into five categories; (i) morphological, (ii) geological, (iii) land cover, (iv) hydrological, and (v) others. Süzen and Kaya^{26}, listed about 18 causative factors in the triggering mechanism of landslides. However, in any given situation, some of these factors may be important whilst others are irrelevant^{18}. These factors come from different sources, and their quality varies widely. Thus, the landslide evaluation exclusively based on a digital elevation model (DEM) has been conducted assuming that topography reflects other causative factors such as hydrology and land use. The availability of global DEMs and recent advances in DEM acquisition techniques encourage this approach.
Accurate topographic input comes from highquality DEMs, along with the geological conditions are usually necessary for producing accurate susceptibility products^{18}. Generally, DEM’s produced using interpolation of contours from a topographic map, radarbased Shuttle Radar Topographic Mission (SRTM) DEM, and stereooptical derived Advanced SpaceBorne Thermal Emission Radiometer (ASTER) DEM are used for susceptibility analysis if no highresolution DEM is available for the studied region. They come with varying spatial resolution; 10 m – 90 m. A coarser DEM describes the terrain less accurately, resulting in the propagation of error on to the secondary derivatives such as slope, aspect, and curvature, etc.^{27}.
While the scale effects of landslide causative factors, particularly topographic variables derived from DEM are wellknown issues in geomorphology, only a few studies have attempted the potential effects it may have on susceptibility models. For example, Guzzetti and others^{28} suggested the use of multiple resolutionDEMs for testing and to opt the best performed one in final susceptibility mapping. However, with the increasing availability of very highresolution DEMs derived from LiDAR and UAV images, researchers tend to use 1–5 m DEM’s in their modeling part expecting that a finest DEM can describe more detailed topography. Paudel et al.^{29} studies, however, argued that the smallestscale variability does not well represent the physical processes because the local topography does not resemble the processes of controlling landslide initiation. On the other hand, Tian et al.^{30} by analyzing 5–190 m DEMs, indicated that the optimal resolution often depends on the chosen size of the study area. Catani et al.^{31} coined the term Mapping Unit Resolution (MUR) to define the raster resolution and performed the scale effects out at six different MUR (10–500 m). Their results are in line with^{29}, where the finer resolutions are found less accurate. Nevertheless, performing sensitivity analysis was recommended when LSM results are utilized for planning and protection purposes in a given area. Conclusive answers for identifying the optimal scale for global reach, therefore needs further investigation, especially for hazard assessments. A poor understanding of scale effect may inadvertently promote frequent use of highresolution DEMs, thus demanding substantial computational requirements. Because disaster management and mitigation require quick responses, timely interventions are necessary. Therefore, a special focus was given in this work to address the scale dependency in detail.
With these objectives in mind, the present work aims at: (i) producing a rich analysis of the effect scale dependency in landslide assessment; (ii) demonstrating the appropriateness of certainty factor model in selecting significant influencing factors; and (iii) address the landslide susceptibility issue for the study area by benefiting from multiple machine learning models. To achieve the aforementioned objectives, we employed an integrated approach comprising three varying resolution elevation models, certainty factor for investigating the relationship between correlated factors and landslide occurrence, and the widely applicable statistical and machine learning models such as Logistic Regression (LR), Random Forest (RF), and Support Vector Machine (SVM) for producing the susceptibility maps. All analyses were conducted in the programming environment R (3.6.0). The source code of this research is publicly made available online (https://github.com/aminevsaziz/lsm_in_Sihjhong_basin) to ensure results reproducibility. ArcGIS (10.4) and SAGA (4.0.1) were used for compilation and visualization of factor and susceptibility maps.
The general idea of selecting LR, RF, and SVM in our analysis is that LR is the most popular susceptibility model^{19}, whereas RF and SVM are the most promising ones^{9,13}. While LR is simple, straightforward, and highly interpretable, however, it cannot solve nonlinear problems. RF, on the other hand, produces a more accurate and robust prediction, but is less descriptive^{9}. SVM though delivers a unique solution for complex problems with its kernel tricks, but the kernelspecific parameter selection is a complex process. A combination of learning models increases the overall understanding of the issue, but the computational requirements vary. Therefore, we also aimed to quantify the average time required to train and test each of these popular models in view of mitigation preparedness.
Study Area and Data Used
Overview of the study area
Taiwan has a land area of 36,000 m^{2}, 26.68% of which is covered by plains, whereas 27.31% is hilly and 46.01% is mountainous. According to the statistics of the National Fire Agency (NFA), many natural disaster events are occurring in Taiwan, include typhoons, flooding, earthquakes, torrential rainfall, windstorms, and landslides. The selected study areaSihjhong watershed is located in the Hengchun Peninsula in the southern part of Taiwan (Fig. 1). Because of sustained economic growth and land development, the steep terrain in this region has undergone frequent modification in land use pattern. Windward portion of the selected watershed in the recent past has suffered from multiple landslides (Fig. 1b) triggered by heavy rainfall during Pacific typhoon seasons. On average, about five typhoons are expected to affect the Island nation a year. In recent years, it is aggravated by global climate change. Rainfall is plenty in the peninsula and annual accumulated rainfall can be reached up to 3600 mm. The altitude of the study area varies from 0 m to 700 m with a mean of 110 m. Moderately gentle to steep hills and mountains are typical of the Hengchun Peninsula. On the west, the study area is bounded by the South China Sea with flat long coastal plains. The average and maximum slope derived from a 5 m LiDAR DEM are 15° and 66° respectively. Geologically, the study area is composed of thick sedimentary strata. The most dominating lithological unit in the Sihjhong area is Shale with alteration sequence. A detailed description of individual lithologic types is provided in the data section. From a disaster perspective, Sihjhong is an important case area with multiple hazards from typhoons (e.g., flood and landslides) in the sight which may aggravate with extreme climate^{32}. Therefore, performing landslide susceptibility analysis is key for providing baseline information to practitioners and lawmakers^{6,11}.
Data used
The multitemporal landslide inventory database for the study area from 2004 to 2014 was portrayed in Fig. 1. Figure 2 shows examples of a landslide inventory map prepared from dynamic timeseries image analysis carried for the study area. This dataset was downloaded from the NGIS Data Warehouse and Web Service Platform (TGOS Portal) developed by the Information Center, Ministry of Interior in Taiwan. The landslide inventory was created by interpreting Formosat2 satellite data and an expert landslide and shaded area delineation system (ELSADS). The accuracy of landslide inventory has been carefully validated manually with the help of aerial images at 25 cm spatial resolution. The overall accuracy of this inventory was tested previously and found to be 98%^{32}. The number and area statistics of landslides and typhoon details for each landslide inventory in the study area from 2004 to 2014 is illustrated in Table 1. Many landslides occurred in 2008 because of short duration and highintensity rainfall.
The causative factors influencing the spatial distribution of landslides have been extensively explained in literature^{6,13,19}. A general summary of these studies suggests that selection of the landslide predisposing factors in a given case should take into account: (i) the characteristics of the study area, (ii) the landslide type, (iii) scale of the analysis, and (iv) the data availability^{13,14,21}. The causative factors selection in this research was based on the aforementioned summarization concerning spatial relationships between landslide occurrence and causative factors comprising topography, hydrology, tectonics, geology, and geomorphology^{9,14,19}. Tectonic factor (distance to fault) was later discarded because faults are not corresponding with the landslides identified in Fig. 3. Moreover, the triggering mechanism for our landslide inventory was attributed to rainfall alone^{33}.
After careful assessment, a total of twelve landslide causative factors were finally selected for this case study, i.e., elevation, slope angle, slope aspect, total curvature, plan curvature, profile curvature, terrain position index (TPI), terrain roughness index (TRI), distance from the road, distance from drainage networks, rainfall, and lithology. All twelve causative factors were processed and analyzed with the assistance of SAGA and ArcGIS® software. The first eight causative factors were derived from three different digital elevation models (DEM).
DEM is a digital grid form of representation for the terrain’s surface. DEM can be created from various technologies, such as Terrestrial Surveying, Aerial Photogrammetry, Light Detection and Ranging (LiDAR), Interferometric Synthetic Aperture Radar (InSAR). The common applications of DEMs include geomorphometric feature extraction, hydrological modeling, geohazard inventory, lightofsight analysis, and landscape modeling and ecosystem management, etc. Highquality DEMs are required for precise applications. For this study, we have used two kinds of DEM’s: the first one is a LiDARderived 5meter DEM (hereafter termed as 5 m LiDAR DEM) derived from investigation results of changes in surface topography and environmental geology caused by Typhoon Morakot, happened on August 2009 from Central Geological Survey in 2013. After field verification, the overall geometric accuracy found between 0.5 and 1.0 m.
The second kind is the ASTER Global Digital Elevation Model with 30meter resolution (hereafter named 30 m ASTER DEM) is a joint product developed and made available to the public by the Ministry of Economy, Trade, and Industry (METI) of Japan and the United States National Aeronautics and Space Administration (NASA). It can be available free of charge to users worldwide from the Land Processes Distributed Active Archive Center or shortly LP DAAC (https://lpdaac.usgs.gov/products/astgtmv002/). The vertical accuracy of ASTER GDEM version 2 had been revealed a standard deviation is 5.9~12.7 meters. (https://asterweb.jpl.nasa.gov/gdem.asp). Additionally, we resampled the 5 m LiDAR DEM into 30 m resolution using bilinear interpolation technique to have a comparison with the 30 ASTER DEM. The source of the road and hydrology network map used in this study is obtained from a digital map of the traffic network produced by the Ministry of Transportation and Communications. Lithology data is digitized from a 1:50000 Geology map produced by the Central Geological Survey (Fig. 3). There are six lithology types contained in the area including Gravel, sand and clay (Type I), Shale and thin alternation of sandstone and shale with thickbedded sandstone and conglomerate lentil (Type II), Sandy conglomerate (Type III), Mudstone and various exotic blocks (Type IV), Thickbedded sandstone, interbedded sandstone and shale (Type V), and Thickbedded sandstone intercalated with conglomerate (Type VI).
Methods
Implemented models
We employed three popular machine learning algorithms to map landslide susceptibilities. While logistic regression (LR) is a parametric machine learning algorithm (learning model that summarizes data with a set of parameters of fixed size  no matter how much data we input at a parametric model, it won’t change its mind); both support vector machine (SVM) and random forest (RF) are nonparametric models (algorithms that do not make strong assumptions about the form of the mapping function; also the complexity grows as the number of training samples increases)^{19,34}. Among these two nonparametric models, RF does not need any real hyperparameters to tune, whereas SVM requires tuning for the right kernel, regularization penalties, and the slack variable^{13,35}. Detailed description and computation of each ML algorithm are provided in the following sections.
Logistic regression
Logistic Regression is a popular statistical modeling method which has been applied widely in many problems such as gene selection in cancer classification and crime analysis^{18,36}. In landslide susceptibility analysis, the LR has also used popularly in many case areas^{19,37}. In the LR, the main mathematical concept is to use the logitthe natural logarithm of an odds ratio, which is expressed as follows:
where: n is the number of the variables used, α_{o} means the intercept, and α_{i} are defined as the coefficients related with the explained variables x_{i}, and prob means the probability of a landslide occurrence which is a nonlinear function of x_{i} is expressed as follows:
Support vector machine
Introduced by Vapnik^{38}, Support Vector Machine (SVM) is a wellknown unsupervised learning machine learning method which has been applied successfully and effectively in landslide susceptibility mapping^{34,39}. The main concept of the SVM is to apply the linear model to carry out the nonlinear class boundaries by nonlinear mapping the input vectors into the new highdimensional feature space where the optimal separating hyperplane is built to separate output classes for classification. More detail, the optimal separating hyperplane is the maximum margin hyperplane, which offers the maximum separation between the output classes, and the training samples which are closest to this hyperplane called support vectors. In the linearly separable problem, the optimal separating hyperplane of binary decision classes can be computed as follows^{40}:
where y is defined as the outcome class, x_{i} means the input variables, and w_{i} mean the weights which determine the hyperplane.
Random forest
Random Forest (RF) is an effective ensemble classifier, which constructs multiple decision trees for classification utilizing a subset of variables randomly selected^{41}. It is a machine learning technique as well, which has been used to solve a lot of realworld problems such as monitoring of land cover, predicting proteinprotein interactions, predicting disease risks^{9,35}. In landslide prediction, the RF has also been applied in several types of research. In literature, the RF is a popular method with high performance as it has several advantages such as (1) It is a nonparametric naturebased method, (2) it is able to determine the importance of variables used, (3) it provides an algorithm to estimate the missing values, and it is flexible for the analysis of classification, regression and unsupervised learning^{42}.
In the RF, one subset of the predictor variables are utilized to construct each tree, and the number of trees (n_{tree}) and the number of the predictors used to build each tree (m_{try}) can be different which depend on the dataset. Using the RF, each tree is constructed from a bootstrap sample of primary training dataset used to estimate the robust error with the testing dataset expressed as follows:
where MSE means mean square error calculated during constructing the classification trees, n is the number of out of bag observation in each tree, \(\overline{{t}_{i}\,}\,\)is defined as the average of whole out of bag predictions^{43}. Percentage of the explained variable is calculated as follows:
where: V_{z} means the total variation of the response variable. At last, the outcome of the RF is one single prediction that is the mean of all aggregations.
Certainty factor (CF)
The certainty factor (CF) model is an approach for handling uncertainty in rulebased systems, which has been broadly used in expert system shell field, additionally, to medical diagnosis studies^{15}. The CF model is one of the probable favorability functions to solve the problem of incorporating heterogeneous data^{44}. The universal theory function is expressed as:
where PP_{a} is the conditional probability (CP) of owning a number of landslide events happen in class a and PP_{s} is the prior probability (PP) of owning a total number of landslide events in the case study area. The value of PP_{s} this study is computed to be 0.012.
The range of CF value varies [−1, 1]. Positive values denote an increasing certainty in landslide occurrence; negative values imply a decrease in the certainty. A CF value near 0 shows that the prior probability is near to the conditional probability, and thus, it is difficult to determine the certainty of landslide occurrence^{15}. The favorability values are acquired by overlapping landslide inventory maps and each data layer and calculating the landslide frequency. The CF model provides a rank measure of certainty in forecasting landslides. The relationship between the landslide sites and used causative factors had been analyzed in this study.
Construction of the geospatial database for the training and the validation dataset
The input dataset obtained from the geospatial database of this experimental research was fed directly into the required models without extra encoding (i.e., dummying or numerically decoding of categorical variables) because the selected models handle efficiently diverse space variables (i.e., numeric and categorical). Also, it is critical to understand that the input dataset is not only for training the models. In the absence of an independent testing dataset, a common approach is to estimate the predictive performance based on resampling the original data. These strategies divide the data into training sets and a testing sets, while ranging in complexity from the popular simple holdout split to Kfold crossvalidation, MonteCarlo Kfold crossvalidation, bootstrap resampling^{45}. They can be used efficiently for models selection, accuracy assessment, and hyperparameters tuning^{3,13}.
In our study, the input dataset was randomly split into two sets (training and testing datasets) by 70:30 ratios, and then the training set was innerly resampled using ten kfolds crossvalidations. The implemented resampling approach is considered as the golden standard for machine learning, because they are found effective as it reduces the split randomness that comes with testtrain split strategy, which allows the input dataset to be used for three different purposes: (1) tuning models hyperparameters, (2) to train models with this subset using after optimal parameters are found, and (3) models validation, assessment, and comparison.
Model configuration and implementation
Some models (i.e., RF and SVM) require a finetuning for its hyperparameters on which the model performance depends. Usually, such feat is achieved by manual tuning using techniques such as grid search, random search, and even gradientbased optimization. However, such techniques have proven to be suboptimal at best, considering the fact that manually exploring the resulting combinatorial space of parameter settings is quite tedious and tends to lead to unsatisfactory results. Moreover, the obtained optimal hyperparameters cannot be reproduced to a certain degree, and that is because of such techniques rely on “Trial and Error” experimenting, which depends on analyzing that learning curves and decide that best learning path. This drawback is so critical especially if the modeling experiment involves complex experiments with a fair amount of data to process and for that fact, we opted for a Stateoftheart algorithm socalled sequential modelbased optimization (SMBO) to finetune models hyperparameters.
Sequential modelbased optimization (Fig. 4) is unique automated approaches for solving algorithms configuration and hyperparameter optimization of expensive blackbox models. SMBO is known to converge for the low computational budget performance is due to: (1) the capability to reason about the quality of experiments before they are run; and (2) advancing from the “adaptive capping” to avoid long run^{3}.
When it comes to model implementation, only RF and SVM require tuning to some of its hypermeters. The overallhyperparameters utilized for each model was summarized along with its value, short description, and the package used to run the model, is in Table 2. The search space for each required hyperparameters was set according to guides and manuals of each package that implement each model.
Only “mtry” and “num.trees” are allowed to fix by the user according to some instructions and strategies. Or else, the left parameters are set exactly to the allowed (or default) values (or range of values) by each package. The number of variables is for each tree (i.e. “mtry”), various heuristics recommended by packages that provided RF are used to set the optimum values (Table 3). These heuristics advise that ranges of 2 to 8 would be excellent for “mtry”. On the other hand, the total number of trees to fit (i.e., “num.trees” for RF) is set to exponential rate via a base of 2 (i.e.2^{i}, i = 5, …, 11). By allowing for the instructions of the used packages and some experimental researches, an optimal value of 2^{5} to 2^{10} was set^{3}.
During tuning, hyperparameters need to be carefully optimized, so as much accuracy the model is achieving, the model selection will be reliable. In general, the tuning process must be a formal and quantified part of the model evaluation yet, in most cases personal experience and intuition, heavily intervene by influencing the process in ways that are hard to quantify or describe^{46}. In this study, three techniques were implemented, i.e., LR, RF, and SVM, only LR is straight forward and does not require any further tuning. The training process was started by searching the optimal parameters using SMBO with 10fold crossvalidation on the training set that represents 70% of the input data to prevent overfitting. The chosen optimum pairs of hyperparameters that have the highest classification accuracy are shown in Table 4.
Models evaluation and comparison
Various performance metrics can be executed for quantitative assessment; however, we consider the Accuracy (Acc) as main metric for hyperparameters tuning and one of the main overall performance indicator metrics for the landslide predictive models. In this study, Acc together with Cohen kappa index (kappa)^{47} and the Area under the ROC Curve (AUC), were used to evaluate the overall performance and the predictive capabilities of the tuned models.
Additionally, model performance was evaluated using one of the most important nonparametric tests called the Friedman test^{48}. The Friedman test is heavily used for multiple comparisons to perceive significant differences between the performances of two or more approaches because the test involves no previous information for the used data and still is valid even if the data are normally distributed and was designated in this study^{49}.
The Friedman test has a null hypothesis, viz., there are no differences between the performances of the landslide models. The pvalue is the probability of refusing the null hypothesis if the hypothesis is true. Then each model is assessed. The higher the pvalue, the more likely that the null hypothesis is rejected.
Another useful use for the Friedman nonparametric test is the ability to obtaining a “Critical differences” diagram of multiple classifiers. A value called “Critical Difference” (i.e. calculated according to the equation below) indicates the critical average rank performance. If the average rank of the classifiers is within the critical difference distance (CD) then they are not statistically significantly different.
where: α is the confidence level, K is the number of models and N is the number of measurements. To calculate q_{α, K} the Studentised range statistic for infinite degrees of freedom divided by \(\,\sqrt{2}\) is used.
Landslide susceptibility map assessments
At the end of the validation and assessment processes, landslide susceptibility maps can be generated to: (1) assess the quality of the generated maps; and (2) check the input dataset for its suitability for later usage in other tasks (i.e., decision making) because it’s common to have some variables that have high correlation or even multicollinearity and these variables need to be check before using them as variables. However, to achieve those goals a key step must be performed. Usually, that step involves assessing the sufficiency and accuracy of the generated susceptibility maps based on the empirical assumption that state: “A model is sufficient and accurate when there is an increasing landslide density ratio when moving from low to high susceptible classes and high susceptibility classes cover small areas extent^{3,9}”. This means, a sufficiency analysis is essentially based on susceptibility maps and can be implemented by: (i) reclassifying the probability pixels produced for the whole study area by each model; (ii) overlying the existing landslide inventory over the susceptibility maps so to be able to obtain representative statistic for each susceptibility class (i.e., landslide density and extent).
Scale Effects of Geomorphometric Factors
It has been proved in the literature that topographic variables coming from a digital elevation model are the prime component for any susceptibility analysis. Furthermore, several studies indicated that the quality of DTMs would affect the overall model results^{27,30}. Therefore, the certainty factor (CF) method had been conducted to analyze the scale effects of geomorphometric factors for two kinds of DEMs in different quality and resolution in this section. For this, the 5 m Lidar DEM is downsampled to 30 meters to have a comparison with ASTER DEM, and then the CF values are calculated according to the landslide characteristics in each geomorphometric factors generated by the three elevation models. The 5 m Lidar DEM, 30 m Lidar DEM, 30 m ASTER DEM, and their derivative factors for the study area is shown in Fig. 5 and 6(a–l), respectively. Subsequently, these DEMs were employed to produce the LSM maps.
Results and Analysis
The relationship between landslide causative factors and landslide occurrence
The relationship between landslide causative factors and landslide occurrence was identified by CF model using 5 m Lidar DEM as shown in Table 5. This table presents the CF value for all causative factors, including the eight geomorphometric factors. Additionally, the CF value statistics of eight geomorphometric factors with the varying resolution of DEMs (5 m LiDAR, 30 m LiDAR and 30 m ASTER DEM) are summarized in Table 6 for the exploration of the scale effects. According to the distribution of the CF value in Table 5, for the higher Elevation, Slope, TRI, and TPI values, the certainty increased in landslide occurrence. With Aspect value of 66–247°, i.e., slope face to northeastsouthwest direction, the certainty increased in landslide occurrence. No matter the type of curvature, the larger or smaller its value, the larger its corresponding CF value. For lithology, Type II, V, and VI corresponds with a positive CF value, i.e., the certainty increased for the shale and thin alternation of sandstone and shale with thickbedded sandstone and conglomerate lentil (Type II), thickbedded sandstone, interbedded sandstone and shale (Type V), and thickbedded sandstone intercalated with conglomerate (Type VI). The thickbedded sandstone intercalated with conglomerate has the least cementation degree and the highest material discontinuity. Therefore, the interlayer slip is most likely to occur. Besides, the closer to the drainage networks, the greater the landslide occurrence.
From Table 6, it is observed that different spatial resolutions of topographic data do not affect largely on the trend of CF values, except for the impact on the curvature factors, comprising total, profile, and plan curvature because curvatures are defined by means of a second derivative of the elevation and a second derivative amplifies greatly even the smallest differences between the DTMs. However, the results on different data quality of DEMs indicate that geomorphometric features cannot be accurately derived from a lower data quality of DEM, e.g., an ASTER DEM. Therefore, the CF values shown a different trend for ASTER DEM compared with ones derived from Lidar DEM with the same resolution. For example, CF values with Slope of 20–30° changed from negative to positive. And ones with TRI of 2–4 changed from positive to negative. Moreover, the CF value of −1 on curvature factors shows that results with a lower data quality of ASTER DEM are unable to render more detailed topographic curvature. It implies that some geomorphometric features derived from different data quality of DEMs will be affected significantly on the landslide susceptibility modeling.
Landslide susceptibility map assessments
Achieving decent models’ performance is not the end road for LSM analysis. Additional steps involve:(1) creating the landslide susceptibility maps for the case study area in the form of probability grids using the validated models; (2) reclassifying susceptibility grids; and (3) analyzing the overall grids and assess its quality. The initial two steps are based on predicting the study area probabilities toward landsliding and afterward, a simple reclassification into five susceptibly classes that vary from very low to very high (Fig. 7) using Table 7 is performed. The last step is critical for understanding the overall pattern of landslides distribution and landslides susceptible areas and can be performed by attaining a landslide density distribution by overlapping the existing inventory map over the generated susceptibility maps and afterward, a summary statistic for the area covered by each susceptibility classes (Fig. 8) is obtained.
A visual analysis of the resulting LSM maps (Fig. 7), shows a smooth surface produced by each model for each DEM dataset. An obvious differentiation between Lidar datasets (5 and 30 meters) maps and ASTER dataset maps is represented in the form of very smooth transitioning from each susceptibility class to another. The results of sufficiency analysis (Fig. 8) were positive as they fulfilled the two required spatial conditions: (1) landslide pixels should belong to the highest susceptible class available; and (2) the extent areas covered by higher susceptible classes need be lower as possible. The results are similar to models evaluation results, LiDAR datasets (i.e., 5 and 30 meters) in particular and RF models, in general, achieve better results than the rest of the models with wellbalanced outcomes that put confidence in the overall LSM produced by either Lidar 5 meters or 30 meters. However, it is very crucial to understand that landslide density in Fig. 7a,c,e have a moderate presence of landslide events in very low susceptibility class despite the models achieved excellent scores regarding performance metric and that is due to how the stable nonlandslide samples are sampled. Usually, misclassifications on the extremes (very low and very high) tend to indicate the overall confidence in the misclassification of the model, but that depends on modeling experiment conditions.
Model evaluation and comparison
The optimum hyperparameters obtained in Table 8 for each model in each respective dataset, were used to train each model and assess the overall performance of the models using performance metrics indicators such Acc, AUC, and Kappa index.
The generated overall rank matrix of the implemented models (Fig. 9) based on performance results (Table 8 and Fig. 10) are generally in favor of RF being ranking top of all model in all datasets, followed up by either SVM or LR depending on the dataset for (i.e. LR on Lidar 30 meters dataset was able to achieve better results than SVM). However, a detailed analysis on the dataset level shows that Lidar 5 meters dataset models achieved far better results than Aster 30 meters dataset models, but surprisingly the highest performance results in term of all metrics were achieved by the resampled Lidar Dataset from 5 meters to 30 meters. These dataset models were able to achieve excellent results exceeding closest dataset models (i.e., Lidar 5 meters) by a margin ranging from 1% to 3%, 1% to 1.5% and 4% to 10% in term of AUC, Acc and Kappa respectively.
Despite the fact, the difference between each dataset models regarding performance results is relatively noticeable. However, Friedman nonparametric test at the significant level α = 5%was performed on models’ performance results in all datasets rather than inside each dataset (Table 9). These results show that the differences in performance between the implemented model are statistically insignificant between datasets because the p value exceeds the significant level of 0.05.
Additionally, the critical difference plot (Fig. 11) generated using the Friedman nonparametric test, shows that there’s a line connecting models indicating that they are within the insignificance range (i.e., critical difference range) of 1.91, which means that there are no statistical differences among all model.
Discussion
Effect of grid resolution and data quality on susceptibility models
Landslide susceptibility assessment is a useful task for landslide hazard management and mitigation^{8,9,50}. However, landslide is a complex natural phenomenon which is controlled by several geoenvironmental factors; thus, it is not easy to be modeled accurately^{8,9}. Datadriven models are proved to be an effective tool for landslide susceptibility modeling^{19,29}. Very recently, a large number of machine learning approaches are adopted and applied successfully for landslide susceptibility assessment^{8,21}. However, the performance of these models depends mainly on the input data. Therefore, it is essential to test and check the quality of the data before providing it as an input in the learning models. Typically, a large portion of the input factors in susceptibility modeling comes from a DEM^{8}. Consequently, the quality of DEM data or more specifically, the DEM derived causative factors used in the model are very crucial input for producing an accurate LSM output. In addition to having an appropriate quality, the scale of selected DEM is also vital in landslide hazard assessment. This is because the details of the topographic information provided in a DEM depends upon its spatial resolution^{29,51}. Several studies consider the DEM resolution as a first filter that assimilated into a model^{52,53}. Researchers are often direct for the highest spatial resolution product for mapping the finest details^{13}. However, an increase in spatial resolution means increased computational requirements for preprocessing the data. Moreover, with different DEM resolution, the primary topographic attributes such as slope angle and curvature exhibit substantial local variations^{54}.
In this study, we have demonstrated the scale effects of geomorphometric factors derived from two DEMs with varying spatial resolutions (i.e., LiDAR and ASTER) in analyzing the landslide susceptibility of Sihjhong watershed region. Contrary to the general expectations, but in line with the findings of Catani et al.^{31} and others^{27,29,30}, our result shows that a fine raster resolution DEM (5 m) does not significantly help in increasing the model prediction accuracy. Accuracies (AUC Values – see Fig. 9) obtained for the three different datadriven models indicate that 30 m resampled LiDAR DEM produces the best fit with the field data. Probable reasons are highlighted below why a finer MUR does not necessarily provide the best results. Firstly, landslide susceptibility assessments are dealing with the local geomorphological processes. Like any other geomorphic processes, landslides are also influenced by the morphology measured at the mesoscale level that is more representative of the hillslope forms and processes of such kind. However, finer DEMs would account for topography variations at the microscale, and probably those forms are not very much related to mesoscale processes like landslides.
Furthermore, the minimum landslide size mapped from the satellite images is 0.1 hectare, hence the LSM results from a 30 m resolution DEM is a good option. Excessive detailing of topography from the highresolution models are discussed in several studies and pointed out that the general trend of relief is often a better predictor of mesoscale processes than detailed information^{55,56}. Additionally, slope and curvature derived from a fine resolution DEM are higher than the coarser resolutions (see Fig. 5); this may result in more number of false positive rates. Similar results were also noticed in other studies^{29,53,57}. Zhang and Montgomery^{51}, portrayed that for many landscapes, a medium resolution grid size explores a rational compromise between improving resolution and data volume for simulating geomorphic and hydrological processes. Therefore, appropriate DEM resolution should be selected depending upon the aim of the modeling, characteristics of the study area, and the availability of data.
On the other hand, subpar quality of DEM can decrease the modeling accuracy as well. Therefore, the CF values showed a different trend for 30 m ASTER DEM compared with the one derived from Lidar DEM with exactly the same resolution. Although the terrain representation by ASTER GDEM used in this study is superior to SRTM‐3 for most landform elements^{58}, their accuracies for forested terrains and low elevated regions remains questionable^{59,60}. Furthermore, when compared with locally derived LiDAR DEMs, their RMSE is found to be large^{60}. This implies that ASTER DEM has inherent artifacts in producing a realistic representation of terrain features. A large part of inherency comes from the processing stage itself as they were developed from a compilation of over 1.2 million ASTER AVNIR scenes, many of it contains clouds obscuring the features. The aforementioned artifacts in ASTER DEM will also inherent to their derivatives^{61}.
Execution time of different susceptibility assessments
A random sampling of nonlandslides points from the overall study area carry some artifacts and randomness to the evaluation process and that randomness can vary in size and effect. This drawback is one of the disadvantages of LSM using ML modeling, and efficiently eliminating those artifacts and randomness is nearly impossible. To overcome such drawback, machine learning needs to have decent performance with less computational time (i.e., execution time).
The results of computational time required for each model (training and testing the final models excluding the time spent on tuning the hyperparameters) and each dataset (Fig. 12) shows that SVM models are at least 50% faster than RF, and LR models are 50 times faster than SVM. Besides, 5 meters Lidar DEM based models required a relatively close computational time to LSM models based of 30 meters (i.e., LR and SVM), except for RF and LiDAR 30 meters dataset slightly require less computational time for LR and SVM models compared to the result of datasets. Note that the preprocessing time for deriving topographical variables from 5 m DEM is much larger than the 30 m DEMs. Therefore, the overall performance results that reported in Tables 8, 9 and Figs 8–12, when combined with the computational process, it is obvious that resampling the LiDAR dataset (i.e. from the original 5 meters to 30 meters) with LR and/or RF models combination would be “Go To” solution as they provide decent results. However, it is widely accepted that no single or particular model can be depicted as the most suitable for all case scenarios, as it depends on the subjective opinion of the decisionmaker of whether the more accurate results matter more than the computational time or viceversa. After all, recent studies^{13,62} suggest that a rather fast and simple model, such as SVM would be much better than an advanced machine learning models like RF, if the consideration was not solely based on the overall performance but on balance of overall performance and the computational time. For instance, SVMs are useful nonlinear classifiers whose goal is not only to classify landslide instances correctly but also to keep the distance between instances and keep the separation of the hyperplane at a maximum. On the other hand, RF models offer an excellent performance with decent interpretability and moderate number of hyperparameters to tune in but require a considerable time budget (they require a lot of time to converge especially if used on largescale analyses) compared to LR models which are the opposite of being simple, fast, easy to implement, and only able to capture the linear relationship between the causative factors and the landslide susceptibility which translate into poor performance. This makes SVM models appealing for susceptibility evaluation considering the number of hyperparameters to tune in. However, if those hyperparameters are inappropriately set, SVM will often lead to unsatisfactory results^{3,13}. Though the computational performance for all the models in this study was quick (i.e., <3 minutes), the aforementioned analysis and discussions will be helpful while dealing with a larger amount of data in the machine learning environment.
Summary and Conclusions
This paper conducts the scale dependency of DEM data in the analysis of landslide susceptibilities. The study area is characterized by steep slopes with frequent debris flows and landslides in the typhoon seasons. The LiDAR DEMs provided unprecedented highquality terrain data for detailed topographic representations. This study tested the appropriateness of such high accurate grid sizes in the susceptibility studies. The obtained results highlight that a fine resolution DEM not necessarily produce an accurate LSM as they found to be carrying excessive information. These results are in line with the findings of some previous studies^{29,30,31}. The results prove that entailing different DEM scales introduced different results for the same models. A 30meter resolution DEM depicting accurate topography could be plausible for LSM as they produced decent levels of generalization of the topography. In fact, higher resolution DEMs introduce more noise, which makes the model perform worse than it supposed to be. Entailing highresolution DEMs (5 meters Lidar) have proven to be hindered on susceptibility models as they feed a steady flow of data 36 times more than 30 meters DEMs which are supposed to theoretically produce better models. However, in reality, the data flow was treated as noise that worsens the overall resulting models instead of enhancing it, which prove that a generalized DEMs of 30 meters used for DEMderived condition factors is much valuable than their 5 meters counterpart. Additionally, inappropriate spatial resolution increases the preprocessing time. For this reason, it is suggested that an analysis should be performed to understand the scale effects of topographic variables on landslide susceptibility mapping. Our results also indicate that the scale effects of topographic variables are mainly caused by the resolution impact on topographic parameter derivation, while factors such as geology and rainfall are insensitive to resolutions. For susceptibility mapping, RF models are found to be the best model in term of performance for the study area, while SVM is more suitable in the decisionmaking process when looking for a balanced LSM model between computational time and overall performance.
Further research is required to test variation over a more continuous range of resolutions (e.g. 10 m, 15 m, and 20 m) in more case studies for reducing some uncertainties behind the obtained results. Also, to enhance the results, deep learning techniques such as convolutional neural network and testing other machine learning models are recommended. The obtained landslide susceptibility maps are based on present and past landslides. However, Future landslides are not foreseeable, and thus the obtained LSM models are obsolete after a given period of time. Thus, the inventory and model should be updated constantly.
References
Petley, D. Global patterns of loss of life from landslides. Geology 40, 927–930 (2012).
Dou, J. et al. Automatic CaseBased Reasoning Approach for Landslide Detection: Integration of ObjectOriented Image Analysis and a Genetic Algorithm. Remote Sens. 7, 4318–4342 (2015).
Merghadi, A., Abderrahmane, B. & Tien Bui, D. Landslide Susceptibility Assessment at Mila Basin (Algeria): A Comparative Assessment of Prediction Capability of Advanced Machine Learning Methods. ISPRS Int. J. GeoInformation 7 (2018).
Chen, W. et al. Spatial prediction of groundwater potentiality using ANFIS ensembled with teachinglearningbased and biogeographybased optimization. J. Hydrol, https://doi.org/10.1016/j.jhydrol.2019.03.013 (2019).
Chen, W. et al. Novel Hybrid Integration Approach of BaggingBased Fisher’s Linear Discriminant Function for Groundwater Potential Analysis. Nat. Resour. Res, https://doi.org/10.1007/s1105301909465w (2019).
Chen, W. et al. Novel hybrid artificial intelligence approach of bivariate statisticalmethodsbased kernel logistic regression classifier for landslide susceptibility modeling. Bulletin of Engineering Geology and the Environment, https://doi.org/10.1007/s1006401814018 (2018).
Chen, W. et al. Landslide Susceptibility Modeling Based on GIS and Novel BaggingBased Kernel Logistic Regression. Appl. Sci, https://doi.org/10.3390/app8122540 (2018).
Pham, T. B. et al. A Novel Hybrid Approach of Landslide Susceptibility Modeling Using Rotation Forest Ensemble and Different Base Classifiers. Geocarto Int. 1–38, https://doi.org/10.1080/10106049.2018.1559885 (2018).
Dou, J. et al. Assessment of advanced random forest and decision tree algorithms for modeling rainfallinduced landslide susceptibility in the IzuOshima Volcanic Island, Japan. Sci. Total Environ. 662, 332–346 (2019).
Hong, H., Miao, Y., Liu, J. & Zhu, A.X. Exploring the effects of the design and quantity of absence data on the performance of random forestbased landslide susceptibility mapping. CATENA 176, 45–64 (2019).
Hong, H. et al. Landslide susceptibility assessment at the Wuning area, China: a comparison between multicriteria decision making, bivariate statistical and machine learning methods. Natural Hazards, https://doi.org/10.1007/s1106901835360 (Springer Netherlands, 2018).
Hong, H., Pourghasemi, H. R. & Pourtaghi, Z. S. Landslide susceptibility assessment in Lianhua County (China): A comparison between a random forest data mining technique and bivariate and multivariate statistical models. Geomorphology, https://doi.org/10.1016/j.geomorph.2016.02.012 (2016).
Dou, J. et al. Evaluating GISBased Multiple Statistical Models and Data Mining for Earthquake and RainfallInduced Landslide Susceptibility Using the LiDAR DEM. Remote Sens. 11, 638 (2019).
Ayalew, L., Yamagishi, H., Marui, H. & Kanno, T. Landslides in Sado Island of Japan: Part I. Case studies, monitoring techniques and environmental considerations. Eng. Geol. 81, 419–431 (2005).
Dou, J. et al. GISBased Landslide Susceptibility Mapping Using a Certainty Factor Model and Its Validation in the Chuetsu Area, Central Japan. In Landslide Science for a Safer Geoenvironment 419–424, https://doi.org/10.1007/9783319050508_65 (Springer International Publishing, 2014).
Guo, C., Montgomery, D. R., Zhang, Y., Wang, K. & Yang, Z. Quantitative assessment of landslide susceptibility along the Xianshuihe fault zone, Tibetan Plateau, China. Geomorphology 248, 93–110 (2015).
Hong, H., Naghibi, S. A., Pourghasemi, H. R. & Pradhan, B. GISbased landslide spatial modeling in Ganzhou City, China. Arab. J. Geosci, https://doi.org/10.1007/s125170152094y (2016).
Budimir, M. E. A., Atkinson, P. M. & Lewis, H. G. A systematic review of landslide probability mapping using logistic regression. Landslides, https://doi.org/10.1007/s1034601405505 (2015).
Dou, J. et al. Optimization of causative factors for landslide susceptibility evaluation using remote sensing and GIS data in parts of Niigata, Japan. PLoS One 10, e0133262 (2015).
Reichenbach, P., Rossi, M., Malamud, B. D., Mihir, M. & Guzzetti, F. A review of statisticallybased landslide susceptibility models. EarthScience Reviews 180, 60–91 (2018).
Dou, J. et al. An integrated artificial neural network model for the landslide susceptibility assessment of Osado Island, Japan. Nat. Hazards 78, 1749–1776 (2015).
Le, L. T., Nguyen, H., Zhou, J., Dou, J. & Moayedi, H. Estimating the Heating Load of Buildings for Smart City Planning Using a Novel Artificial Intelligence Technique PSOXGBoost. Appl. Sci. 9, 2714 (2019).
Le, L. T., Nguyen, H., Dou, J. & Zhou, J. A Comparative Study of PSOANN, GAANN, ICAANN, and ABCANN in Estimating the Heating Load of Buildings’ Energy Efficiency for Smart City Planning. Appl. Sci. 9, 2630 (2019).
Yunus, A. P., Dou, J., Song, X. & Avtar, R. Improved Bathymetric Mapping of Coastal and Lake Environments Using Sentinel2 and Landsat8 Images. Sensors 19, 2788 (2019).
Khosravi, K. et al. A comparative assessment of flood susceptibility modeling using MultiCriteria DecisionMaking Analysis and Machine Learning Methods. J. Hydrol. 573, 311–323 (2019).
Süzen, M. L. & Kaya, B. Ş. Evaluation of environmental parameters in logistic regression models for landslide susceptibility mapping. Int. J. Digit. Earth 5, 338–355 (2012).
Arnone, E., Francipane, A., Scarbaci, A., Puglisi, C. & Noto, L. V. Effect of raster resolution and polygonconversion algorithm on landslide susceptibility mapping. Environ. Model. Softw. 84, 467–481 (2016).
Guzzetti, F., Carrara, A., Cardinali, M. & Reichenbach, P. Landslide hazard evaluation: a review of current techniques and their application in a multiscale study, Central Italy. Geomorphology 31, 181–216 (1999).
Paudel, U., Oguchi, T. & Hayakawa, Y. MultiResolution Landslide Susceptibility Analysis Using a DEM and Random Forest. Int. J. Geosci. 07, 726–743 (2016).
Tian, Y., Xiao, C., Liu, Y. & Wu, L. Effects of raster resolution on landslide susceptibility mapping: A case study of Shenzhen. Sci. China, Ser. E Technol. Sci. 51, 188–198 (2008).
Manzo, G., Tofani, V., Segoni, S., Battistini, A. & Catani, F. GIS techniques for regionalscale landslide susceptibility assessment: the Sicily (Italy) case study. Int. J. Geogr. Inf. Sci. 27, 1433–1452 (2013).
Lin, E.J., Liu, C.C., Chang, C.H., Cheng, I.F. & Ko, M.H. Using the FORMOSAT2 High Spatial and Temporal Resolution Multispectral Image for Analysis and Interpretation Landslide Disasters in Taiwan. J. Photogramm. Remote Sens. 17, 31–51 (2013).
Jeremy Shen. The Key to Access Geospatial Open Data in Taiwan: TGOS GIS Cloud. Available at, http://www.supergeotek.com/index.php/201512_cs_tgos02/# (2014).
Dou, J., Paudel, U., Oguchi, T., Uchiyama, S. & Hayakawa, Y. S. Shallow and DeepSeated Landslide Differentiation Using Support Vector Machines: A Case Study of the Chuetsu Area, Japan. Terr. Atmos. Ocean. Sci. 26, 227 (2015).
Yunus, A. P., Dou, J., Song, X. & Avtar, R. Improved Bathymetric Mapping of Coastal and Lake Environments Using Sentinel2 and Landsat8 Images. Sensors 19, 2788 (2019).
Cawley, G. C. & Talbot, N. L. C. Gene selection in cancer classification using sparse logistic regression with Bayesian regularization. Bioinformatics, https://doi.org/10.1093/bioinformatics/btl386 (2006).
Dou, J. et al. Torrential rainfalltriggered shallow landslide characteristics and susceptibility assessment using ensemble datadriven models in the Dongjiang Reservoir Watershed, China. Nat. Hazards. https://doi.org/10.1007/s11069019036594 (2019).
Vapnik, V. N. Statistical Learning Theory (Adaptive and Learning Systems for Signal Processing, Communications and Control Series). WileyInterscience, Chichester (WileyInterscience, 1998).
Hong, H. et al. Application of fuzzy weight of evidence and data mining techniques in construction of flood susceptibility map of Poyang County, China. Sci. Total Environ, https://doi.org/10.1016/j.scitotenv.2017.12.256 (2018).
Kim, K. Financial time series forecasting using support vector machines. Neurocomputing 55, 307–319 (2003).
Breiman, L. E. O. Random Forest. Mach. Learn. 5–32, https://doi.org/10.1023/A:1010933404 (2001).
RodriguezGaliano, V. F., Ghimire, B., Rogan, J., ChicaOlmo, M. & RigolSanchez, J. P. An assessment of the effectiveness of a random forest classifier for landcover classification. ISPRS J. Photogramm. Remote Sens, https://doi.org/10.1016/j.isprsjprs.2011.11.002 (2012).
Liaw, A. & Wiener, M. Classification and Regression by randomForest. R news, https://doi.org/10.1177/154405910408300516 (2002).
Chung, C.J. & Fabbri, A. G. The representation of geoscience information for data integration. Nonrenewable Resour. 2, 122–139 (1993).
Pfeiffer, R. M., Molinaro, A. M. & Simon, R. Prediction error estimation: a comparison of resampling methods. Bioinformatics 21, 3301–3307 (2005).
Bergstra, J., Yamins, D. & Cox, D. D. Making a Science of Model Search: Hyperparameter Optimization in Hundreds of Dimensions for Vision Architectures. ICML 28, 115–123 (2013).
Cohen, J. A Coefficient of Agreement for Nominal Scales. Educ. Psychol. Meas, https://doi.org/10.1177/001316446002000104 (1960).
Friedman, M. The Use of Ranks to Avoid the Assumption of Normality Implicit in the Analysis of Variance. J. Am. Stat. Assoc. https://doi.org/10.1080/01621459.1937.10503522 (1937).
MartínezÁlvarez, F., Reyes, J., MoralesEsteban, A. & RubioEscudero, C. Determining the best set of seismicity indicators to predict earthquakes. Two case studies: Chile and the Iberian Peninsula. KnowledgeBased Syst. 50, 198–210 (2013).
Zhu, Z., Wang, H., Peng, D. & Dou, J. Modelling the Hindered Settling Velocity of a Falling Particle in a ParticleFluid Mixture by the Tsallis Entropy Theory. Entropy 21, 55 (2019).
Zhang, W. & Montgomery, D. R. Digital elevation model grid size, landscape representation, and hydrologic simulations. Water Resour. Res, https://doi.org/10.1029/93WR03553 (1994).
Dai, W. et al. Effects of DEM resolution on the accuracy of gully maps in loess hilly areas. CATENA 177, 114–125 (2019).
Catani, F., Lagomarsino, D., Segoni, S. & Tofani, V. Landslide susceptibility estimation by random forests technique: sensitivity and scaling issues. Nat. Hazards Earth Syst. Sci. 13, 2815–2831 (2013).
Deng, Y., Wilson, J. P. & Bauer, B. O. DEM resolution dependencies of terrain attributes across a landscape. Int. J. Geogr. Inf. Sci. 21, 187–213 (2007).
Tarolli, P. & Dalla Fontana, G. Hillslopetovalley transition morphology: New opportunities from high resolution DTMs. Geomorphology 113, 47–56 (2009).
Penížek, V., Zádorová, T., Kodešová, R. & Vaněk, A. Influence of Elevation Data Resolution on Spatial Prediction of Colluvial Soils in a Luvisol Region. PLoS One 11, e0165699 (2016).
Penna, D., Borga, M., Aronica, G. T., Brigandì, G. & Tarolli, P. The influence of grid resolution on the prediction of natural and roadrelated shallow landslides. Hydrol. Earth Syst. Sci. 18, 2127–2139 (2014).
Hayakawa, Y. S., Oguchi, T. & Lin, Z. Comparison of new and existing global digital elevation models: ASTER GDEM and SRTM3. Geophys. Res. Lett, https://doi.org/10.1029/2008GL035036 (2008).
Tetsushi Tachikawa. ASTER Global Digital Elevation Model Version 2 – Summary of Validation Results. Japan Sp. Syst, https://doi.org/10.1017/CBO9781107415324.004 (2011).
Avtar, R., Yunus, A. P., Kraines, S. & Yamamuro, M. Evaluation of DEM generation based on Interferometric SAR using TanDEMX data in Tokyo. Phys. Chem. Earth, https://doi.org/10.1016/j.pce.2015.07.007 (2015).
Fisher, P. F. & Tate, N. J. Causes and consequences of error in digital elevation models. Prog. Phys. Geogr. 30, 467–489 (2006).
Camilo, D. C., Lombardo, L., Mai, P. M., Dou, J. & Huser, R. Handling high predictor dimensionality in slopeunitbased landslide susceptibility models through LASSOpenalized Generalized Linear Model. Environ. Model. Softw. 97, 145–156 (2017).
Wright, M. N. & Ziegler, A. ranger: A fast implementation of random forests for high dimensional data in C++ and R. J. Stat. Softw. 77, 1–17 (2015).
Meyer, D., Dimitriadou, E., Hornik, K., Weingessel, A. & Leisch, F. e1071: Misc Functions of the Department of Statistics, Probability Theory Group (Formerly: E1071), TU Wien. (2017).
Acknowledgements
This research is partially supported by Japan Society for the Promotion of Science. ASTER data are distributed by the Land Processes Distributed Active Archive Center (LP DAAC), located at the U.S. Geological Survey (USGS) Center for Earth Resources Observation and Science (EROS) http://LPDAAC.usgs.gov.
Author information
Authors and Affiliations
Contributions
J.D. and K.T.C. led the research, formulated the research questions and defined the manuscript contents in line with discussions with A.P.Y. J.D. coordinated and prepared the manuscript, with contributions from K.T.C and A.P.Y.; K.T.C. together with J.D., A.M., and B.T.P. analyzed the background literature, controlling factors and performed the susceptibility assessments; J.D., KTC, A.P.Y. A.M. and B.T.P. contributed to the preparation of figures items at various stages. All authors revised and approved the manuscript.
Corresponding author
Ethics declarations
Competing Interests
The authors declare no competing interests.
Additional information
Publisher’s note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Chang, KT., Merghadi, A., Yunus, A.P. et al. Evaluating scale effects of topographic variables in landslide susceptibility models using GISbased machine learning techniques. Sci Rep 9, 12296 (2019). https://doi.org/10.1038/s41598019487732
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598019487732
This article is cited by

Landslide type inference based on statistical analysis of a highresolution digital elevation model in Gorce National Park, Poland
Scientific Reports (2024)

Assessing landslide susceptibility using improved machine learning methods and considering spatial heterogeneity for the Three Gorges Reservoir Area, China
Natural Hazards (2024)

Study on the influence of input variables on the supervised machine learning model for landslide susceptibility mapping
Environmental Earth Sciences (2024)

Review on the progress and future prospects of geological disasters prediction in the era of artificial intelligence
Natural Hazards (2024)

Knowledge discovery in weather forecasting: mining fuzzy image association rules with finetuned CNN and fuzzy HIFP algorithm
Evolving Systems (2024)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.