Abstract
Rockburst is a severe geological hazard that restricts deep mine operations and tunnel constructions. To overcome the shortcomings of widely used algorithms in rockburst prediction, this study investigates the ensemble trees, i.e., random forest (RF), extremely randomized tree (ET), adaptive boosting machine (AdaBoost), gradient boosting machine, extreme gradient boosting machine (XGBoost), light gradient boosting machine, and category gradient boosting machine, for rockburst estimation based on 314 real rockburst cases. Additionally, Bayesian optimization is utilized to optimize these ensemble trees. To improve performance, three combination strategies, voting, bagging, and stacking, are adopted to combine multiple models according to training accuracy. ET and XGBoost receive the best capabilities (85.71% testing accuracy) in single models, and except for AdaBoost, six ensemble trees have high accuracy and can effectively foretell strong rockburst to prevent large-scale underground disasters. The combination models generated by voting, bagging, and stacking perform better than single models, and the voting 2 model that combines XGBoost, ET, and RF with simple soft voting, is the most outstanding (88.89% testing accuracy). The performed sensitivity analysis confirms that the voting 2 model has better robustness than single models and has remarkable adaptation and superiority when input parameters vary or miss, and it has more power to deal with complex and variable engineering environments. Eventually, the rockburst cases in Sanshandao Gold Mine, China, were investigated, and these data verify the practicability of voting 2 in field rockburst prediction.
Similar content being viewed by others
Introduction
Rockburst is a geological calamity often confronted in deep mine operations or deep tunnel excavations, and it has the manners of rock breaking and the sudden release of energy from wall rock1. The occurrence of rockburst is generally relevant to lithology, geological structure, surrounding rock mass properties, terrain, and etc. Rockburst, which occurs in many countries2,3,4,5, is considered a severe danger to the security of employees and equipments in underground construction. Rockburst is a “cancer” in deep mines6, killing many South African gold mine employees7. With more and more constructions in underground excavations, efficient prediction and prevention of rockburst have become an increasingly crucial topic.
According to Russnes’s method8, the rockburst intensity can be classified into four levels (i.e., none, light, moderate and strong). The nature of rockburst is complex and nonlinear, and it is a big challenge to predict rockburst. Numerous technologies have been put forward to evaluate rockburst in the last few decades. These methods include empirical methods, numerical simulation, experimental methods, and intelligent algorithms3, 9, 10.
The empirical methods are often applied in the trial implementation phase of underground constructions, including single and multi-index indicators. The single indicators include stress index, energy index, brittleness index, depth index, and so on3. The multi-index indicators utilize mathematical methods or other methods to combine the significant factors that are able to control rockburst. The empirical methods are simple and easy to implement. However, they are poorly applicable and only effective in a specific area. Jing et al.11 introduced a lot of numerical simulation experiments on rockburst prediction. Wen et al.12 applied strain energy density to simulate and investigate the rockburst mechanism. Chen et al.13 utilized discontinuity deformation methodology to assess rockburst. Numerical simulation can reveal the failure process of rock14. Nevertheless, it is sensitive to input parameters and hard to simulate the dynamic behavior of rockburst. Moreover, the constitutive model in the numerical simulation may not demonstrate the real propriety of the rock. Gong et al.15 researched the rockburst tendency of red sandstone by rock experiments. He et al.16 adopted indoor experimental methods to study and classify rockburst. Rock mechanics experiments can give some essential information, which is beneficial to study rock properties17. However, they are challenging to reproduce real engineering environment and are limited by monitoring and measurement techniques. The empirical method has a narrow scope of application18. Numerical simulation for rockburst prediction has high requirements on simulation methods, mechanical constitutive model, and rockburst mechanism10. The rock mechanics test for evaluating rockburst requires samples preparation and adequate types of equipment19, which is expensive and time consuming. In contrast, the intelligent algorithm is low cost, only focuses on input and output parameters, and has wider applicability4. The intelligent algorithm is more worthy for rockburst prediction efficiently and timely with the growing development of big data and artificial intelligence.
Since Feng et al.20 utilized neural networks to predict rockburst, many intelligent algorithms have been applied. Table 1 summarizes intelligent algorithms for rockburst prediction in recent years. Each intelligent algorithm has its advantages for specific problems. However, any of these algorithms cannot be perfectly performed in all problems according to the ‘No Free Lunch theory’. There are inevitably some disadvantages in each intelligent algorithm when applied in practical engineering. The discriminant analysis21 and logistic regression22 are simple and easy to interpret. However, they cannot be applied to complex problems and high-dimensional data. Decision trees23,24,25 can be used for data with missing values, but they tend to overfit. Support vector machine23, 26, 27 has a solid theoretical basis, and it is not easy to overfit. However, it performs poorly in multiple classification problems28. The k-nearest neighbor is efficient and straightforward23, 29, but it is sensitive to irrelevant features. Bayes model23, 30 is simple and fast in the calculation. However, it requires that features are independent distribution, which is difficult to satisfy in practice. Although neural networks23, 29, 31, 32 can deal with more complex problems, they have many hyperparameters to be turned33.
The single model has low robustness, cannot get the optimal solution for all problems, and its performance changes with the variation of engineering environment or input parameters. Accordingly, scholars have attempted to adopt ensemble models to combine multiple models to overcome the shortcomings of a single model. Nonetheless, there are only a few studies in the area of rockburst. Moreover, there is no detailed research on the selection and application of ensemble models in rockburst prediction. To fill the gaps, the present study considers seven models based on decision trees and three combination strategies for rockburst estimation in complex and variable engineering conditions. The seven models include random forest (RF), extremely randomized tree (ET), adaptive boosting machine (AdaBoost), gradient boosting machine (GBM), extreme gradient boosting machine (XGBoost), light gradient boosting machine (LighGBM), and category gradient boosting machine (CatBoost), which all adopt decision trees (DTs) as the basic classifier due to the low bias and high variance of DTs39. Three combination strategies are voting, bagging and stacking. These seven models have good performance in machine learning (ML) tasks, but there are no detailed investigations of applying them to rockburst. Furthermore, applying combination strategies to combine multiple single models can make the rockburst models more robust and powerful. Apart from that, Bayesian optimization is implemented to optimize these models. It is significant to note that Bayesian as a highly efficient optimization model, has been widely used in hyperparameter optimization of ML area40.
The rest of this study is organized as follows: “Methodology” section describes the techniques and the data from real cases for simulation. Section “Simulation” presents the model parameter optimization and combination, which exhibits the process of selection and integration of the base classifier in detail. In “Results and discussion” section, all models are evaluated to select an optimal model. Moreover, the selected model conducts the sensitivity analysis and is tested for engineering practicability.
Methodology
Ensemble trees
Random forest and extremely randomized tree
RF is the ML model composed of \(K\) decision trees. The process to construct the RF is shown in Fig. 1. ET is similar to RF, and the main differences between them are as follows: first, RF uses bootstrap sampling to build a random sample subset, while ET utilizes all original samples, which can reduce the deviation. Secondly, the choice of the split point is different. The RF selects the optimal split point, while the ET randomly chooses the split point, which can reduce the variance. The choice of random split point adds more randomness to the model and speeds up the calculation speed.
In Eqs. 1 and 2, \(D\) represents the dataset, \(\left| y \right|\) is the class number, \(p\) is the proportion of each class to the total dataset, \(Gini(D^{V} )\) is the Gini value of the class \(V\), \(\left| D \right|\) represents the number of instances, \(\left| {D^{V} } \right|\) represents the number of instances of the class \(V\), and \(a\) represents the feature that needs to be divided.
Boosting model
Boosting model sequentially combines multiple poor learners to build a robust model. The steps to develop the boosting model are presented in Fig. 2. AdaBoost constructs many poor learners from the training data, then linearly synthesizes them into a strong model41. Compared with AdaBoost, GBM is a more robust model, which can optimize any differentiable loss function42. XGBoost43, LightGBM44, and CatBoost45 are the development and extension of GBM. The detailed comparison between these three models can be referred to previous investigations46.
Combination strategy
This study uses three combination strategies: voting, stacking and bagging to combine multiple models. Figure 4 displays the combination process. Voting is a commonly used method to combine the output of multiple classifiers. In this research, simple soft voting is adopted. Individual classifier \(h_{i}\) outputs a \(l\) dimensionality vector \((h_{i}^{1} (x),...,h_{i}^{l} (x))^{T}\) when inputting sample \(x\). The simple soft voting method calculates and outputs the average value of the output of each classifier.
Bagging uses bootstrap sampling to generate different base classifiers. Given a training dataset with m instances, a training subset of m instances can be obtained with the replacement sample. Some of the original instances are selected many times, and others are not selected. Repeating the process t times, t training subsets including m instances are obtained. Each training subset is used to develop a base classifier. Voting is adopted to aggregate t base classifiers in the classification task.
Stacking combines individual classifiers by training classifier, and the individual classifier is called first-level learner, and the connector is called second-level learner. Stacking first trains the first-level learners utilizing the initial database, then forms a new database to train the second-level learners using the outcomes of the first-level learners as input features and the corresponding initial markers as new markers.
Bayesian optimization
Bayesian optimization (BO) is suitable for complex problems whose objective function cannot be expressed47. BO chooses the next estimation points according to previous outcomes. BO consists of the surrogate model and acquisition function47. The goal of the surrogate model is to match the detected points into the objective function. The acquisition function decides to use different points by balancing exploration and exploitation. The Bayesian model can discover the most likely optimum area for the present and avoid missing better parameters in unknown areas.
Gaussian process regression is often chosen as the surrogate model in BO40. The acquisition functions include the probability of improvement48, expected improvement49, 50, and upper/lower confidence bound (UCB/LCB)51. To match the acquisition function to the surrogate model, GP-Hedge is introduced to select an appropriate acquisition function in each BO iteration51, 52.
Data
A database including 314 real rockburst cases is established and used for modeling. Table 2 lists different sources of this database. The maximum tangential stress (\({\sigma }_{\theta }\)), the uniaxial compressive strength (\({\sigma }_{c}\)), the tensile strength (\({\sigma }_{t}\)), the stress ratio (\({\sigma }_{\theta }/{\sigma }_{c}\)), the brittleness ratio (\({\sigma }_{c}/{\sigma }_{t}\)), and the elastic strain energy index (\({W}_{et}\)) are selected as the input variables in this study by referring to the previous research31, 34, 53. Pearson correlation coefficients (Eq. 3) between the six variables are calculated. Table 3 shows correlation coefficients between variables. Figure 3 displays the statistics and distribution of each variable.
According to Fig. 4, the database is split into a training set (80%) and a test set (20%). The training set is employed to construct seven models based on trees. fivefold cross-validation is employed for model selection. BO is utilized to optimize the hyperparameters of models. The voting, bagging, and stacking strategies are applied to combine these optimized models to develop ensemble models in predicting/evaluating rockburst. The test set is implemented to assess the capability of models. Finally, the optimal model is utilized to conduct the sensitivity analysis and how it can be applied to engineering projects.
Simulation
Model metrics
Accuracy was applied to estimate the global performance of the model. The \(F_{1}\) combined precision and recall and was utilized to assess the performance of each classification.
In Eq. 4, \(m\) is the number of samples, \(\overline{{y_{i} }}\) represents the predicted labels, \(y_{i}\) represents the actual labels, and \(I( \cdot )\) is one if the conditions in brackets are true and zero, otherwise. In Eqs. 5 and 6, TP is the true positive, FP is the false positive, and FN is the false negative.
Model hyperparameters optimization
The hyperparameters range
The training set was adopted to train the seven ensemble models based on DTs. Z-score was used to process the input variables (Eq. 8). The open-source Python library, Scikit-learn59, was used to construct RF, ET, AdaBoost, and GBM models. The open-source Python libraries, XGBoost43, LightGBM44, and CatBoost45, were utilized to build the XGBoost, LightGBM, and CatBoost models, respectively. Table 4 presents the hyperparameters optimization range in seven models.
In Eq. 8, \(\overline{x}\) is the mean value of the data and \(\sigma\) is the standard deviation of the data.
The objective function
Before hyperparameters optimization, the objective function should be defined. In ML, the cross-entropy loss function is a method to measure classifier performance (Eq. 9). It is generally believed that the classifier performs better when the cross-entropy loss function obtains a smaller value. In this paper, we adopted the cross-entropy loss function in fivefold cross-validation as the objective function. Figure 5 shows the steps to calculate the objective function.
In Eq. 9, \(m\) is the number of instances and \(p_{\bmod el} [y_{i} \in C_{{y_{i} }} ]\) is the prediction probability of the model in the actual label.
The process of BO
In this research, the Scikit-Optimize 61 was used to perform the BO. The surrogate model in BO adopted the Gaussian process (GP) regression, the acquisition function utilized the GP-Hedge, and the noise was assumed to be Gaussian distribution. The kernel function was an important part of the GP regression. Table 5 tabulates the kernel function parameters of GP regression. Figure 6 illustrates the process that BO optimized the hyperparameters. In this research, the iteration N was set to 50. BO can minimize the objective function within the parameter range so that the performance of the model can reach optimum. In addition, Fig. 7 presents the objective function convergence of seven models in 50 iterations. It reflects the variation of the objective function with the iteration process. Different models had different values of the objective function in the initial state, which was related to the random selection of the initial point in BO. With the iteration progress, BO was constantly balancing the process of exploration and utilization, and the value of the objective function was shrinking. After 50 iterations, BO can find the minimum value of the objective function and return the optimum value of hyperparameters. Table 6 shows the optimized parameter value and the training accuracy in seven models. The training accuracies in the seven models varied greatly. XGBoost had the highest training accuracy, and AdaBoost had the worst training performance.
Nu controlling the smoothness of the learned function.
Model combination
Voting combination
According to the accuracy results in the training set (Fig. 8), multiple models were combined by the simple soft voting method. From XGBoost to AdaBoost, the model was added to the voting combination model in order of accuracy in the training set. Table 7 presents the final six voting combination models. It can be seen that with the addition of some models with lower training accuracy, the training accuracy of the voting combination model was gradually decreasing.
Bagging combination
The seven models were used as base classifiers in bagging ensemble models. Bagging fitted each base classifier on a random subset of the initial training set and then combined their prediction outcomes by voting to build an eventual ensemble model. The number of base estimators in the bagging ensemble model was set to 10. Table 8 displays the final seven bagging combination models. Except for the AdaBoost model, the training accuracies of other models that adopted the bagging combination were reduced.
Stacking combination
In stacking combination, we adopted the seven models as the first-level learners, and the second-level learners adopted LR. Like “Voting combination” section, voting combinations, multiple models were combined in turn among first-level learners based on the performance in the training set. Table 9 displays the final seven stacking combination models.
Results and discussion
Model performance evaluation
The individual model performance evaluation and comparison
The test set is applied for evaluating the seven base models. Table 10 presents the \(F_{1}\) and accuracy of the test set in seven base models. In the individual model, ET and XGBoost perform best, and AdaBoost performs worst. When considering the accuracy in the test set, it can be concluded that the capacity ranking is ET, XGBoost > RF > GBM > CatBoost > LightGBM > AdaBoost. Besides, apart from AdaBoost, these single models have high \(F_{1}\) in strong rockburst, and these suggest that ensemble trees have superior capability to forecast massive rockburst hazards.
Six other widely used ML models, LR, SVM, KNN, ANN, DT, and Naive Bayes, are also developed based on the training set and evaluated by the test set. Their hyperparameters adopt the default value in Scikit-learn. Figure 9 shows the performance comparison of ensemble trees and other ML models. DT model suffers from serious overfitting, and the ensemble trees have better generalization than DT. Except for AdaBoost, the ensemble trees have higher testing accuracy than other ML models. These indicate that the proposed ensemble trees solution can get more accurate rockburst prediction results.
Investigate the strength of voting combination models
Due to the difficulty of selecting base learners, not all voting combination models improve performance compared to their base classifiers. Therefore, the test set is employed to assess the capability of the voting combination models. Table 11 shows the \(F_{1}\) and accuracy of the test set in six voting combination models. It is found that the voting 2 has an outstanding \(F_{1}\) in a single type of rockburst prediction and the highest accuracy. Besides, six voting combination models have the same performance in predicting strong rockburst. Figure 10 presents the accuracy improvement of the voting combination model in the test set compared to the individual optimal classifier. Voting 1, voting 2, and voting 3 perform better than the individual model on the testing set. The performances of the three models increase by 1.59%, 3.18%, and 1.59%, respectively. With the combination of poor performance base learners, voting 4, voting 5, and voting 6 perform poorly than the single optimal model. Figure 11 compares the \(F_{1}\) of the voting 2 model and its base classifiers. It can be seen that voting 2 has better performance than XGBoost, ET, and RF models in the single rockburst category. The results suggest that the voting 2 combining high accuracy and diversity is the best in voting combination models.
The effect of bagging integration on single model performance
Bagging combines independent base classifiers, which reduces the error. The bagging combination models are evaluated by the test set to determine the enhancement of models (i.e., RF, ET, etc.) performance before and after bagging integration. Table 12 presents the \(F_{1}\) and accuracy of the test set in the seven bagging combination models. Figure 12 displays the accuracy improvement of the bagging combination model in the test set compared to the individual classifier. After the bagging combination, except for the ET model, the accuracies in the test set of other models increase, and the accuracy of the AdaBoost increases by 23.82%. XGBoost, GBM, and LightGBM achieve the best performance after adopting the bagging combination. Figure 13 compares the \(F_{1}\) of XGBoost, GBM, and LightGBM and their bagging combination models. The bagging 1 that adopts XGBoost as the base learner has great improvement for predicting the none intensity of rockburst. Bagging 4 and bagging 6 perform better than their base classifiers in the prediction of a single rockburst category.
Explore the power of the stacking combination models
Stacking is a learning combination method, and it is of importance to match the appropriate first-level learners (i.e., RF, ET, etc.) to second-level learner (i.e., LR). The performance of stacking combination models in the test set can reflect whether the model combination is appropriate. Table 13 shows the \(F_{1}\) and accuracy of the test set in seven stacking combination models. Stacking 5 is the optimal model in terms of accuracy and \(F_{1}\). Figure 14 illustrates the accuracy improvement of the stacking combination models in the test set. Compared with the previous two combination strategies, the rockburst prediction performance of the stacking combination with the individual classifier is not ideal, and only the stacking 5 performs better than the individual optimal classifier in the seven stacking combination models. Figure 15 compares the \(F_{1}\) of stacking 5 and its base classifiers. Stacking 5 has an improvement in predicting the rockburst of light intensity compared to its base classifiers. Contrasted to GBM, the performance of stacking 5 in the strong rockburst prediction is weakened. We assume that the reason for the poor performance of the stacking models might be that the LR does not validly match models based on DTs.
Summary
For this part, seven ensemble trees are evaluated and compared with other ML models. Except for AdaBoost, the proposed ensemble trees have superior rockburst estimation results than other ML models. The XGBoost and ET perform best in the single models, and the accuracies in the test set are 85.7%. In the voting combination models, the voting 2 consisting of XGBoost, ET, and RF, is the best, and the accuracy in the test set is obtained 88.89%. The bagging combination models which adopt XGBoost, CatBoost, and LightGBM as base classifiers are optimal and their accuracies in the test set are obtained as 87.30%. In stacking combination models, the stacking 5, which utilizes XGBoost, ET, RF, CatBoost, and GBM as first-level learners and LR as the second-level learner, has the most outstanding performance, and the accuracy in the test set is achieved as 87.30%. It is found that voting 2 is the best model for rockburst prediction in all proposed models.
Analysis of the adaptation and superiority of applying combination model
In the previous section, we find that voting 2 is the most excellent combination model, and XGBoost, ET, and RF are the best three single models. In this part, sensitivity analysis is conducted to determine the adaptation and superiority of applying the voting 2 model for rockburst cases. The permutation feature importance algorithms59 are introduced to discover the crucial input parameters affecting rockburst. The relative importance of input parameters in voting 2 and its base classifiers are calculated, as shown in Fig. 16a. Although the variables with less importance are different in the four models, the pivotal variables are consistent. The relative importance scores of input variables in the four models are averaged. Figure 16b displays the mean importance score of the six input variables. The importance ranking of parameters influencing the rockburst is \(W_{et}\) > \(\sigma_{\theta }\) > \(\sigma_{\theta } /\sigma_{c}\) > \(\sigma_{c} /\sigma_{t}\) > \(\sigma_{c}\) > \(\sigma_{t}\). The \(W_{et}\) is the most critical factor that affected the rockburst. Energy-absorbing bolts and pressure relief blasting can be implemented to absorb the strain energy in deep excavation engineering to prevent rockburst62.
To inspect the adaptation and superiority of voting 2, the number of input parameters are varied, and the performances in voting 2 and three base classifiers are recorded and compared. According to the importance of variables influencing rockburst, some variables are reduced based on the original training and test sets to generate five datasets. Table 14 lists the variations of input parameters and generated five datasets. The five datasets are used to train and evaluate voting 2, XGBoost, RF, and ET. Table 15 tabulates the training and test results of four models in six datasets with different input parameters. According to Fig. 17, with the change of input parameters, the performance of the voting 2 in the training set is close or better to the single optimal model. Depending on Fig. 18, it can be seen that the single models have great differences in capacities for predicting rockburst with the variation of input parameters. For instance, XGBoost performs best for rockburst evaluation in 6 input parameters, but in the absence of \({\sigma }_{c}\), the estimation results of XGBoost can be worse. On the contrary, although RF has optimal performance in only 3 and 2 input parameters, it performs worse than XGBoost with the increase of input parameters. As for ET, when only \({W}_{et}\) is available for evaluating rockburst, it is not a good model. In practical engineering, some input parameters are difficult to obtain or missing, and adopting the single model for rockburst prediction might lead to disappointing outcomes. By contrast, the voting 2 model always has the optimal capability in the test set with different input parameters and can deal with the variation or missing of input data.
A ranking system63 is introduced to evaluate the performance of four models in different training and test sets comprehensively, as shown in Table 15. The training and testing accuracies of four models in the same dataset are ranked. The higher the accuracy, the higher the ranking score. The total rank in a model is obtained by adding the ranks considering the whole six datasets. The final rank is the sum of total ranks in training and test sets. The voting 2 has the highest final rank, indicating that the combination model has the most remarkable capacity in training and testing phases with different input parameters. The results suggest that the voting 2 has better robustness than single models and can copy with polytropic engineering environments.
Engineering application
The Sanshandao Gold Mine is located in Shandong Province, China, as shown in Fig. 19. To meet the production needs, the production of the Sanshandao Gold Mine is going deeper strata. Under the deep and high-stress environment, rockburst is a geological hazard threatening mine production. Figure 20 shows some rockburst sites in Sanshandao Gold Mine. To carry out the rockburst assessment, eight groups of rock specimens from eight locations in the Sanshandao Gold Mine were carried out in rock mechanics experiments. According to the test requirements, the rock samples were processed into two specifications of Φ 50 × 100 mm and Φ 50 × 25 mm, as shown in Fig. 21. The Brazilian splitting tensile tests were carried out by the INSTRON 1342 rock mechanics test system with the rock samples of Φ 50 × 25 mm. The uniaxial compression tests and the loading and unloading tests utilized the INSTRON 1346 rock mechanics test system with the rock samples of Φ 50 × 100 mm. Figure 22 illustrates these three rock mechanics experiments.
\(\sigma_{c}\), \(\sigma_{t}\), and \(W_{et}\) were obtained by rock mechanics experiments, and \(\sigma_{\theta }\) was calculated according to the stress of the surrounding rock. Through field observation and evaluation, the rockburst grade was obtained. Table 16 tabulates the rock mechanical parameters and rockburst grade in eight different regions.
To verify the practicability of the combination model, the voting 2 is applied to predict rockburst in Sanshandao Gold Mine. Meanwhile, four empirical criteria methods are used for rockburst prediction, as shown in Table 17. Additionally, Table 18 presents the rockburst prediction results. The voting 2 model has the best performance with 100% accuracy compared to the other methods, which suggests the combination model has superior engineering practicability.
Conclusion
-
1.
This study comprehensively introduced and evaluated the application of the seven ensemble trees in rockburst prediction. The performance ranking of seven models is XGBoost and ET > RF > GBM > CatBoost > LightGBM > AdaBoost. Except for AdaBoost, in the tree-based models, the testing accuracy ranges (76.2%, 85.71%) and \(F_{1}\) of strong rockburst ranges (0.86, 0.91). The ensemble trees have superior capacities than other ML models in general. Besides, the ensemble tree models are beneficial to forecast the occurrence of strong rockburst for protecting the safety of workers and facilities in underground engineering. Not only that, these tree-based models have fewer parameters to tune, and they are easy to apply to the field.
-
2.
To improve robustness and capability, three combination strategies, including voting, bagging, and stacking, were used to combine multiple models. The testing accuracy of voting combination models is range (80.95%, 88.89%), testing accuracy of bagging combination models is range (66.67%, 87.3%), and testing accuracy of stacking combination models is range (80.95%, 87.3%). The combination models have better capacity than single models, and they are suitable for huge and expensive projects that need to forecast rockburst precisely. It is worth noting that the voting 2 model, which adopts simple soft voting to combine XGBoost, RF, and ET, has an accuracy of 88.89% and is the most excellent in all models.
-
3.
Sensitivity analysis is applied to analyze the adaptation and strength of the voting 2 model compared to single models. The single model has different performances with different input parameters and is susceptible to the variation of parameters. In contrast, the combination model (i.e., voting 2) has better robustness and can receive the optimal capability when the input parameters vary. The results suggest that the combination model has better applicability for rockburst evaluation on-site when some parameters miss or are difficult to obtain.
-
4.
The real rockburst cases from Sanshandao Gold Mine, China, are measured and recorded. These datasets validate the practicability (100% accuracy) and advantage of the voting 2 model compared to empirical methods. Furthermore, the validation data can be employed to expand the rockburst database for building more strong models in the future.
-
5
The limitations of this study are that the performance of the stacking ensemble models could not achieve the desired effect. The second-level learner only considers the LR model in this paper, which is narrow, and it is necessary to explore appropriate second-level learners to match tree models in the future. It consumes more time and computing power to train the combination models than the single models, and fortunately, the limitation can be solved with the development of computation techniques.
References
Simser, B. Rockburst management in Canadian hard rock mines. J. Rock Mech. Geotech. Eng. 11, 1036–1043 (2019).
Keneti, A. & Sainsbury, B.-A. Review of published rockburst events and their contributing factors. Eng. Geol. 246, 361–373 (2018).
Zhou, J., Li, X. & Mitri, H. S. Evaluation method of rockburst: State-of-the-art literature review. Tunn. Undergr. Space Technol. 81, 632–659 (2018).
Pu, Y., Apel, D. B., Liu, V. & Mitri, H. Machine learning methods for rockburst prediction-state-of-the-art review. Int. J. Min. Sci. Technol. 29, 565–570 (2019).
Liang, W., Dai, B., Zhao, G. & Wu, H. A scientometric review on rockburst in hard rock: Two decades of review from 2000 to 2019. Geofluids 2020, 1–17 (2020).
Suorineni, F., Hebblewhite, B. & Saydam, S. Geomechanics challenges of contemporary deep mining: A suggested model for increasing future mining safety and productivity. J. S. Afr. Inst. Min. Metall. 114, 1023–1032 (2014).
Cai, M. Prediction and prevention of rockburst in metal mines—a case study of Sanshandao gold mine. J. Rock Mech. Geotech. Eng. 8, 204–211 (2016).
Russenes, B. Analysis of rock spalling for tunnels in steep valley sides. In Norwegian Institute of Technology (1974).
Afraei, S., Shahriar, K. & Madani, S. H. Developing intelligent classification models for rock burst prediction after recognizing significant predictor variables, Section 1: Literature review and data preprocessing procedure. Tunn. Undergr. Space Technol. 83, 324–353. https://doi.org/10.1016/j.tust.2018.09.022 (2019).
Wang, J. et al. Numerical modeling for rockbursts: A state-of-the-art review. J. Rock Mech. Geotech. Eng. 13, 457–478 (2020).
Jing, L. A review of techniques, advances and outstanding issues in numerical modelling for rock mechanics and rock engineering. Int. J. Rock Mech. Min. Sci. 40, 283–353 (2003).
Weng, L., Huang, L., Taheri, A. & Li, X. Rockburst characteristics and numerical simulation based on a strain energy density index: A case study of a roadway in Linglong gold mine, China. Tunn. Undergr. Space Technol. 69, 223–232 (2017).
Chen, G., He, M. & Fan, F. Rock burst analysis using DDA numerical simulation. Int. J. Geomech. 18, 04018001 (2018).
Xiao, P., Li, D., Zhao, G. & Liu, M. Experimental and numerical analysis of mode I fracture process of rock by semi-circular bend specimen. Mathematics 9, 1769 (2021).
Gong, F.-Q., Luo, Y., Li, X.-B., Si, X.-F. & Tao, M. Experimental simulation investigation on rockburst induced by spalling failure in deep circular tunnels. Tunn. Undergr. Space Technol. 81, 413–427 (2018).
He, M., e Sousa, L. R., Miranda, T. & Zhu, G. Rockburst laboratory tests database—application of data mining techniques. Eng. Geol. 185, 116–130. https://doi.org/10.1016/j.enggeo.2014.12.008 (2015).
Han, Z., Li, D., Zhou, T., Zhu, Q. & Ranjith, P. Experimental study of stress wave propagation and energy characteristics across rock specimens containing cemented mortar joint with various thicknesses. Int. J. Rock Mech. Min. Sci. 131, 104352 (2020).
He, S. et al. Damage behaviors, prediction methods and prevention methods of rockburst in 13 deep traffic tunnels in China. Eng. Fail. Anal. 121, 105178 (2021).
Gong, F.-Q., Wang, Y.-L. & Luo, S. Rockburst proneness criteria for rock materials: Review and new insights. J. Central South Univ. 27, 2793–2821 (2020).
Feng, X. T. & Wang, L. N. Rockburst prediction based on neural networks. Trans. Nonferrous Metals Soc. China 4, 7–14 (1994).
Zhou, J., Shi, X. Z., Huang, R. D., Qiu, X. Y. & Chen, C. Feasibility of stochastic gradient boosting approach for predicting rockburst damage in burst-prone mines. Trans. Nonferrous Met. Soc. China 26, 1938–1945 (2016).
Li, N. & Jimenez, R. A logistic regression classifier for long-term probabilistic prediction of rock burst hazard. Nat. Hazards 90, 197–215 (2018).
Zhou, J., Li, X. & Mitri, H. S. Classification of rockburst in underground projects: Comparison of ten supervised learning methods. J. Comput. Civ. Eng. https://doi.org/10.1061/(asce)cp.1943-5487.0000553 (2016).
Pu, Y., Apel, D. B. & Lingga, B. Rockburst prediction in kimberlite using decision tree with incomplete data. J. Sustain. Min. 17, 158–165 (2018).
Ghasemi, E., Gholizadeh, H. & Adoko, A. C. Evaluation of rockburst occurrence and intensity in underground structures using decision tree approach. Eng. Comput. 36, 213–225 (2020).
Pu, Y., Apel, D. B. & Xu, H. Rockburst prediction in kimberlite with unsupervised learning method and support vector classifier. Tunn. Undergr. Space Technol. 90, 12–18 (2019).
Pu, Y., Apel, D. B., Wang, C. & Wilson, B. Evaluation of burst liability in kimberlite using support vector machine. Acta Geophys. 66, 973–982 (2018).
Cervantes, J., Garcia-Lamont, F., Rodríguez-Mazahua, L. & Lopez, A. A comprehensive survey on support vector machine classification: Applications, challenges and trends. Neurocomputing 408, 189–215 (2020).
Yin, X. et al. Strength of stacking technique of ensemble learning in rockburst prediction with imbalanced data: Comparison of eight single and ensemble models. Nat. Resour. Res. 30, 1795–1815 (2021).
Li, N., Feng, X. & Jimenez, R. Predicting rock burst hazard with incomplete data using Bayesian networks. Tunn. Undergr. Space Technol. 61, 61–70 (2017).
Zhou, J., Guo, H., Koopialipoor, M., Jahed Armaghani, D. & Tahir, M. M. Investigating the effective parameters on the risk levels of rockburst phenomena by developing a hybrid heuristic algorithm. Eng. Comput. https://doi.org/10.1007/s00366-019-00908-9 (2020).
Li, J. Z. M. K. E. Machine Learning Approaches for Long-term Rock Burst Prediction (March 2020).
Moayedi, H., Mosallanezhad, M., Rashid, A. S. A., Jusoh, W. A. W. & Muazu, M. A. A systematic review and meta-analysis of artificial neural network application in geotechnical engineering: Theory and applications. Neural Comput. Appl. 32, 495–518 (2020).
Zhou, J., Koopialipoor, M., Li, E. & Armaghani, D. J. Prediction of rockburst risk in underground projects developing a neuro-bee intelligent system. Bull. Eng. Geol. Environ. 79, 4265–4279 (2020).
Lin, Y., Zhou, K. & Li, J. Application of cloud model in Rock Burst prediction and performance comparison with three machine learning algorithms. IEEE Access 6, 30958–30968. https://doi.org/10.1109/access.2018.2839754 (2018).
Zhang, J., Wang, Y., Sun, Y. & Li, G. Strength of ensemble learning in multiclass classification of rockburst intensity. Int. J. Numer. Anal. Meth. Geomech. 44, 1833–1853. https://doi.org/10.1002/nag.3111 (2020).
Wang, S.-M. et al. Rockburst prediction in hard rock mines developing bagging and boosting tree-based ensemble techniques. J. Central South Univ. 28, 527–542 (2021).
Xie, X., Jiang, W. & Guo, J. Research on Rockburst prediction classification based on GA-XGB model. IEEE Access 9, 83993–84020 (2021).
Olaru, C. & Wehenkel, L. A complete fuzzy decision tree technique. Fuzzy Sets Syst. 138, 221–254 (2003).
Snoek, J., Larochelle, H. & Adams, R. P. Practical bayesian optimization of machine learning algorithms. Adv. Neural Inf. Process. Syst. 25, 2951–2959 (2012).
Freund, Y. & Schapire, R. E. A decision-theoretic generalization of on-line learning and an application to boosting. JCoSS 55, 119–139 (1997).
Friedman, J. H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 29, 1189–1232 (2001).
Chen, T. & Guestrin, C. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 785–794.
Ke, G. et al. Lightgbm: A highly efficient gradient boosting decision tree. Adv. Neural. Inf. Process. Syst. 30, 3146–3154 (2017).
Dorogush, A. V., Ershov, V. & Gulin, A. CatBoost: Gradient boosting with categorical features support. arXiv preprint arXiv:1810.11363 (2018).
Al Daoud, E. Comparison between XGBoost, LightGBM and CatBoost using a home credit dataset. Int. J. Comput. Inf Eng. 13, 6–10 (2019).
Cui, J. & Yang, B. Survey on Bayesian optimization methodology and applications (in Chinese). J. Softw. 29, 3068–3090 (2018).
Kushner, H. J. A new method of locating the maximum point of an arbitrary multipeak curve in the presence of noise. J. Basic Eng. 86, 97–106 (1964).
Mockus, J., Tiesis, V. & Zilinskas, A. The application of Bayesian methods for seeking the extremum. Towards Glob. Optim. 2, 2 (1978).
Jones, D. R., Schonlau, M. & Welch, W. J. Efficient global optimization of expensive black-box functions. JGO 13, 455–492 (1998).
Archetti, F. & Candelieri, A. Bayesian Optimization and Data Science (Springer, New York, 2019).
Hoffman, M., Brochu, E. & de Freitas, N. In UAI. 327–336 (Citeseer).
Xue, Y., Bai, C., Qiu, D., Kong, F. & Li, Z. Predicting rockburst with database using particle swarm optimization and extreme learning machine. Tunn. Undergr. Space Technol. 98, 103287.103281–103287.103212 (2020).
Xue, Y. et al. Prediction of rock burst in underground caverns based on rough set and extensible comprehensive evaluation. Bull. Eng. Geol. Environ. 78, 417–429 (2019).
Ran, L., Ye, Y., Hu, N., Hu, C. & Wang, X. Classified prediction model of rockburst using rough sets-normal cloud. Neural Comput. Appl. 31, 8185–8193 (2019).
Jia, Q., Wu, L., Li, B., Chen, C. & Peng, Y. The comprehensive prediction model of rockburst tendency in tunnel based on optimized unascertained measure theory. Geotech. Geol. Eng. 37, 3399–3411 (2019).
Du, Z., Xu, M., Liu, Z. & Xuan, W. Laboratory integrated evaluation method for engineering wall rock rock-burst. Gold (2006).
Wu, S., Wu, Z. & Zhang, C. Rock burst prediction probability model based on case analysis. Tunn. Undergr. Space Technol. 93, 103069.103061–103069.103015 (2019).
Pedregosa, F. et al. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 12, 2825–2830 (2011).
Liang, W., Sari, A., Zhao, G., McKinnon, S. D. & Wu, H. Short-term rockburst risk prediction using ensemble learning methods. Nat. Hazards 104, 1923–1946 (2020).
Head, T., MechCoder, G. L. & Shcherbatyi, I. scikit-optimize/scikit-optimize: v0. 5.2. Zenodo (2018).
Xiao, P., Li, D., Zhao, G. & Liu, H. New criterion for the spalling failure of deep rock engineering based on energy release. Int. J. Rock Mech. Min. Sci. 148, 104943 (2021).
Zorlu, K., Gokceoglu, C., Ocakoglu, F., Nefeslioglu, H. & Acikalin, S. Prediction of uniaxial compressive strength of sandstones using petrography-based models. Eng. Geol. 96, 141–158 (2008).
Zhang, L., Zhang, D. & Qiu, D. Application of extension evaluation method in rockburst prediction based on rough set theory. J. China Coal Soc. 35, 1461–1465 (2010).
Lee, P., Tsui, Y., Tham, L., Wang, Y. & Li, W. Method of fuzzy comprehensive evaluations for rockburst prediction (in Chinese). Chin. J. Rock Mech. Eng. 17, 493–501 (1998).
Zhang, C., Zhou, H. & Feng, X.-T. An index for estimating the stability of brittle surrounding rock mass: FAI and its engineering application. Rock Mech. Rock Eng. 44, 401–414 (2011).
Funding
This research was financially supported by the National Natural Science Foundation of China (52074349).
Author information
Authors and Affiliations
Contributions
D.L. drafted and reviewed the manuscript; Z.L. collected the data and conducted the modeling; D.J.A. reviewed this paper; P.X. conducted the rock mechanics experiments; J.Z. reviewed this paper.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Li, D., Liu, Z., Armaghani, D.J. et al. Novel ensemble intelligence methodologies for rockburst assessment in complex and variable environments. Sci Rep 12, 1844 (2022). https://doi.org/10.1038/s41598-022-05594-0
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-022-05594-0
This article is cited by
-
A Transparent and Valid Framework for Rockburst Assessment: Unifying Interpretable Machine Learning and Conformal Prediction
Rock Mechanics and Rock Engineering (2024)
-
Rockburst Intensity Grade Prediction Based on Data Preprocessing Techniques and Multi-model Ensemble Learning Algorithms
Rock Mechanics and Rock Engineering (2024)
-
Machine learning for predicting hepatitis B or C virus infection in diabetic patients
Scientific Reports (2023)
-
Weighted ensembles of artificial neural networks based on Gaussian mixture modeling for truck productivity prediction at open-pit mines
Mining, Metallurgy & Exploration (2023)
-
Intelligent hybrid model to classify failure modes of overstressed rock masses in deep engineering
Journal of Central South University (2023)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.