1 INTRODUCTION
Blasting, as a common method for tunnel excavation and rock breakage, is widely used in mining and tunneling engineering. The distribution of rock fragmentation after blasting directly affects downstream operations like shoveling, loading, and milling (Amoako et al., 2022; Enayatollahi et al., 2014; Inanloo Arabi Shad et al., 2018; Kinyua et al., 2022; Shehu et al., 2022). Oversized fragments require secondary crushing to improve the efficiency of shovels and loaders, resulting in higher explosive costs and safety issues. Finer particles, on the contrary, are more susceptible to water and easily cemented, increasing the risk of chute blockage. In addition, blasting-induced rock fracture behavior characterization cannot be overlooked in engineering. To address this problem, several experiments and theoretical analyses of rock fracture mechanisms under different conditions have been conducted (Han, Li, Li, et al., 2023; Han, Li, Wang, et al., 2023). Since the above phenomenon occurs in most cases, it is crucial for engineers to quickly obtain rock fragment size after the blasting operation.
Rock fragment size can be measured by direct and indirect methods. The most well-known direct method is the sieving method. Although the sieving method can render more precise results, it is time-consuming, costly, and only suitable for small-scale blasting tests (Amoako et al., 2022; Bamford et al., 2021; Kinyua et al., 2022; Qiao et al., 2021). As a result, empirical models and indirect methods have been widely applied in engineering. The Kuz–Ram model (Cunningham, 1987), the extended Kuz–Ram model (Hekmat et al., 2019; Lawal, 2021; Morin & Ficarazzo, 2006), the KCO model (Ouchterlony, 2005), the SveDeFo model (Hjelmberg, 1983), and the Kou–Rustan model (Kou & Rustan, 1993) are the most commonly used empirical models. However, the empirical models also have their limitations: (1) they are established based on a single mathematical equation, thus failing to capture the complex relationships between the influential factors and rock fragment size (Amoako et al., 2022; Li, Yang, et al., 2021), and (2) the key influential factors that affect rock fragmentation, such as rock properties, charge structure, and detonation method, cannot be explained well quantitatively.
Artificial intelligence (AI)-based methods have displayed excellent advantages in terms of complex nonlinear problems in recent years, and are thus being extensively used in engineering fields, such as rock physical and mechanical parameter prediction (Ceryan et al., 2013; He et al., 2021; Li et al., 2020; Momeni et al., 2015), rock lithology classification (Li et al., 2022a, 2022b), and rock burst prediction (Li et al., 2022a, 2022b; Zhou et al., 2020). The remarkable performance of AI-based methods has attracted much attention from scholars engaged in rock fragmentation prediction. For instance, Kulatilake et al. (2010) trained a shallow artificial neural network (ANN) to predict X50. The Levenberg–Marquardt algorithm was adopted to optimize the hidden layer units, and the trained model was proven to be practical in mines. Enayatollahi et al. (2014) developed a two-hidden layer ANN model, and the experimental results demonstrated that the ANN model outperformed the regression models and the empirical Kuz–Ram model. Dimitraki et al. (2019) introduced a simple ANN model for rock fragment size prediction. The model required only three input parameters (blastability index, powder factor, and quantity of blasted rock pile) and produced better results with a coefficient of determination (
) value of 0.8, which was higher than that of the regression model (
value of 0.7). Hasanipanah et al. (2018) integrated an adaptive neuro-fuzzy inference system model and an optimization algorithm for the prediction of X50. The model predictive performance outperformed the multiple regression equation (MR) method. Moreover, Asl et al. (2018) developed a hybrid model using ANN and the firefly optimization algorithm. The values of
and the root mean square error (RMSE) were 0.94 and 0.10, respectively, and the optimized model reduced rock fragmentation size by 32.9%. Li, Koopialipoor, et al. (2021) proposed an optimized support vector regression (SVR) model by combining five optimization algorithms, and the optimized model achieved satisfactory performance with 19 input parameters.
Although ANN methods have achieved reliable results in rock fragmentation prediction, the performance of ensemble learning algorithms deserves in-depth study. Ensemble learning algorithms such as random forest (RF), adaptive boosting (AdaBoost), gradient boosting (GBoost), and extremely randomized trees (ERT) are simple and practical. Accordingly, they have received considerable attention in recent years. Zhang et al. (2021) and Zhou et al. (2019) used an RF-based method for material strength prediction, and the trained model showed good prediction capabilities. Zhou et al. (2020) developed an RF model to estimate the ground vibration by using the Bayesian optimization algorithm (BOA). The model obtained 92.95% and 90.32% accuracy on the training and testing sets, respectively. Dai et al. (2022) combined an RF model and a particle swarm optimization algorithm to estimate the backbreak after blasting, and the
reached 0.9961 on the testing data set. AdaBoost, another ensemble learning algorithm, has also been used in a wide range of applications, including rock mass classification (Liu et al., 2020), rockburst prediction (Wang et al., 2021), rock strength estimation (Liu et al., 2022), tunnel boring machine (TBM) performance prediction (Zhang et al., 2020), and engineering geology (Wu et al., 2020; Yang et al., 2019). According to Geurts and Louppe (2011), the ERT model had a higher prediction accuracy than the RF model. Overall, the studies mentioned above have verified the excellent abilities of ensemble learning algorithms. This paper aims to use four ensemble learning algorithms to predict rock fragmentation. In addition, the BOA, proposed by the American scholar Pelikan et al. (1999), is utilized to obtain the optimal hyperparameters of the ensemble learning models. BOA boasts of advantages of fewer iterations and higher processing speed. Accordingly, it has been extensively utilized to fine-tune the parameters of machine learning (ML) models (Canayaz et al., 2022; Egberts et al., 2022; Snoek et al., 2012). In all, four hybrid models, namely, RF-BOA, AdaBoost-BOA, GBoost-BOA, and ERT-BOA, are established to predict rock fragmentation.
This paper is organized as follows: In Section 2, the background of the four ensemble learning algorithms and the BOA is introduced. In Section 8, the process of hybrid model development is described. In Section 13, the prediction capabilities of the four hybrid models (RF-BOA, AdaBoost-BOA, GBoost-BOA, and ERT-BOA) are analyzed and discussed, and the optimal hybrid model for rock fragmentation prediction is confirmed. Finally, sensitivity analysis is implemented to investigate the effects between the input variables and the output X50.
2 METHODOLOGY
In this paper, four ensemble learning algorithms, namely, RF, adaptive boosting, gradient boosting, and ERT, are used to predict rock fragmentation, and the structures are shown in Figure 1. Furthermore, the BOA is adopted to determine the best parameters of all models to improve model prediction accuracy.
Flowchart of the ensemble learning algorithms. (a) Random forest method; (b) adaptive boosting method; (c) gradient boosting method (a–c are modified from Chen et al. [
2022]); and (d) extremely randomized trees method (modified from Yang et al. [
2021]).
2.1 RF
RF is a commonly used method for addressing classification and regression issues. The RF model has strong generalization capabilities, and thus can achieve better results even with noise data sets (Zhang et al.,
2022). Typically, only two parameters need to be fine-tuned: the maximum depth of the tree and the number of estimators, which makes it more user-friendly. For regression prediction, the RF model generates
N regression trees and renders the final result by integrating the output of all the independent regression trees, as shown in Equation (
1). The structure is shown in Figure
1a.
(1)
where
indicates the final result;
represents the input vector;
is the result of each decision tree; and
is the number of regression trees.
2.2 Adaptive boosting
AdaBoost is a typical ensemble boosting method that consolidates weak learners with strong learners. It gives full consideration to the weights of all learners so as to obtain a more accurate and less overfitted model. Its framework is shown in Figure 1b and the procedure of the algorithms is shown as follows:
(1)
Determine the regression error rate of the weak learner.
Obtain the maximum error;
(2)
where
indicate the input data; m is the number of samples; and
represents the weak learner of the t-th iteration.
Estimate the relative error for each sample by using the linearity loss function;
(3)
Determine the regression error rate;
(4)
where
denotes the weights of the sample
.
(2)
Determine the weight coefficient
of the weak learner.
(5)
(3)
Sample weight updating for the
iteration round.
(6)
(7)
where Zt denotes the normalization factor.
(4)
Build the ultimate learner.
(8)
f (x) is the median value of αt ht (x)(t = 1, 2, …, T).
2.3 Gradient boosting
GBoost is another empirical ensemble algorithm derived from a gradient descent, whose flowchart is shown in Figure
1c. Given a set of training data sets
, the maximum iterations number
, and the squared error loss function
, a strong learner could be obtained as follows:
(1)
Initialize the base learner.
(9)
(10)
where
indicates the squared error loss function;
is a constant; and
is the target value.
(2)
Calculate the negative gradient.
Then, the leaf node region
of the t-th regression tree (
is the number of leaf nodes of a regression tree
) could be obtained according to
.
(11)
(3)
Obtain the optimal-fitted value
for the leaf region j = 1, 2, …, J, and update the strong learner.
(12)
(13)
where I is the identity matrix.
(4)
Finally, a strong learner can be obtained.
(14)
The advantage of GBoost consists of its flexibility in choosing the loss function, which makes it possible to use any continuously derivable loss function. Also for this reason, it is feasible to adopt some robust loss functions to make the model more robust. Therefore, the GBoost method has achieved many successful applications and more modified versions have been developed (Chen & Guestrin, 2016; Ke et al., 2017; Prokhorenkova et al., 2018).
2.4 ERT
Figure 1d shows the architecture of the ERT model, which is similar to that of the RF model. The ERT model sets K features randomly from the entire sample. Each of these K features yields K classification nodes, and the node with the highest score is considered the final split node. According to Geurts and Louppe (2011), ERT showed better prediction ability than the RF model. Hameed et al. (2021) applied ERT to obtain the discharge coefficient for side weirs, and the model prediction results were more reliable than the field engineer's analysis results. Djarum et al. (2021) explored ML algorithms for river water quality prediction using linear discriminant analysis. The results revealed that the ERT-based model outperformed other ensemble models. In addition, research has been carried out to study the application of the ERT model to predict stock prices (Pasupulety et al., 2019; Polamuri et al., 2019).
2.5 BOA
Compared with the commonly used grid and random search algorithms, BOA takes full advantage of prior information to determine the parameters that maximize the target function globally. The algorithm is composed of two parts: (1) Gaussian process regression, which aims to determine the values of the mean and variance of the function at each point, and (2) the acquisition function, which is used to obtain the search position of the next iteration, as shown in Figure 2. BOA boasts of advantages of fewer iterations and higher processing speed, because of which it has been applied in several fields. Li, Zhao, and Ma (2022) proposed an intelligent model for rockburst prediction using the BOA for hyperparameter optimization. Lahmiri et al. (2023) used the BOA to obtain the optimal parameters of models for house price prediction. Bo et al. (2023) developed an ensemble classifier model to assess tunnel squeezing hazards using the BOA method, and the optimal values of the 17 parameters were obtained.
Schematic of the Bayesian optimization algorithm (Greenhill et al.,
2020).
4 RESULTS, ANALYSIS, AND DISCUSSION
4.1 Evaluation indexes
Three evaluation metrics, namely,
RMSE,
MAE, and
, are used to evaluate the predictive performance of all models. The lower the
RMSE and
MAE values, the better the model performs. This indicates that the prediction results are closer to the measured values. Conversely, the greater the value of
, the more robust the model will be, and the maximum value is equal to 1.
(15)
(16)
(17)
where
,
, and
represent the target value, the prediction result, and the average of all the target values, respectively.
The prediction capabilities of the four standalone ensemble models, the hybrid models (RF-BOA, AdaBoost-BOA, GBoost-BOA, and ERT-BOA), and the previous studies (Amoako et al., 2022; Hudaverdi et al., 2011) are compared in this section. Table 4 shows the testing data set, which is completely consistent with the previous studies.
Table 4. Testing data set (Hudaverdi et al.,
2011; Kulatilake et al.,
2010).
ID |
S/B |
H/B |
B/D |
T/B |
fp (kg/m3) |
XB (m) |
E (GPa) |
X50 (m) |
Ad23 |
1.11 |
4.44 |
18.95 |
1.67 |
1.25 |
1.63 |
16.90 |
0.21 |
Ad24 |
1.28 |
3.61 |
18.95 |
1.67 |
0.89 |
0.61 |
16.90 |
0.20 |
Db10 |
1.15 |
4.35 |
20.00 |
1.75 |
0.89 |
1.00 |
9.57 |
0.35 |
En13 |
1.24 |
1.33 |
27.27 |
0.78 |
0.48 |
1.11 |
60.00 |
0.47 |
Mg8 |
1.10 |
2.40 |
30.30 |
0.80 |
0.55 |
1.23 |
50.00 |
0.44 |
Mg9 |
1.00 |
2.67 |
27.27 |
0.89 |
0.75 |
0.77 |
50.00 |
0.25 |
Mr12 |
1.25 |
6.25 |
31.58 |
0.63 |
0.48 |
1.03 |
32.00 |
0.20 |
Ru7 |
1.13 |
5.00 |
39.47 |
3.11 |
0.31 |
2.00 |
45.00 |
0.64 |
Sm8 |
1.25 |
2.50 |
28.57 |
0.83 |
0.42 |
0.50 |
13.25 |
0.18 |
Oz8 |
1.20 |
2.40 |
28.09 |
1.00 |
0.53 |
0.82 |
15.00 |
0.23 |
Oz9 |
1.11 |
3.33 |
30.34 |
1.11 |
0.47 |
0.54 |
15.00 |
0.17 |
Where the symbols Ad, Db, En, Mg, Mr, Ru, Sm, and Oz indicate the abbreviation of Akdaglar quarry, the Dongri–Buzurg open-pit manganese mine, the Enusa mine, the Murgul copper mine, Mrica quarry, the Reocin underground mine, the Soma basin mine, and Ozmert quarry, respectively.
4.2 Optimization of the hybrid models
The hyperparameters for the RF model that need to be optimized include the maximum number of estimators, maximum features, and maximum decision tree depth. However, the stability of the AdaBoost algorithm is dependent on the parameters' learning rate, the maximum decision tree depth, and the loss function. The hyperparameters of the GBoost algorithm that need to be optimized are maximum number of estimators, maximum decision tree depth, learning rate, and maximum features. The main optimization parameters of the ERT method include maximum features, maximum decision tree depth, minimum number of samples required to split an internal node, and minimum number of samples required to be at a leaf node. To optimize the corresponding parameters of all the ensemble learning models, BOA is used to obtain the optimal results with 500 iterations.
In this section, the prediction performance of RF-BOA, AdaBoost-BOA, GBoost-BOA, and ERT-BOA hybrid models is systematically evaluated and compared. Figure 8a depicts the loss convergence of the four hybrid models during the training process. The loss values of hybrid models RF-BOA and ERT-BOA converge to the same level and yield comparable prediction performance, as shown in Figure 8a,b. The values of
, RMSE, and MAE of RF-BOA are 0.88, 0.05, 0.05, and those of ERT-BOA are 0.89, 0.05, and 0.04, respectively. Although the minimum loss value of the AdaBoost model is smaller than that of the RF-BOA and ERT-BOA models, the entire loss curve is more oscillating. Obviously, the hybrid model GBoost-BOA is the best-performing model among the four hybrid models. The loss value, on the one hand, converges to the lowest level; the value of
is the highest (
); and the values of RMSE and MAE are 0.03 and 0.02, which are also the best. The optimal hyperparameters of the four hybrid models are obtained by using the BOA, as shown in Table 5.
(a) Training loss of the four hybrid models and (b) evaluation results of the four hybrid models on the testing data set.
Table 5. Optimized values of the hyperparameters for different machine learning models.
Models |
Hyperparameters |
Optimal value |
Search space |
RFR |
Maximum number of estimators |
65 |
1–200 |
Maximum features |
5 |
1–10 |
Maximum depth of the tree |
8 |
1–50 |
ABR |
Learning rate |
0.1 |
0–1 |
Maximum number of estimators |
96 |
1–200 |
The loss function |
Square |
Linear, square, etc. |
GBR |
Maximum number of estimators |
125 |
1–200 |
Maximum depth of the tree |
6 |
1–50 |
Learning rate |
0.15 |
0–1 |
Maximum features |
6 |
1–10 |
ERT |
Maximum features |
4 |
1–10 |
Maximum depth of the tree |
5 |
1–50 |
Minimum number of samples required to split an internal node |
2 |
2–30 |
Minimum number of samples required to be at a leaf node |
1 |
1–30 |
Abbreviations: ABR, adaptive boosting regression; ERT, extremely randomized trees; GBR, gradient boosting regression; RFR, random forest regression.
Besides, Figure 9a,b show the rock fragmentation prediction results of the four hybrid models. The abscissa axis indicates the measured values; the ordinate axis denotes the model prediction results; and the green dotted line
represents the reference line. The closer the prediction result is to the reference line, the better the model performs. As can be seen, the prediction results of the GBoost-BOA hybrid model almost match the reference line, whereas the other three hybrid models perform slightly worse. Accordingly, the model prediction capability is ranked in descending order as follows: GBoost-BOA, AdaBoost-BOA, ERT-BOA, and RF-BOA.
Model prediction results on the testing data set. (a) RF-BOA prediction result, (b) AdaBoost-BOA prediction result, (c) GBoost-BOA prediction result, and (d) ERT-BOA prediction result.
4.3 Comparison between hybrid models and standalone models
Figure 10 summarizes the prediction results of all standalone ensemble learning models and the corresponding hybrid models on the testing data set. The red pentagrams indicate the prediction results of the standalone ensemble learning model; the solid balls denote the prediction results of the hybrid models; and the green dotted line represents the reference line. As shown in Figure 10, compared with solid balls, the red pentagrams are further away from the reference line, indicating that the hybrid model is more robust than standalone models and that the optimization algorithm is useful for the improvement of the model prediction capability. In addition, a box plot is applied to show the detailed errors between all the predictive model results and the measured values, as shown in Figure 11. Obviously, the prediction performance of the four hybrid models is better than that of the corresponding standalone models.
(a–d) Comparison results of standalone and hybrid model prediction performance.
Box-plot of errors between the model prediction results and the measured values.
Furthermore, as shown in Figure 12, the evaluation indices
, RMSE, and MAE are calculated to quantitatively compare the prediction capabilities of the standalone models and the hybrid models. All hybrid models have higher
values than the standalone models, with an improvement rate greater than 15%. The largest amplitude of improvement appears between the AdaBoost–BOA hybrid model and the corresponding standalone model, reaching up to 37.44%. The values of RMSE and MAE of all the hybrid models, on the contrary, are lower than those of the standalone models, and the reduction rates of RMSE and MAE are more than 30% and 25%. According to the evaluation metrics, all hybrid models outperform the corresponding standalone models.
(a–c) Evaluation comparison of the standalone and hybrid models on
R
2,
RMSE, and
MAE, respectively.
4.4 Comparison results between the hybrid model and previous literature methods
Furthermore, several studies on rock fragmentation prediction are proposed using this data set. Table 6 shows a comparison of performance between the previous studies and the GBoost–BOA hybrid model.
Table 6. Comparison results of
X
50 predicted by different methods. m
ID |
X50 |
X50-GBoost-BOA |
X50-K |
X50-MVR |
X50-ANN |
X50-SVR |
Ad23 |
0.21 |
0.22 |
0.12 |
0.19 |
0.21 |
0.20 |
Ad24 |
0.20 |
0.20 |
0.13 |
0.15 |
0.21 |
0.19 |
Db10 |
0.35 |
0.35 |
0.09 |
0.16 |
0.21 |
0.52 |
En13 |
0.47 |
0.41 |
0.48 |
0.39 |
0.44 |
0.38 |
Mg8 |
0.44 |
0.40 |
0.42 |
0.40 |
0.38 |
0.41 |
Mg9 |
0.25 |
0.29 |
0.33 |
0.24 |
0.25 |
0.25 |
Mr12 |
0.20 |
0.19 |
0.27 |
0.14 |
0.15 |
0.14 |
Ru7 |
0.64 |
0.64 |
0.71 |
0.51 |
0.68 |
0.61 |
Sm8 |
0.18 |
0.19 |
0.38 |
0.17 |
0.19 |
0.19 |
Oz8 |
0.23 |
0.24 |
0.22 |
0.17 |
0.17 |
0.18 |
Oz9 |
0.17 |
0.20 |
0.25 |
0.17 |
0.17 |
0.19 |
Note: The results of the X50-K, X50-ANN, and X50-SVR were presented by Amoako et al. (2022), and the result of the X50-MVR was reported by Hudaverdi et al. (2011). Abbreviation: MVR, multivariate regression.
Figure 13 displays the comparison of the evaluation metrics between the existing models and the GBoost–BOA model. As can be seen, the existing models are the same as those described in Table 6. First, it can be concluded that the prediction methods based on ML or ANN-based prediction methods (such as the GBoost-BOA, ANN, and SVR model) outperform traditional methods (Kuz–Ram and MVR methods). Furthermore, as illustrated in Figure 14, the median values of the box plots for the GBoost-BOA, ANN, and SVR models are lower than those of the Kuz–Ram model and the MVR method. Second, the GBoost–BOA hybrid model shows the best prediction performance compared with Kuz–Ram, MVR, ANN, and SVR.
Evaluation comparison of different methods on
R
2,
RMSE, and
MAE.
Variability between predictions and target values for different models.
In summary, the findings of this paper show that the hybrid model GBoost–BOA developed in this paper outperforms the models developed in the previous studies (Amoako et al., 2022; Hudaverdi et al., 2011) as well.
4.5 Sensitivity analysis
In addition, sensitivity analysis was conducted to better understand the intrinsic relationships between the seven independent variables and rock fragmentation. The relevancy factor is a commonly used method to illustrate the sensitivity scale (Bayat et al.,
2021; Chen et al.,
2014). Accordingly, it is applied in this paper to assess the influence of each variable on rock fragmentation. It is common knowledge that the absolute value of the RF between the independent and dependent variables is positively correlated with the influence of the variable on rock fragmentation. In some cases, it is also necessary to specify whether the correlation is positive or negative. The main form of the sensitivity relevancy factor (
SRF) is as follows:
(18)
where
denotes the mean value of all data for variable
l (
l includes
S/B,
H/B,
B/D,
T/B,
f
p,
X
B, and
E);
represents the
ith value of variable
l;
indicates the number of the variable data; and
and
are the
ith measured value of variable
l and the average value of the prediction results, respectively.
Figure 15 shows the results of the sensitivity analysis. Four factors (B/D, T/B, XB, E) show a positive impact on rock fragmentation, whereas S/B and fp exert a negative impact. The RF value of H/B is close to zero and thus can be ignored. Among all variables, XB (
is 0.73) has the greatest effect on rock fragmentation, followed by T/B, E, and B/D. The sensitivity analysis results in this paper are highly consistent with previous studies (Mehrdanesh et al., 2018; Sayadi et al., 2013).
Sensitivity analysis result.
5 CONCLUSIONS
In this study, hybrid predictive models combined with ensemble learning algorithms and Bayesian optimization method were developed for rock fragmentation prediction. In total, 102 samples were used to evaluate the prediction performance of the predictive models. The evaluation results revealed that the hybrid models outperformed the corresponding standalone models. The hybrid model GBoost–BOA achieved far better prediction results than other hybrid models. Specifically, it obtained the highest R2 value of 0.96, and the smallest values of RMSE and MAE, 0.03 and 0.02, respectively, on the testing data set. Moreover, the proposed GBoost–BOA hybrid model showed better accuracy and generalization ability than other models developed previously.
In addition, the sensitivity analysis results indicated that four factors (XB, T/B, E, and B/D) showed positive impacts, and the sensitivity relevancy factors were 0.73, 0.63, 0.60, and 0.54, respectively. S/B and fp reflected negative influences, with the values of the sensitivity relevancy factor being −0.27 and −0.35. The contribution of H/B in the prediction of rock fragmentation was much less and thus negligible. Therefore, XB, T/B, and E were the primary influencing factors among all the variables.