1 INTRODUCTION
Tunnel boring machine (TBM) boasts of the advantages of fast excavation, high quality, economy, environmental protection, and lower labor intensity (Qian et al., 2002). It is capable of simultaneously executing multiple processes, including excavation, debris exclusion, and support (Du & Du, 2010). With the continuous advancement in construction technology and increasing mechanization, TBM, as an advanced semiautomated tunnel construction method, is increasingly being used for the excavation of rock tunnels (Liu et al., 2016). There is a poor match between the rock mass state parameters and the tunneling parameters. When the geological conditions change, it becomes difficult to adjust the tunneling parameters and excavation schemes promptly and efficiently (Li, 2022; Zhang et al., 2018). The accurate perception of rock mass information ahead of the TBM face, particularly the information about rock mass classification, facilitates the main drivers in optimizing and adjusting the tunneling parameters. Therefore, the objective of this study is to develop a neural network for accurately predicting rock mass classification in TBM tunneling.
Traditional geological forecasting methods, such as advanced drilling (Kim et al., 2018; Yue et al., 2002) and seismic wave detection (Li et al., 2010), necessitate additional time and cost. During excavation, TBM is influenced by the conditions of the rock mass. The changes in surrounding rock response and tunneling parameters result from the interaction between the TBM and the rock mass (Galende-Hernández et al., 2018; Zhu et al., 2020). The perception of rock machine information and the identification of rock mass classification have always been major issues in tunnels excavated by TBM. Various machine learning techniques, such as artificial neural network (ANN), support vector classifier, k-nearest neighbor, and decision tree algorithm, have been introduced to address such classification problems (Ghasemi et al., 2017; Młynarczuk et al., 2013; Shahriar et al., 2008). Yamamoto et al. (2003) developed a real-time TBM tunneling control system capable of predicting geological conditions. Jung et al. (2019) developed a universal model for predicting ground conditions ahead of the tunnel face by utilizing the advantages of ANN in handling irregular TBM operational data from four different sites. Li (2022) established a model called self-organizing maps-support vector machine (SOM-SVM) for the prediction and recognition of rock mass classification. The results obtained demonstrated that the model could render strong prediction accuracy and robustness. Random forest (RF) and adaptive cost (AdaCost) models were used to identify the rock mass classification in real time based on TBM tunneling parameters (Fu, 2021; Zhu et al., 2020). In a variety of actual projects, a deep neural network model was used to predict the classification of rock mass (Mao et al., 2021; Nie et al., 2019). The outcomes showed that the prediction accuracy of rock mass classification was higher than 85%. The long short-term memory (LSTM) model was used by Erharter et al. (2019) to establish the correlation between TBM tunneling parameters and the classification of rock mass. Although the above-mentioned studies on machine information perception and rock mass classification prediction have produced excellent outcomes, the following issues remain unsolved. Some models are criticized for their overly slow response due to excessive complexity. Some other models only include TBM tunneling parameters as input variables for predicting the classification of rock mass, failing to cover rock mass parameters as input variables as well.
Apart from the influence of the basic characteristics of machine learning models on the accuracy of rock mass classification predictions, the rationale within the standards for rock mass classification is also critical to the prediction process. Various classification methods of the rock mass quality index based on rock mass stability, such as the Beckman and Quarles method, rock quality designation, rock mass rating, and hierarchical clustering, have been developed to thoroughly analyze the complex geological conditions (Gong et al., 2021; Li et al., 2023a; 2023b; Salimi et al., 2016). The gated recurrent unit (GRU) neural network is widely used in geotechnical and power engineering owing to its continuous advancements in deep learning (Xiao et al., 2023; Yue et al., 2021). To improve the accuracy and operational efficiency of the rock mass classification prediction model, this study proposed a neural network that combines principal component analysis (PCA) (Ren & Yu, 2011) and GRU (Chung et al., 2014). The neural network was applied in the TBM project of Lanzhou Water Source Construction, which involved the classification of rock mass based on the HC method.
2 PCA–GRU NEURAL NETWORK
2.1 PCA
PCA is a traditional statistical analysis method primarily suitable for processing high-dimensional data with a strong correlation between variables (Ren & Yu, 2011). The core concept of PCA involves forming linear combinations of the original indicators to create a set of new, independent, and comprehensive indicators (Duan et al., 2020). These new indicators contain most of the original information, thereby achieving the objective of dimension reduction. The algorithm is outlined as follows (Wu et al., 2019):
Suppose there are
m samples and
n evaluation indicators. The
jth indicator of the
ith sample takes the value of
and the raw data variable indicator matrix
is constructed.
(1)
where
is a raw data variable metric. The composite index is set after dimension reduction, namely, the principal component
. Then, the linear combination of indicator vectors for the raw data variable's indicator matrix is established as follows:
(2)
Where lnn is for variable coefficient. Equations (1) and (2) satisfy the following conditions: the sum of the squared coefficients in each equation is 1, and any two principal components are independent and unrelated to each other.
The steps to solve the principal component from the covariance matrix are as follows:
Step 1: Normalize the raw data.
(3)
(4)
where
and
represent the mean value and the standard deviation of the
jth indicator, respectively.
Step 2: Calculate the Pearson correlation coefficient matrix between indicators:
(5)
where
denotes the correlation coefficient between the
kth indicator and the
lth indicator;
and
are symmetric matrices of each other. The calculation equation is as follows:
(6)
Step 3: Calculate the eigenvalues and eigenvectors of the Pearson correlation coefficient. The eigenvalues are denoted as
, satisfying
, and the unitized eigenvectors corresponding to the eigenvalues are denoted as
.
Step 4: Determine the number of principal components. To calculate the cumulative contribution ratio of the principal components, the previous
principal components corresponding to the eigenvalues greater than 1 would be generally taken. The cumulative contribution ratio is greater than 85%.
(7)
(8)
In the above equations,
is the variance contribution ratio for the sth principal component and
is the cumulative contribution ratio of the previous
principal components.
Step 5: Calculate the corresponding score of the extracted principal components. The matrix of principal component coefficients is
. If the previous
principal components are extracted from the original indicators, then:
(9)
where
is the standardized matrix of raw data and
is the normalized indicator variable.
2.2 GRU neural network
The GRU neural network, first proposed by Chung et al., has demonstrated excellent performance in time-series forecasting and found widespread applications in transportation, power engineering, and other fields (Chung et al., 2014; Yue et al., 2021). The architecture of the GRU neural network is depicted in Figure 1.
Architecture of the gated recurrent unit neural network.
The expressions for the reset gate and update gate are defined as follows (Xiao et al.,
2023):
(10)
(11)
(12)
(13)
In the above equations,
and
represent the output of the reset gate and update gate, respectively;
represents the sigmoid function;
denotes the hidden state of the neurons at the previous moment;
,
, and
represent the weight matrices of the update gate, reset gate, and temporary output, respectively;
,
, and
are bias functions;
is the input value of the GRU neural network at the present moment; and
and
refer to the vectors of information that provide temporary unit state and hidden layer output at the moment of
, respectively.
2.3 PCA–GRU neural network
To extract the corresponding principal components from raw data, PCA dimension reduction processing is performed using SPSS software. Subsequently, the PCA–GRU neural network is constructed, using the principal components as the input layer for the GRU neural network. The model construction process of the PCA–GRU neural network is illustrated in Figure 2.
Model construction process of the principal component analysis–gated recurrent unit neural network.
As depicted in Figure 2, the feature sequence obtained through PCA dimension reduction serves as the input for the GRU model. Typically, 80% of the data postdimensionality reduction is allocated to the training set and the remaining 20% is allocated to the test set. The data are subjected to normalization, and the model's internal training facilitates denormalized output.
3 ACTUAL PROJECT
3.1 Project overview
In this study, the existing data in the TBM project of Lanzhou Water Source Construction (Jiang, 2020; Yan & Jiang, 2020) were used to verify the effectiveness of the PCA–GRU model in predicting the rock mass classification. The excavation diameter of the tunnel is 5.46 m, and the final diameter with lining is 4.60 m (Yan & Jiang, 2020). The length of the TBM construction section for the tunnel is approximately 27.0 km. The tunnel is excavated using two double-shield TBMs combined with drilling and blasting in certain segments (Yan & Jiang, 2020). Among the excavation sections, the length of the TBM1 construction section is 12.426 km, while that of the TBM2 construction section is about 13.259 km. The main design parameters of the two double-shield TBMs are shown in Table 1.
Table 1. Main design parameters of tunnel boring machine (TBM) (Jiang,
2020).
Parameter |
TBM1 |
TBM2 |
Cutterhead diameter (mm) |
5480 |
5480 |
Cutter number |
37 |
30 |
Cutter diameter (mm) |
432 (center cutter), 483 |
432 (center cutter), 483 |
Maximum cutter spacing (mm) |
86 |
83 |
Range of rotation speed (r/min) |
0–10.3 |
0–8.7 |
Cutterhead power (kW) |
1800 |
2100 |
Maximum cutterhead thrust (kN) |
22 160 |
11 900 |
Cutterhead nominal torque (kN·m) |
3458 |
4210 |
Maximum tunneling speed (mm/min) |
120 |
120 |
According to the HC method, the rock mass classification for the TBM1 construction section indicates that the predominant rock mass type is Class II, approximately 48% (Yan & Jiang, 2020). Following this, Class III and Class IV account for approximately 25% and 23%, respectively, while Class V is less prevalent, making up around 4%. The percentage distribution of the rock mass classes for the Lanzhou Water Source Construction TBM project is shown in Figure 3. Class III rock masses are primarily characterized by poor integrity, whereas Class II rock masses show comparatively good integrity. Class IV and V rock masses are typically fragmented and demonstrate extremely poor integrity.
Percentage distribution of the rock mass classes for the Lanzhou Water Source Construction tunnel boring machine project.
3.2 Data collection
In this study, three different classifications of rock mass—Class II, Class III, and Class IV—were utilized (Jiang, 2020). The initial data set collection included nine parameters: uniaxial compressive strength (UCS), Brazilian tensile strength (BTS), rock integrity index (Kv), rock wear resistance index (CAI), deformation modulus (E0), Poisson's ratio (μ), net tunneling rate (PR), rotation speed (RPM), and cutterhead thrust (TF). The selection of these parameters was guided by principles of parameter availability, model complexity, and their frequency of use in previous research (Jiang, 2020). The final sample data set consisted of 150 carefully selected sets of representative engineering measured data. The geological daily report of the tunnel construction site was used to extract tunneling parameters such as TF, RPM, and PR. Average values were then computed for various tunneling sections.
The statistical description of the input variables in the sample data set is presented in Table 2. The percentage distribution of rock mass classes in the sample data set is shown in Figure 4. The percentage distribution of different lithologies in the sample data set is shown in Figure 5. Violin with box depicting the input variables for three rock mass classifications in the sample data set is shown in Figure 6.
Table 2. Basic descriptive statistics of input variables in the sample data set (Jiang,
2020).
Input variable |
Sample size |
Minimum |
Maximum |
Average value |
Standard deviation |
UCS (MPa) |
150.000 |
20.000 |
90.000 |
52.567 |
20.964 |
BTS (MPa) |
150.000 |
1.000 |
10.000 |
5.200 |
3.322 |
Kv |
150.000 |
0.450 |
0.850 |
0.622 |
0.120 |
CAI |
150.000 |
0.750 |
3.490 |
2.034 |
1.062 |
E0 (GPa) |
150.000 |
3.000 |
14.000 |
8.447 |
3.919 |
μ |
150.000 |
0.200 |
0.310 |
0.252 |
0.041 |
PR (mm/min) |
150.000 |
34.800 |
75.750 |
56.541 |
9.560 |
RPM (r/min) |
150.000 |
4.000 |
7.600 |
5.907 |
0.877 |
TF (MN) |
150.000 |
2.100 |
9.750 |
5.812 |
2.082 |
Percentage distribution of rock mass classes in the sample data set: (a) Training data set and (b) test data set.
Percentage distribution of different lithologies in the sample data set: (a) Training data set and (b) test data set.
Violin with box depicting input variables for three rock mass classifications in the sample data set. (a) Represents uniaxial compressive strength, (b) represents uniaxial tensile strength, (c) represents rock wear resistance index, (d) represents rock integrity index, (e) represents deformation modulus, (f) represents Poisson's ratio, (g) represents net tunnelling rate, (h) represents rotation speed, and (i) represents cutterhead thrust.
The training data set shows variation in the percentages of different rock mass classes, as depicted in Figure 4. The percentages of various rock mass classes in the test data set, however, correspond to the real engineering situation shown in Figure 4. The correspondence between rock mass classes in the test data set and the actual engineering scenario is advantageous for a more thorough examination of the model's predictive performance. As shown in Figure 5, there are variations in the percentages of different lithologies in the training data set. However, the percentages of the test data set for various lithologies remain unchanged. It is valuable to assess the model's predictive performance on rock mass classes based on the consistency across various lithologies. As depicted in Figure 6, violin with box depicting input variables for three rock mass classifications in the sample data shows significant variation among the three rock mass classifications. However, from Figure 6a,c,e,f,g,h,i, it becomes evident that there is an overlap in certain input variables between Class III and Class IV rock masses. This overlap may affect the predictive performance of the subsequent model for Class III and Class IV rock masses.
3.3 PCA dimension reduction
First, the raw data are standardized according to the criteria presented in Equation (
14). This process counters the impact of various dimensions among input variables in the sample data set and eliminates interference from data with either large or small absolute values on the analysis results.
(14)
In the above equation,
represents the normalized variable value;
denotes the raw variable value;
is the average of the data for this set; and
is the standardized deviation of the data for this set.
The standardized data were subjected to PCA using SPSS software. The correlation coefficients between the various input variables are depicted in Figure 7. Additionally, Table 3 presents the cumulative contribution ratio and the principal component contribution ratio for each principal component. According to experience, a correlation coefficient reaching 0.5 indicates a particularly strong correlation between input variables, suggesting potential mutual influence. From Figure 7, it is evident that many correlation coefficients among the input variables exceed 0.5, potentially influencing the prediction results of rock mass classification.
Correlation coefficients among various input variables.
Table 3. Eigenvalues and cumulative contribution ratio for each principal component.
Component |
Eigenvalue |
Principal component contribution ratio (%) |
Cumulative contribution ratio (%) |
1 |
7.0530 |
78.37 |
78.37 |
2 |
1.1200 |
10.00 |
88.37 |
3 |
0.3661 |
4.07 |
92.44 |
4 |
0.2822 |
3.13 |
95.57 |
5 |
0.1443 |
1.60 |
97.17 |
6 |
0.1230 |
1.37 |
98.54 |
7 |
0.0658 |
0.73 |
99.27 |
8 |
0.0455 |
0.51 |
99.78 |
9 |
0.0199 |
0.22 |
100.00 |
The first two principal components have a cumulative contribution ratio of 88.37%, as indicated in Table 3. This means that the first two principal components together account for 88.37% of the original data, satisfying the requirements since their respective eigenvalues are greater than 1. Therefore, these two principal components are selected for further analysis. Table 4 illustrates the variance extraction ratio of raw variable information by the first two principal components.
Table 4. Variance extraction ratio of raw data by the first two principal components.
Parameter |
Principal component F1 |
Principal component F2 |
UCS (MPa) |
0.971 |
0.024 |
BTS (MPa) |
0.934 |
0.154 |
Kv |
0.719 |
0.594 |
CAI |
0.913 |
−0.285 |
E0 (GPa) |
0.878 |
−0.351 |
μ |
−0.846 |
0.501 |
PR (mm/min) |
−0.902 |
−0.169 |
RPM (r/min) |
0.867 |
0.118 |
TF (MN) |
0.916 |
0.157 |
As depicted in Table 4, the first and second principal components effectively capture the majority of information from the raw data. The first principal component is primarily associated with UCS, BTS, Kv, CAI, E0, μ, PR, RPM, and TF, while the second principal component is mainly related to Kv and μ. Following PCA preprocessing, the coefficients of component scores for each principal component are obtained, as presented in Table 5. The individual values for each principal component score are determined through a linear combination of the standardized raw data values with their respective principal component score coefficients, as outlined in Table 6.
Table 5. Coefficients of component scores for each principal component.
Parameter |
Principal component F1 |
Principal component F2 |
UCS (MPa) |
0.365 |
0.023 |
BTS (MPa) |
0.351 |
0.145 |
Kv |
0.270 |
0.560 |
CAI |
0.343 |
−0.269 |
E0 (GPa) |
0.330 |
−0.331 |
μ |
−0.318 |
0.473 |
PR (mm/min) |
−0.339 |
−0.159 |
RPM (r/min) |
0.326 |
0.111 |
TF (MN) |
0.344 |
0.148 |
Table 6. Scores associated with each principal component.
Data set order number |
Principal component F1 |
Principal component F2 |
1 |
4.54 |
0.67 |
2 |
3.87 |
0.47 |
3 |
4.86 |
0.80 |
4 |
2.32 |
−0.30 |
… |
… |
… |
148 |
−1.55 |
1.12 |
149 |
−3.08 |
0.61 |
150 |
−2.94 |
0.63 |
Given the slow computational time associated with the PCA dimension reduction technique using SPSS software, this study endeavors to improve the PCA–GRU model. These improvements are designed to expedite the prediction process for rock mass classification, ensuring timely and accurate results. To streamline the preprocessing steps for rock mass classification, a MATLAB program has been developed.
The program seamlessly integrates standardization and PCA dimension reduction techniques, facilitating the extraction of principal components F1 and F2 directly from the raw data. These processed principal components are subsequently input into the GRU model for computation, eliminating the need for cumbersome intermediate steps.
3.4 Performance indices
The key evaluation metrics, namely, accuracy (A), precision (P), and recall (R), are frequently used to assess the predictive performance of models. The accuracy (A) metric evaluates the proportion of correctly predicted outcomes among all observations made by the models. Precision (P) measures how many of the positive predictions were correct, emphasizing the accuracy of positive classifications. On the other hand, recall (R) highlights the models' sensitivity to positive instances by measuring the percentage of real positive instances that the models have correctly identified.
Equation (
15) for
A, (
16) for
P, and (17) for
R express these metrics. Using the previously mentioned equations, macro precision (
MP) and macro recall (
MR) are calculated to comprehensively evaluate model strengths and weaknesses. Macro recall (
MR) denotes the average recall of models across various rock mass classifications, while macro precision (
MP) indicates the average precision of models. This comprehensive assessment approach ensures a thorough and nuanced understanding of each model's overall performance in the classification of rock masses.
(15)
(16)
(17)
where
TP refers to instances in which the models correctly identify the true value;
FN denotes situations where the true value is correct, but the models predicate it as incorrect;
FP denotes situations where a value is mistakenly recognized as correct by the models, despite the true value being wrong; and
TN denotes cases where the models correctly identify an incorrect value, even though the true value is wrong.
4 RESULTS AND ANALYSIS
4.1 Selection of hyperparameters
The main hyperparameters for the PCA–GRU model, the GRU model, and the LSTM model include the hidden layer, dropout layer, fully connected layer, number of neurons, maximum number of training cycles, minimum batch size, dropout rate, initial learning rate, learning decay factor, and gradient threshold. The RF model is characterized by two hyperparameters: decision trees (DT) and minimum leaf size (LFmin). Similarly, the SVM model has two hyperparameters, namely, the penalty factor (C) and the radial basis function gamma (g).
The quantity of neurons and the number of hidden layers in the PCA–GRU, GRU, and LSTM models can affect how well the models predict outcomes. As indicated in Table 7, the aforementioned hyperparameters were adjusted. Additionally, a layer of dropouts with a ratio of 0.2 was incorporated into the models to prevent overfitting. A maximum of 1000 training epochs was set for the models, with a minimum batch size of 100. The fixed values were 0.2 for the learning decay factor, 1 for the gradient threshold, and 0.01 for the initial learning ratio. The fully connected layer was limited to 1, with four neurons, to simplify the adjustment of hyperparameters.
Table 7. Selection of hyperparameters.
Hyperparameter |
1 |
2 |
3 |
4 |
5 |
6 |
7 |
The number of hidden layers |
1 |
1 |
1 |
1 |
1 |
2 |
3 |
The number of neurons |
10 |
20 |
40 |
80 |
160 |
80 |
80 |
DT |
2 |
4 |
6 |
8 |
10 |
4 |
4 |
LFmin |
1 |
1 |
1 |
1 |
1 |
2 |
3 |
C |
2 |
4 |
6 |
8 |
10 |
10 |
10 |
g |
0.01 |
0.01 |
0.01 |
0.01 |
0.01 |
0.10 |
1.00 |
The minimum leaf size (LFmin) and the number of DT in the RF model can affect the model's ability to predict outcomes. The radial basis function parameter (g) and the penalty factor (C) in the SVM model may have an effect on the model's predictive ability. As a result, the hyperparameters mentioned above are modified as shown in Table 7.
As Figure 8 illustrates, the PCA–GRU model selects hyperparameter 6 with the best predictive performance, while the GRU and LSTM models select hyperparameter 7 with the best performance. By selecting hyperparameter 4, the RF model's predictive performance is at its best. The SVM model's predictive ability is not significantly affected by changes to its hyperparameters. The optimal hyperparameters are then determined by taking the values that correspond to each model's best predictive performance in the previously mentioned situations.
Predictive performance of models for different hyperparameters. (a) Represents accuracy, (b) represents macro precision, and (c) represents macro recall. GRU, gated recurrent unit; LSTM, long short-term memory; PCA, principal component analysis; RF, random forest; SVM, support vector machine.
Figure 9 shows the computational time of the PCA–GRU model for different hyperparameters. The PCA–GRU model takes approximately 20 s, as depicted in Figure 9. This quick processing time enables rapid responses for the classification prediction of rock masses in TBM tunneling.
Computational time of the principal component analysis (PCA)–gated recurrent unit (GRU) model for different hyperparameters. The numbers 1–7 stand for different hyperparameters.
4.2 Prediction results of five models
The predictions for the rock mass classification by the PCA–GRU model were compared with those of the GRU, LSTM, RF, and SVM models based on optimal performance corresponding to the best hyperparameters mentioned earlier in this paper. Confusion matrices for rock mass classification test data of the PCA–GRU, GRU, LSTM, RF, and SVM models are presented in Figure 10. The radar chart comparing predictions for rock mass classification by the PCA–GRU, GRU, LSTM, RF, and SVM models is shown in Figure 11. The numbers 1, 2, and 3 in Figure 10 correspond to the Class II, III, and IV rock masses, respectively. The confusion matrix's diagonal elements indicate how many correct classifications the model achieves for every specific class. On the other hand, misclassification is indicated by the off-diagonal elements, which show differences between the predicted and real classes. These matrix elements can be used to compute a variety of performance metrics, such as accuracy (A), precision (P), and recall (R), to fully assess the efficacy of the model.
Confusion matrices for rock mass classification test data of the principal component analysis–gated recurrent unit (GRU) (a), GRU (b), long short-term memory (c), random forest (d), and support vector machine models (e).
Radar chart comparing the rock mass classification prediction results of the principal component analysis (PCA)–gated recurrent unit (GRU), GRU, long short-term memory (LSTM), random forest (RF), and support vector machine (SVM) models. (a) Represents accuracy, (b) represents macro precision, and (c) represents macro recall.
The PCA–GRU model predicts rock mass classification with A, MP, and MR of 0.9667, 0.963, and 0.9763, respectively, as illustrated in Figure 11. These values are significantly higher than those of the GRU, LSTM, RF, and SVM models. The prediction results indicate that the PCA–GRU model outperforms the GRU, LSTM, RF, and SVM models in terms of overall performance in predicting the rock mass classification.
The PCA–GRU model shows high precision (P) and recall (R) in predicting rock masses of Class II, III, and IV, as illustrated in Figure 10. On the other hand, the GRU, LSTM, RF, and SVM models show lower precision (P) for Class III rock masses and lower recall (R) for Class IV rock masses. Certain input variables overlap, explaining why models like GRU, LSTM, RF, and SVM have lower recall (R) in Class IV rock mass prediction and lower precision (P) in Class III prediction. Among these variables are UCS, CAI, E0, μ, PR, RPM, and TF in relation to these two rock mass classes. The models may incorrectly classify Class IV rock masses as Class III due to this overlap.
The PCA dimension reduction technique has reduced the nine raw input variables to two comprehensive indicator variables that accurately represent the original data's features. This aids in eliminating input variable overlap in the distribution of Class III and IV rock masses. As a result, Class II, III, and IV rock masses can be successfully identified using the PCA–GRU model. The prediction outcomes demonstrate how effectively the PCA–GRU model predicts the rock mass classification of TBM projects.
4.3 Random sampling analysis
The model's ability to generalize may be impacted by the uneven distribution of various rock mass classes and lithologies, which change in percentage between the training and test data sets. To test the model's generalization, the ratio of the training data set to the test data set is maintained at 8 : 2, and both the training and test data sets are shuffled at random. Ten randomly shuffled data sets are used to compare the predictions so as to improve the reliability of the results. Figure 12 shows the model predictions obtained from 10 data sets that are shuffled at random.
Model predictions based on 10 randomly shuffled data sets. (a) Represents accuracy, (b) represents macro precision, and (c) represents macro recall. GRU, gated recurrent unit; LSTM, long short-term memory; PCA, principal component analysis; RF, random forest; SVM, support vector machine.
The predictive performance metrics A, MP, and MR of the PCA–GRU model are clearly higher than those of the GRU, LSTM, RF, and SVM models, as shown in Figure 12. The results of the research suggest that the PCA–GRU model performs better in terms of generalization than the GRU, LSTM, RF, and SVM models. This makes it more appropriate for conditions where the distribution of various rock mass classes and lithologies change in percentage.
ACKNOWLEDGMENTS
This study was funded by the National Key R&D Program of China, Grant/Award number: 2018YFC1504801, 2018YFC1504902; the National Natural Science Foundation of China, Grant/Award number: 51522903, 51774184; The State Key Laboratory of Hydroscience and Hydraulic Engineering of Tsinghua University, Grant/Award number: 2019-KY-03; the Yujie Project of North China University of Technology, Grant/Award number: 216051360020XN199/006; the Key Technology of Intelligent Construction of Urban Underground Space of North China University of Technology, Grant/Award number: 110051360022XN108-19; and the Research Start-up Fund Project of North China University of Technology, Grant/Award number: 110051360002.