Abstract

The existing pavement performance prediction methods are limited to single-factor predictions, which often face the challenges of high cost, low efficiency, and poor accuracy. It is difficult to simultaneously solve the temporal, spatial, and exogenous dependencies between pavement performance data and maintenance, the service life of highways, the environment, and other factors. Digital twin technology based on the building information modeling (BIM) model, combined with machine learning, puts forward a new perspective and method for the accurate and timely prediction of pavement performance. In this paper, we propose a highway tunnel pavement performance prediction approach based on a digital twin and multiple time series stacking (MTSS). This paper (1) establishes an MTSS prediction model with heterogeneous stacking of eXtreme gradient boosting (XGBoost), the artificial neural network (ANN), random forest (RF), ridge regression, and support vector regression (SVR) component learners after exploratory data analysis (EDA); (2) proposes a method based on multiple time series feature extraction to accurately predict the pavement performance change trend, using the highway segment as the minimum computing unit and considering multiple factors; (3) uses grid search with the k-fold cross validation method to optimize hyperparameters to ensure the robustness, stability, and generalization ability of the prediction model; and (4) constructs a digital twin for pavement performance prediction to realize the real-time dynamic evolution of prediction. The method proposed in this study is applied in the life cycle management of the Dalian highway-crossing tunnel in Shanghai, China. A dataset covering 2010–2019 is collected for real-time prediction of the pavement performance. The prediction accuracy evaluation shows that the mean absolute error (MAE) is 0.1314, the root mean squared error (RMSE) is 0.0386, the mean absolute percentage error (MAPE) is 5.10%, and the accuracy is 94.90%. Its overall performance is better than a single model. The results verify that the prediction method based on digital twin and MTSS is feasible and effective in the highway tunnel pavement performance prediction.

1. Introduction

In early 2020, China proposed a plan to accelerate the modernization of urban governance with advanced application of technologies, such as the Internet, big data, and artificial intelligence to support the transformation and upgrading of traditional infrastructures, and merge into innovatively infrastructures, such as intelligent transportation infrastructures. According to government statistics, as of the end of 2019, the summed length of highway tunnels of China reached 18.9666 million meters. To comprehensively improve the management level of urban highway tunnels, academia and industry are paying great attention to the life cycle management of urban highway tunnels based on digital twins. A digital twin, proposed by Michael Grieves in 2002, is a mapping relationship between the physical space and virtual space and is intelligent. It maps, records, simulates, predicts, and manages the running track of the life cycle of objects in the physical world and digital virtual space. The life cycle management of urban highway tunnels emphasizes the transformation from traditional corrective maintenance to data-driven preventive maintenance, in which the accurate and timely prediction of highway pavement performance plays a crucial role.

There are two types of research methods in the field of pavement performance prediction. One type is based on traditional statistical models, and the other type is derived from modern machine learning models. Methods based on traditional statistical models are relatively extensive, including deterministic methods such as mechanical prediction models and mechanical-empirical prediction models [16], as well as probabilistic methods such as Markov probability models and Bayesian probability models [710]. The deterministic model is based on empirical classification, mechanical analysis, or mechanism analysis to predict the pavement performance. Eghbalpoor [1] et al. proposed a three-dimensional finite element model of asphalt pavement to study the damage to the pavement caused by high cyclic loads and their healing effect. Zhang [2] et al. proposed a research method based on morphological dynamics to predict the temperature stress change trend of long-term aging asphalt pavement materials. Shirzad [4] et al. explored the influence of the asphalt mixture ratio on pavement performance and established a mechanical experience model to predict rutting. The probabilistic model is an empirical model dominated by a large amount of measured data. Bianchini [8] et al. proposed an IF-THEN fuzzy neural inference model to predict pavement performance. Li [11] et al. used the method of K-means and type-2 fuzzy set to predict the long-term traffic volume. Chen [7] et al. used a Markov chain Monte Carlo (MCMC) simulation to analyze the failure probability of pavement prevention and treatment. Abed [10] et al. used the Monte Carlo simulation method to predict the change probability of pavement performance. The existing traditional statistical models are limited to single-factor predictions, in which single independent variables can hardly reflect the essence of pavement performance. They usually focus on the statistical inference and interpretability of a small sample under certain hypothesis testing conditions. They usually use a small sample size, mainly to represent the relationship between the data and the result variables and the importance of these relationships, and the calculation accuracy is limited. With the increase in the service life of urban highway tunnels, pavement performance is subjected to long-term coupling effects of loads, maintenance situations, environmental conditions, material properties, and other factors, and the resulting pavement diseases aggravate the uncertainty of the driving comfort and safety of the pavement. The predictive effect of traditional statistical methods is sometimes suboptimal in practical engineering.

Studies have shown that machine learning models are superior to traditional statistical methods, such as curve fitting, exponential smoothing, and the autoregressive moving average methods. The accuracy of prediction can be optimized through feature selection, hyperparameter tuning, and integrated learning technology. Gong [12] et al. developed a random forests regression (RFR) model to predict the international roughness index (IRI) performance of flexible pavement based on long-term historical time series data information of disasters, traffic, gas, and structure. The comparison found that it was significantly better than the regularized linear regression model. Xiao and Nie [13] used the regression model and time series GM (1, 1) to predict pavement performance. The maximum difference between the time series predicted value and the regression model predicted value can be reduced by 1%. Bi [14] et al. proposed a time series algorithm suitable for single-factor distributed version to predict the workload of the next stage. Lintonen and Raty [15] proposed a new self-learning stopping criterion for multivariate time series which is a positive-unlabeled (PU) learning case. Liu [16] et al. improved the accuracy of pavement temperature prediction by using a gradient boosting extreme learning machine (GBELM). Gong [17, 18] considered parameters such as materials, structure, traffic, and environment and successfully developed the Mechanistic-Empirical Pavement Design Guide (MEPDG) deep neural network model and GBM method for pavement rutting prediction, and the accuracy was improved. Hu [19] obtained risk indicators from a neural network prediction model and found that the application of data analysis improved the similarity of the predictive performance of the overall pavement safety level. Wang [20] et al. proposed a training method to automatically determine the parameters of network structure for the deep belief network (DBN) with transfer learning (TL-GDBN). Gao [21] et al. proposed 6 non-BP swarm intelligence search algorithms (BBO, PSO, GA, ACO, ES, and PBOL) for model training and developed a new dendritic neuron model (DNM) that considers nonlinearity. Wang and Kumbasar [22] used two swarm intelligence algorithms (PSO and BBBC) to optimize parameters. The machine learning model provides a new method for effectively dealing with multiple factors, ambiguity, randomness, and other issues. It can consider multiple factors and processes high-dimensional data at the same time. It focuses more on the balance between accuracy and complexity. The training set learns the predictive model, and the testset evaluates and verifies its accuracy. No hypothesis conditions are required. The larger the sample size is, the faster the model converges. The generalization integration framework [23], proposed by Wolpert in 1992, is a serial multilevel learning structure with flexibility and uncertainty that can achieve both bagging integration and boosting integration. Lin [24] et al. proposed a stacking model (SMVP) based on real datasets for predicting changes in public bicycle traffic flow. Compared with the traditional single model, the certainty level is increased by 25.58%. Zhai [25, 26] et al. predicted and analyzed the average daily polyvinyl acetate (PM2.5) concentration in Beijing, China, and developed a stacked ensemble model, including the stacking of LASSO, AdaBoost, XGBoost, the GA-MLP-optimized multilayer perceptron, and the support vector regression model (SVR). The results showed that the stacked combination model had high interpretability. Jiang [27] et al. also proposed a stacked ensemble model, which used RF, ERT, XGBoost, LightGBM, RNN, bidirectional RNN, LSTM, and GRU as basic learners, and used logistic regression as the second-level metalearner. In different application scenarios, the stacked models constructed by researchers such as Li [28], Yang [29], and Feng [30] et al. had good results, suggesting an idea for establishing stacked models to predict pavement performance in the field of urban highway tunnel pavement performance prediction.

In the field of architecture, engineering, construction (AEC), BIM has proved to be an intelligent and parametric digital modeling method that can support the full life cycle of buildings. Since 2019, researchers have begun to conduct applied research on digital twins in the AEC field based on the BIM model [31, 32]. A digital twin refers to a virtual model that is completely consistent with physical entities in the real world and can simulate an entity’s behavior and performance in a real environment in real-time. Digital twin technology has the characteristics of traceability, high integration, real-time, and high precision [33, 34]. In the design phase, digital twins can be used for model design, collision detection, pipeline synthesis, and hardcover design. Lu [35] et al. proposed a semiautomatic digital twin system based on images and CAD drawings, aiming at the time-consuming problem of BIM construction. In the construction phase, digital twins can be used for cost budgeting [36], quality management [37], construction collaboration [38], and schedule management [39]. In the operation and maintenance phase, digital twins can be used for equipment asset management [40], safety and prevention management, building space management, and building environment management. Shim [40, 41] et al. proposed a bridge maintenance program based on the concept of digital twins. The program combines a 3D information model-based maintenance information management system with a digital detection system that uses image processing to enhance the decision-making process in terms of the bridge maintenance process reliability.

Therefore, this paper proposes a tunnel pavement performance prediction method based on a digital twin + MTSS for the life cycle operation and maintenance management of urban highway tunnels. Considering many factors, such as the maintenance condition, the service life of highways, and the environment, along with the actual pavement performance data of urban highway tunnels, this method adopts machine learning methods to improve the prediction accuracy, which is conducive to extending the service life of urban highway tunnel pavement, and provides support for operational management. The remainder of this paper is organized as follows: the second section introduces the digital twin for pavement performance prediction and the MTSS prediction model. The third section applies the proposed method in practical engineering projects and compares it with the single model to evaluate the performance of the MTSS model and verifies the applicability of the method. Finally, the fourth section summarizes the whole paper.

2. Theoretical Basis of Regression Models

2.1. EXtreme Gradient Boosting (XGBoost)

XGBoost is a highly scalable gradient boosted decision tree system. It has the advantages of good robustness and strong scalability. In recent years, it has shown excellent performance in many fields, such as information technology and software engineering [42], environmental science [30], and economics and finance [36]. XGBoost performs a second-order Taylor expansion on the loss function and uses a quadratic function to approximate the loss function. To prevent overfitting, regularization is introduced into the loss function to constrain the number and weight of leaf nodes, balance the complexity of the model, and optimize each iteration until the following equation is minimized [43].where , is the loss function, is the regularization term, is the number of leaf nodes, and and are the regularization parameters.

2.2. Gradient Boosting Decision Tree (GBDT)

GBDT is an ensemble method of boosting. It uses the CART decision tree as the basic learner. The GBDT is different from RF that RF can be generated in parallel by the voting method, while GBDT is an ensemble of weight-based weak classifiers, which is the accumulation of the branch results of each tree, and can be generated only serially. Therefore, the GBDT is more concerned with reducing the bias in the training process, while the RF is more concerned with reducing the variance in the training process. At the same time, GBDT has good robustness and high scalability [44].

2.3. Light Gradient Boosting Machine (LightGBM)

LightGBM is an open-source distributed gradient boosting framework based on the decision tree algorithm developed by the Microsoft Research Asia Distributed Machine Learning Toolkit (DMTK) team [45]. It is an improved decision tree integration model based on GBDT and XGBoost [46]. Two aspects are improved. (1) Gradient-based one-sided sampling (GOSS): GOSS algorithm sorts the gradients. Samples with a large gradient are selected by percentage, and those with a small gradient are randomly selected to optimize the sampling of the training sample set and reduce the sample points when calculating the gradient. (2) Exclusive feature bundling (EFB): mutually exclusive features are bound to reduce the feature dimension when selecting split points. In a data environment with large training samples and high-dimensional features, it exhibits a faster training speed and efficiency, requires lower memory usage, offers better accuracy, can handle larger-scale data, and supports parallel learning.

2.4. Random Forest (RF)

RF is an ensemble bagging model proposed by Breiman [47] that combines the CART tree and random space method without assuming a prior distribution. Each tree is built by the independent random sampling method. The original data subset is sampled and generated by the bootstrap method, and a random subset of the feature attributes is randomly choosed according to a specific criterion. The selection of the training set is random, and each training set is independent of the others. Random forests generally perform better than single decision trees because the results of random forests are determined by voting on the results of multiple decision trees. Each decision tree in a random forest has its own result, and the random forest selects the result with the largest number of votes as its final result by counting the results of each decision tree.

2.5. Linear Regression

The linear regression method assumes that the relationship between independent variables and target variables is linear. This simplicity often makes linear regression the best choice for small sample analysis and also makes these models relatively easy to interpret and understand, not suitable for too many predictive variables [48].

2.6. Least Absolute Shrinkage and Selection Operator (LASSO)

The least absolute shrinkage and selection operator (LASSO) is a regression analysis method with strong robustness and a strong generalization ability for simultaneous feature selection and regularization. The norm is introduced to generate sparse solutions at the feature level, which will shrink the unimportant coefficients in the parameters to be estimated, so as to reduce redundancy and make the model more stable when processing different data, thus improving the accuracy and robustness of regression prediction [25].

2.7. Ridge Regression

Ridge differs from LASSO in terms of regularization items, and Ridge introduces the norm. Ridge regularization uses the sum of squares of regression coefficients as a penalty function and compresses regression coefficients, making it difficult to compress the coefficients of redundant predictive variables to 0 and easier to overcompress more important regression coefficients [49].

2.8. Artificial Neural Network (ANN)

The ANN structure usually consists of three main parts: the input layer, the hidden layer, and the output layer. Each layer is composed of many neurons. The input layer is used to load known parameters and directly determine the output results [25], such as performance index and maintenance type in this study. The hidden layer contains all known information and establishes a nonlinear relationship between the input parameters and the output results. The output layer is used to obtain the unknown prediction result Y. As shown in Figure 1, this study has set up 2 hidden layers with 128 neurons in each layer. The selection range of the hidden layers was set as [32, 64, 128], and the nonlinear hyperbolic tangent (tanh) activation function was adopted. The detailed neural network structure diagram is as follows:

2.9. Support Vector Regression (SVR)

The core principle of the support vector machine regression algorithm SVR is to find a regression plane so as to minimize the overall observation error and maximize deviation [50]. The traditional regression methods consider the prediction correct if and only when the regression fitting function is completely equal to . For example, is commonly used to calculate the loss in linear regression [51]. However, SVR believes that as long as deviates from not too much, the prediction can be considered correct. Instead of calculating losses, the threshold a is set to only calculate the loss of the data point . The SVR model is irrelevant of the dimension of the input space and depends on the number of SVs. The goal of SVR is to solve the following basic problems [25]:where and are used to determine the fitting function , is the regularization constant, and are the relaxation variables to define the up and down deviation.

3. Methods

3.1. Digital Twin for Pavement Performance Prediction

This paper proposes a digital twin for the pavement performance prediction of urban highway tunnels. It is driven by the continuous and circular flow of information and data from the physical world to the digital world and then to the physical world. It adopts a new method by the deep integration of global perception technology and machine learning to establish the digital mapping between the virtual pavement BIM model and physical highway entity. It realizes data sharing, visual analysis, and the intelligent decision support functions of pavement performance trend changes. The digital twin consists of three key parts (Figure 2). (1) Data collection module: the data include pavement performance, maintenance records, tunnel highway structure, traffic flow, and tunnel environment. The pavement performance data come from pavement performance indicators, such as vehicle-mounted laser sensors and high-precision accelerometers. The maintenance records are stored in the operation and maintenance management system. (2) Prediction module: within the stacking integrated learning framework of the MTSS, a pavement performance prediction model based on multiple differentiated models is established to predict the future performance trend change of each segment of the pavement. (3) Parametric analysis module: in the Dynamo environment, which supports visual parameter design, the multisource spatial-temporal data of the BIM model and physical entities are integrated, and the inheritance relationship between the node components is constructed. A Python application is implemented with RevitAPI to realize the visual evolution of the pavement performance.

Figure 3 shows the learning process of digital twins. First, the pavement performance data of the physical environment are collected, and the datasets are divided to build the MTSS prediction model. Then, in the Dynamo environment, the physical pavement entity and the virtual digital BIM model are mapped, the predicted data are synchronized, and the BIM model is assigned. Both the virtual environment and the physical environment contain the bidirectional synchronization process of data acquisition and realization of pavement performance prediction state change [52]. Among them, in the visualization state of pavement performance prediction, the physical pavement entity changes through the digital mapping by using the virtual pavement BIM model. This state change is transferred to the parameter analysis module through the MTSS prediction model. Because the pavement segments may be degraded in the future, the digital twin is used to visualize the decay trend of pavement performance and provide the decision analysis service of the highway tunnel pavement performance prediction and maintenance decision-making by predicting the maintenance timing, analyzing the maintenance cost, and adopting preventive maintenance strategy. The ultimate purpose of all these measures is to extend the life of the highways.

3.2. Multiple Time Series Stacking Prediction Model
3.2.1. Time Series Feature Extraction Based on Multiple Time Series

Feature extraction is very important for the performance and generalization ability of the model. It is used to trim, transform, and refine the input factors, which can improve the interpretability and prediction accuracy of the prediction model [53]. The feature engineering used in this paper includes temporal feature extraction, feature transformation, feature filtering, and feature standardization and one-hot for feature transformation and model-based method for feature filtering [54]. In the preliminary exploratory data analysis (EDA) of the dataset, it is found that the pavement performance indicators, maintenance information, and environmental information dataset have the basic characteristics of the time series. However, it is difficult to process multiple time series in parallel.

The general feature filtering method filters only by the correlation between a single factor and the dependent variable, while the model-based feature filtering method filters by the comprehensive effect of each factor, which can make full use of the combinative characteristics among factors and retain more useful factors. In view of the rapid increase in the number of features as well as feature extraction of time series information, in order to avoid feature collinearity and decreasing of the prediction accuracy by adding useless information, a model-based feature filtering method was adopted to filter the redundant features according to the 95% cumulative contribution threshold (proportion) of the feature factors. Therefore, this paper proposes a method based on multiple time series feature extraction to construct and enhance features and uses a tree-based model to select potentially important features to improve the accuracy of the model.

The time series feature refers to the information contained in the historical sequence. The traditional regression prediction method uses the feature data of the time section at time to predict the target value at time . This method uses only the data of a single time section. However, although the time series method can make use of the characteristics of historical sequences over a period of time, it applies only to the characteristics of a single series and has a strong restriction on the richness of the features. The multiple time series feature extraction method proposed in this paper combines historical sequence information with multiple factors and simultaneously uses the temporal features of multiple factors as the factor input of the regression model, that is, the factor feature data from time to time are used to predict the target value at time . This feature extraction method can effectively combine the features of multiple heterogeneous factors at the same time. Due to the use of temporal features, the change pattern along the time dimension (such as the curvature) of factors can be fully utilized, which can provide richer and more valuable information for prediction and improve the prediction accuracy.

The temporal feature extraction process is shown in Figure 4, where is the target time series, is the characteristic time series, is the predicted target scalar value of the corresponding sample at time , and is the factor input of the corresponding sample at time . is a vector composed of to characteristic data extracted from each characteristic time series by characteristic time series. The sample formed by feature extraction constructs the samples of the regression model.

3.2.2. Filter for Component Learners

In order to make the stacking learning model achieve the optimal prediction effect and improve the robustness of the model, not only the predictive ability of a single individual learner must be considered but also necessary to consider the combined effects of individual learners. Individual learners with stronger learning ability will help improve the overall predictive effect of the model. In the first level, heterogeneous models with a large degree of difference are selected as the component learners, which can maximize the advantages of various algorithms, and at the same time, each differentiated model can learn from each other’s strengths [55]. Therefore, this paper conducted a stacking experiment to optimize the stacking method of the component learners when building the model. A total of 9 individual learners, XGBoost, GBDT, LightGBM, RF, Linear, LASSO, Ridge, ANN, and SVR, are initially selected. The essence of the stacking of different models is to observe data from different data space perspectives and data structure perspectives. Based on this principle, base learners should be different from each other. Therefore, this paper uses the Pearson correlation coefficient to calculate the differences of each model. By selecting algorithms with large differences, we maximize the advantages of different algorithms and obtain the best prediction effect. Pearson correlation formula is shown as equation (3), where X and Y are two single models’ outputs.

The stronger the learning ability of each individual learner in stacking and the lower the degree of correlation, the better the prediction effect of the stacked model. The threshold value of correlation between component learners is set to be 0.95. Above the threshold, the correlation is stronger, and less than the threshold, the correlation is weak. As shown in Figure 5, XGBoost, ANN, RF, Ridge, and SVR are selected as the first level of component learners, and SVR as the second level of metalearner. XGBoost has a higher correlation with GBDT and LightGBM, which are both trees boosting algorithms. The correlation between XGBoost and RF is relatively low. This is because although RF is also a tree-based integrated class model, it is a bagging type. The correlations between Ridge and LASSO and linear is strong because Ridge and LASSO are both improved versions of linear regression. ANN is quite different from other models, so the correlation is low.

After the filtering of correlation analysis and combination optimization experiment, as shown in Figure 6, this paper selects XGBoost, ANN, RF, Ridge, and SVR as component learners of level 1 and SVR as metalearner of level 2.

3.2.3. Fusion of the MTSS Prediction Model

This paper first uses the k-fold (k = 5) method to train each component learner with the training set and then uses the predicted values from the testset of component learners to train the metalearner. Stacking ensemble learning divides the original dataset into several data subsets, where is the input vector of the sample with th temporal features, is the corresponding predicted value of the sample , and is the number of temporal features, that is, each feature vector is . In this paper, the five-fold cross validation method is used. This paper randomly divides the data into five equal size and disjoint subsets of the original dataset, where defines and as the -fold training set and testset, respectively. One of them is taken as the testset, and the other four are used as the training set to train the model each time. is calculated based on the testset. Then, the average value is calculated to obtain :

The process of stacking learning is as follows:Step 1Divide the original sample data into training data and holdout data and divide the training data into K subsets of basically equal sizeStep 2Uses the k-fold cross validation method to train each individual learner on the training data and validate it on the holdout dataStep 3Repeat step 2 until all the component learners are trainedStep 4Use the first level output as the input of the second-level metalearner and train the second-level metalearner, namely, the meta-model (SVR)

As shown in Figure 7, the first level prediction algorithm contains base learners. The training set is trained with the th algorithm to obtain the base learner . In k-fold cross validation, for each sample in the -fold testset , the prediction of the base learner is denoted as . After cross validation, the output data of base learners form a new data sample , which is used as the input data of the second level. Train the metalearner of the second-level prediction model to detect and correct the prediction errors in the first layer learning algorithm. The SVR model in the second level outputs the final prediction results. The stacking learning framework improves the robustness and generalizability of the prediction model by combining the output predictions of multiple models to obtain an improvement in the overall prediction accuracy. In order to prevent overfitting, the training set of the component learner cannot be directly used in the training of the metalearner. It is necessary to divide the original data reasonably.

3.2.4. Hyperparameter Optimization

The multilevel stacking learning system involves the setting of many parameters. Nevertheless, the combination of parameters of different values has a greater impact on the prediction accuracy of the model, and unreasonable parameter settings can easily cause underfitting or overfitting of the model. Hyperparameter tuning is an optimization method that can be carried out to preprocess or postprocess any learning algorithm. Grid search can improve the generality of the model. To make the MTSS model perform optimally and prevent underfitting or overfitting, it is necessary to optimize the parameters of each base learner and metalearner. First, the single base learner is predicted and analyzed on the original dataset. The regularization technique is used to avoid overfitting. The k-fold cross validation is adopted to divide the dataset into multiple training sets and testsets. The training set is used to construct the model, and the testset is used to evaluate or verify the performance of the model. The structure of the base learner and metalearner algorithm has certain complexity. It is necessary to define many parameters, such as learning_rate, max_depth, and n_estimators. Among the main parameters, learning_rate can control the model weights’ updating speed. Max_depth is the maximum depth of a tree. The larger the max_depth is, the more complex the model will be. N_estimators is the number of decision trees. Gamma refers to the complexity penalty term at the level of tree structure. If the value is larger, the number of leaf nodes in the tree will be smaller. Lambda represents the penalty for the output weight of leaf nodes to avoid excessive weight of leaf nodes and improve the robustness of the model. The stacking model tuning is different from single model tuning. First, the original sample data are divided into training set and holdout dataset. Second, using the training set, the k-fold cross validation is combined with the grid search method to optimize the parameters of each component learner. In this paper, k is set as five, and the training set is divided into five parts, that is, there are five different subtraining sets and subtestsets. For each component learner, five models are generated based on the training of subtraining sets, respectively, and these five models are, respectively, used to predict their corresponding subtestsets. The optimal parameters are selected with the best five models’ testset average accuracy. The parameter range of grid search is carried out according to the hyperparameter set of each model in Table 1. Then, based on the training set generated by the first level model, the parameters of the second level model are also optimized by the combination of k-fold cross validation and the grid search method. The total stacking model is validated on the validation set generated by the first layer model. Finally, the experimental results show that the parameter tuning method can effectively optimize the parameters of the stacking model, and the prediction effect of the stacking model has been improved. The hyperparameter range of each model algorithm grid search method is shown in Table 1.

3.3. Parametric Analysis Module

The parametric analysis module is based on Dynamo’s visual programming environment, which associates the physical tunnel entity with the BIM model by constructing the mapping relationship between the components. It realizes the integration of real-time dynamic evolution visualization of pavement performance prediction and data science and provides support for managers’ decision-making. The workflow is shown in Figure 8, including six steps: importing the Revit highway BIM model, creating highway segment elements, importing a Python script, importing MTSS highway segment performance prediction results, creating a chromatogram, and visualizing the changing trend of highway segments. First, this paper collects the data information of the 3D BIM model created in Autodesk Revit. Then, we link the MTSS prediction model as a custom node to the Dynamo environment. Finally, the predicted pavement performance trend is displayed in the form of graph visualization to support the decision-making reasoning process.

Step 1Import the Revit highway BIM modelDynamo is a lightweight model engine with a series of built-in graphics and bidirectional conversion nodes [56]. In this paper, the four nodes categories, all elements of a category, element geometry, PolyCurve, and ByjoinedCurves, which are wrapped by RevitAPI, are used to realize the component function of importing the highway BIM model into the dynamo engine from Revit.Step 2Create highway segment elementsAs shown in Figure 8, according to the basic principle of bounding box regression in computer graphics, the bounding box of the highway BIM model in Figure 9(a) is created, and the computational element with the center point of the cube as the minimum (Figure 9(c)) is generated to realize the cut split and create the component function of highway segment elementsStep 3Import Python script to highway segment elementsThe bounding box-generated border range is much larger than the site boundary of the BIM model of the highway surface (Figure 9(b)), and graph screening is more complicated than a numerical screening calculation. Therefore, to improve the efficiency of graphic calculation, this article customizes the Python script and links it to Dynamo by calling the DesignScript library, judging the position of the unit cube according to the projective method and filtering out the unit cubes belonging to the pavement boundary as a highway segment element.Step 4Import MTSS highway segment performance prediction resultsThrough the file. Path node (Figure 8(d)), the dataset of the MTSS model prediction results is imported into Dynamo in the form of an Excel file to provide data support for the visualization of the pavement performance prediction digital twinStep 5Create a chromatogramCreate a gradient color chromatogram based on the three primary colors of red, yellow, and blue, map the predicted value to the chromatogram in Figure 10, and display the changes in the performance of the physical image elements of the highway segment through colors to enhance the visualization of the pavement performanceStep 6Visualize the changing trend of highway segmentsTaking each highway segment as a unit, finite element analysis is used to show the future change trend of the current highway segment. The red part indicates that the pavement performance will be reduced, and corresponding maintenance measures need to be taken. The pavement performance trend should be improved. Finally, according to the numerical results of the pavement performance prediction, the future trend of the highway segment is visually displayed in different colors to provide support for maintenance decision-making.

4. Case Study

4.1. Materials

With a total length of 2500 meters and four lanes in both directions, the Dalian highway tunnel in Shanghai is the first cross-river project in China, which simultaneously adopts two large-scale shields to promote the new technology. This paper uses the historical dataset of the Dalian highway tunnel, including maintenance information and measured IRI data from 2010 to 2019. After data preprocessing, 80% of the dataset is randomly selected as the training set to construct the model, and the other 20% is used as the testset to evaluate and verify the performance of the model. Then, this paper uses the trained model to predict the pavement performance in 2020.

4.2. Data Processing

Whereas the maintenance data of the Dalian highway come from the records of special projects and subprojects over the years, the dataset only records information such as east-west lines, fast and slow lanes, and light numbers. There is no direct record of highway segment elements, which does not meet the data requirements of the MTSS prediction model. Therefore, according to the actual tunnel space structure and the light number arrangement rules, this paper maps each maintenance record to the highway segment elements in accordance with the recorded natural decay, minor repair, or overhaul maintenance to solve the problem that the current maintenance record does not satisfy the prediction model to subdivide the highway into segments.

4.3. Result of Feature Selection
4.3.1. Results Based on Temporal Feature Extraction

The measured data of pavement roughness are affected by factors such as the service life of highways, maintenance level, COVI, traffic volume, and other factors. Among them, the service life of highways, maintenance conditions, and history performance data have the most significant influence on pavement performance prediction. Clearly, the pavement performance prediction is affected by multiple time series, and the measured data have multiple temporal features. The study of Kitaha et al. showed that the performance change was more reliable than the value itself [57]. Therefore, this paper selects IRI of the previous five years as the feature factors describing IRI, defines as the maintenance feature factors, and defines as the feature factor of the service life of highways. The extraction process based on the time series features is shown in Figure 11. The results of the extraction are shown in Tables 2-3. The sample is composed of each record and the predicted value of the pavement performance.

4.3.2. The Significance of Feature Factors

To verify the temporal, spatial, and exogenous dependencies between pavement performance data and multiple factors, this paper uses the MTSS model to score , , and multiple feature factors in the year during the training process. The feature importance score is calculated based on the gain of the number of branches in the tree structure, mainly calculating the sum of the number of occurrences of features in all trees. If the number of occurrences is higher, the characteristic factor is more important. In Figure 12, the feature factors for 2020 are automatically ranked. According to the significance of the feature factors, it is found that the importance of maintenance features in the year accounts for the largest proportion, that is, the maintenance situation in the predicted year has the greatest impact on pavement performance prediction; , that is, pavement performance changes in the previous year have a greater impact on the predicted year; the overhaul has a greater impact on than the minor repair. The closer the value is to zero, the greater the feature importance is of . It can be seen that maintenance types have a great impact on pavement performance.

4.4. Analysis of MTSS Prediction Model Results
4.4.1. MTSS Prediction Model Evaluation

Without the loss of generality, to evaluate the accuracy of the MTSS prediction model, this paper compares the stacking method with a single SVR, Ridge, XGBoost, ANN, and RF machine learning models using the same data to conduct a multicriterion evaluation. MAE and RMSE are standard indicators for model evaluation in machine learning. Their values are often used to directly measure the pros and cons of the prediction model results, and they can also better reflect the degree of deviation of the predicted value from the true value. The lower the value is, the smaller the error and the higher the prediction accuracy. When the outliers have a very large degree of individual deviation, even if the number of outliers is small, the performance of the standard indicators will deteriorate. Therefore, this paper also adds an MAPE indicator to solve the problem of the robustness of the evaluation indicator by replacing the average with the quantile of the error. Suppose is the true value of sample , is the predicted value of sample , and is the number of samples, the evaluation index is defined aswhere denotes the average absolute error between the predicted value and the actual value, denotes the root mean square error, and denotes the average relative error.

In Table 4, the stacking model has the best performance in comparison with the single model. The MAE of the stacking model is only 0.1314, the RMSE is 0.0386, and the MAPE is only 5.10%. The results show that the accuracy of the predicted value obtained by the MTSS model can reach 94.90%, which is higher than that of other single models. In the single prediction model, the SVR prediction error is small and the accuracy is the highest.

Without considering multiple timing features, the MAE of stacking was 0.1808, the RMSE was 0.0643, and the MAPE is 7.23% (Table 5). Considering multiple timing features, the MAE of stacking was 0.1314, the RMSE was 0.0386, and the MAPE is 5.10% (Table 4). Regarding the benchmark performance without considering multiple timing features, the MTSS prediction model shows better results that MAE is decreased by 0.0494, RMSE is decreased by 0.0257, and MAPE is decreased by 2.13%. Without considering the timing characteristics, the accuracy of model prediction is significantly reduced. It can be seen that multitemporal feature extraction is effective for improving the accuracy of the model, and the supplement of historical information can improve the prediction accuracy of the model.

In summary, the stacking model integrates a variety of heterogeneous prediction algorithms and can use the advantages of multiple algorithms in different data spaces and structures to construct better models. Therefore, the MTSS model is better than a single machine learning in terms of pavement performance prediction. The MTSS model shows better performance.

4.4.2. MTSS Prediction Model Results

From the analysis of the forecast cycle of the entire pavement of the Dalian highway, Figure 13 shows the comparison trend of the actual IRI value in 2019 and the IRI predicted values under three different maintenance assumptions in 2020. The pavement performance will decline in turn under the assumption of overhaul, minor repair, and natural decay, which is in line with the normal law. In other words, the pavement performance after major repair is obviously higher than that after minor repair, and the pavement performance after minor repair is higher than that after natural decay.

According to the prediction of each segment in Figure 14, the IRI will decrease by 0.2 under the assumption of natural decay for the 1100 km slow lane from Puxi to Pudong (Eastern Route) in 2020. Under the assumption of minor repair, the IRI will be improved by 0.2. Under the assumption of overhaul, the IRI will be improved by 0.4. In 2020, on the 1100 km segment of the Puxi to Pudong (Eastern Route) expressway, the IRI decreases by 0.1 under the assumption of natural decay; under the assumption of minor repair, the IRI is improved by 0.2; under the assumption of overhaul, the IRI is increased by 0.6. The IRI will decrease by 0.05 under the assumption of natural decay for the 600 km slow lane from Pudong to Puxi (West Route) in 2020. Under the assumption of minor repair, the IRI will be improved by 0.3; under the assumption of overhaul, IRI will be improved by 0.6. The IRI will decrease by 0.1 in the 100 km segment of the expressway from Pudong to Puxi (West Route) in 2020 under the assumption of a natural recession. Under the assumption of minor repair, IRI will be improved by 0.2; under the assumption of overhaul, IRI will be improved by 0.6.

Figure 15 shows the comparison between the predicted value and the actual value of some highway segments. In scatter Figure 16, diagonals identify that the line X equals to Y, and the X-axis and Y-axis, respectively, represent the actual value and predicted value of IRI on the verification set. The closer the scatter is to the diagonal line, the better the prediction accuracy is. It is obvious that the model has a good fitting effect, which further indicates that the prediction effect of the MTSS model is feasible.

4.5. Application of Digital Twins

According to Shanghai Engineering Construction Code DGJ08-92-2013 “Technical Regulations for Urban Pavement Maintenance,” the main highway network connected at both ends of the Dalian highway tunnel belongs to the primary and secondary arterial highways. It needs to be classified according to the ratings in Table 6 to map the pavement performance values and the chromatogram.

Applying the digital twin for pavement performance prediction proposed in Section 3.1, taking each section of the Dalian highway tunnel as an example, as shown in Figure 17, the performance numerical results of the visualized prediction highway segments are used to enhance the visualization by the color mapped in the chromatogram. According to the pavement performance evaluation standard level, the pavement performance trend is predicted by the MTSS prediction model in the Dynamo visualization parametric analysis module (Figure 18), and the change law of the highway segment performance varies under different maintenance assumptions.

5. Conclusions

In this paper, a highway tunnel pavement performance prediction approach based on a digital twin + MTSS is proposed. This method adopts digital twin technology and machine learning technology to predict the pavement performance data through the established MTSS model. The multiple time series feature extraction method and the automatic hyperparameter tuning technology of the MTSS model are the key factors improving the accuracy of pavement performance prediction. The stacking integrated learning method can effectively reduce the risk of poor generalization performance of a single model and falling into a local minimum. The relative error between the predicted value and the actual value of the MTSS model is small, which improves the accuracy of the prediction. The visual information service of the digital twin provides a reliable data-driven management method for preventive maintenance, which helps to improve the management efficiency of prediction and decision-making.

Data Availability

The online access data used to support the findings of this study have not been made available because the data source involves the internal life cycle operation and maintenance management system and is not open to the public.

Conflicts of Interest

The authors declare that there are no conflicts of interest.

Acknowledgments

This research was supported by the National Natural Science Foundation of China (71701121), the Chinese Ministry of Education of Humanities and Social Science Project (17YJC630021), and the Key Program of the Shanghai Science and Technology Commission: research on intelligent identification and control strategy optimization of the shield tunneling state.