Research Article  Open Access
A New Least Squares Support Vector Machines Ensemble Model for Aero Engine Performance Parameter Chaotic Prediction
Abstract
Aiming at the nonlinearity, chaos, and smallsample of aero engine performance parameters data, a new ensemble model, named the least squares support vector machine (LSSVM) ensemble model with phase space reconstruction (PSR) and particle swarm optimization (PSO), is presented. First, to guarantee the diversity of individual members, different single kernel LSSVMs are selected as base predictors, and they also output the primary prediction results independently. Then, all the primary prediction results are integrated to produce the most appropriate prediction results by another particular LSSVM—a multiple kernel LSSVM, which reduces the dependence of modeling accuracy on kernel function and parameters. Phase space reconstruction theory is applied to extract the chaotic characteristic of input data source and reconstruct the data sample, and particle swarm optimization algorithm is used to obtain the best LSSVM individual members. A case study is employed to verify the effectiveness of presented model with real operation data of aero engine. The results show that prediction accuracy of the proposed model improves obviously compared with other three models.
1. Introduction
With increasing demands in the field of operation safety, asset availability, and economy, the health monitoring of aero engine has been widely considered as the key prerequisite for the competition of an airline company. One of the main tasks of health monitoring is to predict the performance parameter of aero engine. By predicting and analyzing the trend of performance parameters, one can obtain valuable information to avoid future risk and loss due to faults or accidents and reduce associated maintenance costs [1]. Therefore, it is necessary to design a high accurate and robust prediction model for aero engine performance parameter (AEPP).
A variety of traditional time series prediction approaches have already been proposed for this problem, such as fuzzy rule [2], Kalman filter [3], grey prediction [4], ARMA [5], and multiple regression [6]. These approaches are very mature in theory, but the accuracy is not always high and the robustness is not always satisfied in the application [7]. With the development of artificial intelligence techniques, recent studies for AEPP prediction are mainly focused on artificial neural network (ANN) [8, 9] and support vector machine (SVM) [10, 11].
Compared with standard SVM, least squares support vector machine (LSSVM) adopts equality constraints and a linear KarushKuhnTucker system, which has a more powerful computational ability in solving the nonlinear and smallsample problem [12, 13]. In addition, LSSVM eliminates local minima and structure design complexity of ANN. Therefore, LSSVM is a good choice for AEPP prediction model designing. However, the modeling accuracy of a single LSSVM is not only influenced by the input data source, but also affected by its kernel function and regularization parameters [12]. Thus, several main disadvantages are worth to be addressed. Firstly, using a datadriven technique to design an LSSVM model, data source should be considered as the first factor. AEPP data is different from the pure random system: that is, the chaotic characteristic of AEPP data should be extracted to reconstruct input data samples before modeling. Secondly, as two common parameter optimized methods for LSSVM, conventional crossvalidation and grid search methods have several defects, such as high time consuming and a priori knowledge requirement.
In addition, although a single LSSVM with optimal parameters and reconstructed input data samples may have an excellent prediction performance under certain circumstances, because its kernel function is fixed, it perhaps has some kinds of inherent bias under other cases. In the literatures, due to the super robustness and generalization, ensemble model has been proved to be an effective way to reduce biases of single model. Ensemble model can make full use of diversity to compensate for disadvantages among the individual members, and the reasonable combination strategy is believed to be able to produce better prediction accuracy and generalization than single model [13–16]. By using combining submodels, the multilayer networks of LSSVMs ensemble have been discussed deeply, which is very encouraging and promising for further research [13], but up to now, the application of LSSVM ensemble model for AEPP prediction is relatively fresh and untouched in the open literature.
For ensemble model design, there are two points that should be considered. One is that the selected individual members need to exhibit much diversity (disagreement) and accuracy. The other is the effectiveness of the combination strategy [17]. For the diversity of individual members, it is an easy and common way to build individual members by using data decomposition [18]. However, this method is proved to be effective when the original data sample is sufficient, and it is not suitable for smallsample data. Compared with the existing combination strategies such as simple averaging weighting method, mean squared error weighting method, and least squares estimation weighting averaging method, intelligent method based combination strategies include ANN combiner and SVM combiner which have become the current trend [18]; however, ANN combiner cannot avoid falling into local optima, and for SVM combiner it is not easy to select appropriate kernel function, so it is necessary to further improve these ensemble strategies.
As previously mentioned, in this paper, a new PSRPSO(SK)LSSVM(MK)LSSVM ensemble (PPLLE) model for AEPP prediction is proposed. Firstly, a set of diverse single kernel LSSVMs are created as base predictors. Subsequently these individual member LSSVMs output the primary prediction results independently. Finally, all the primary prediction results are combined to produce the most appropriate prediction results by another particular multiple kernel LSSVM. In the process of modeling, phase space reconstruction (PSR) theory is applied to extract the chaotic characteristic of input data source and reconstruct the data samples. Particle swarm optimization algorithm is used to search the best parameters for LSSVM members to ensure their prediction accuracies.
The rest of this paper is organized as follows. The next section provides a brief introduction to the related knowledge. Section 3 formulates the proposed PPLLE model. For illustration purpose, the detailed application on AEPP prediction and model comparisons is proposed in Section 4. Section 5 concludes this study.
2. An Overview of the Related Knowledge
2.1. Data Samples Reconstruction Based on PSR Theory
Although the nonlinear chaos behavior is the main challenge confronting the chaotic data series prediction, the underlying data generating mechanisms can still be explored by PSR theory [19]. By means of the ability of revealing the nature of dynamic system state, the PSR theory is useful in system characterization, in nonlinear prediction, and in estimating bounds on the size of the system [20, 21].
According to Takens’ theorem, for the nonlinear time series , the current state information can be represented by an dimensional vector:where is the delay time, is the embedded dimension, they are two important parameters for phase space construction, and is the mapping relation between the inputs and outputs.
The autocorrelation function of time series at the first minimum value is taken as the delay time of the reconstructed phase space; we writewhere is the mean of .
To calculate the correlation dimension, the correlation integral needs to be computed:where is the selected radius and , is the Heaviside function.
The correlation dimension is calculated by the formula as below:
Suppose , ; then we can reconstruct data samples .
2.2. Least Squares Support Vector Machine
LSSVM is the least squares form of a standard SVM; it was firstly proposed by Suykens and Vandewalle [12]. LSSVM uses a set of linear equations during the training process and chooses all training data as support vectors, so it has excellent generalization and low computation complexity [12–14].
In LSSVM, the regression issue can be expressed as the following optimization problem:where is a nonlinear function which maps the input data into a higher dimensional space. is the error at time , is the bias, and is the regulation constant.
According to the Lagrange function and KarushKuhnTucker theorem, the LSSVM for nonlinear functions can be given as below:where is the Lagrange multiplier and is the kernel function which is applied to substitute the mapping process and avoid computing the function .
Typical kernel functions include linear kernel function, polynomial kernel function, radial basis kernel function, sigmoid kernel function, and multiple kernel function. Some of them are listed as follows:(1)linear kernel function (LKF):(2)polynomial kernel function (PKF):(3)Gaussian radial basis kernel function (RBF):(4)sigmoid kernel function (SKF):(5)multiple kernel function (MKF):
The nonlinear mapping ability of LSSVM is mainly determined by its kernel function form and relevant parameters setting: that is, various kernel functions or parameters have different influence on the prediction ability of LSSVM predictor (the parameters setting will be discussed in the next section). As to kernel function, the LKF is suited to expressing the linear component of the mapping relation, and the RBF possesses a wider convergence domain and an outstanding learning ability and high resolution power, while the PKF has a powerful approximation and generalization ability. Meanwhile, kernel functions can also be divided into local kernel function and global kernel function. For the global kernel function, it has the overall situation characteristic and is commonly good at fitting the sample points which are far away from the testing points, but the fitting effect is not perfect on the sample points which are near the testing points, and vice versa to the local kernel function [15]. Each kind of kernel function has its own advantages and disadvantages; the prediction performances of LSSVM with different kernel functions are not identical.
Here, we define the LSSVM configured with a multiple kernel function as the multiple kernel LSSVM (MKLSSVM); otherwise, we call it the single kernel LSSVM (SKLSSVM).
2.3. Parameters Optimized Based on PSO
Particle swarm optimization (PSO) algorithm is a popular swarm intelligence evolutionary algorithm used for solving global optimization problem [22]. It can search the global optimal solution in different regions of the solution space in parallel.
In PSO, the position of each particle represents a solution to the optimization problem. is the position vector and is the velocity vector of the th particle. Similarly, represents the best position of the th particle which has been achieved, and represents the best position among the whole particle group.
The values of position and velocity of the particle are updated as follows:where and are the acceleration constant, and are two random numbers in the range , and is inertia weight factor. To improve the convergence speed of PSO, , , and of PSO are adjusted by using the formulas as below:where expresses the maximum iteration number and is the current iteration number.
3. Overall Process of Designing the PPLLE Model
The core idea of the ensemble model lies in that all the individual members are accurate as much as possible and diverse enough, and it adopts an appropriate ensemble strategy to combine these outputs of the selected members [13–18].
3.1. Selection of the Appropriate Individual Member Predictors
For LSSVM prediction model, several diverse strategies, such as data diversity, parameter diversity, and kernel diversity, have been proved effectively for the creation of ensemble members with much dissimilarity [13]. Because kernel function has a crucial and direct effect on the learning and generalizing performance of LSSVM, various kernel functions can be used to create diverse LSSVMs. In this study, independent SKLSSVMs, such as LKFLSSVM, PKFLSSVM, and RBFLSSVM, are selected as individual member LSSVM predictors.
3.2. Combination of the Selected Individual Member LSSVM Predictors
After the diverse individual member LSSVM predictors have been selected, the other key question is how to determine the weight coefficient of each individual predictor, that is, how to construct the combiner effectively. As depicted in previous section, the MKF integrates the advantages of global kernel function and local kernel function and offsets some shortages of both simultaneously. Hence, another special MKLSSVM is chosen as the combiner. In this paper, the MKF is composed of a RBF and a PKF: the former is a typical local kernel function and the latter is a representative globe kernel function. A similar MKLSSVM model has high prediction accuracy and generalization ability, which has been proved with chaotic time series by Tian et al. [15].
3.3. Overall Process of Designing the PPLLE Model
The basic framework of the proposed PPLLE model is given in Figure 1, where is the number of the individual member LSSVM predictors.
As shown in Figure 1, there are three main stages in the basic framework which can be summarized as follows.
Stage 1 (sample dataset reconstruction and partition). The data source is reconstructed as data samples by using PSR; then the reconstructed data samples are divided into two indispensable subsets: training subset and testing subset.
Stage 2 (individual member creation and prediction). Based on kernel function diversity principle, independent SKLSSVMs are created as the individual member. Each SKLSSVM is trained by using the training subset. Accordingly, the computational results of the SKLSSVM predictors can be obtained, respectively. In the process of SKLSSVM creating, PSO is used to optimize parameters of each member SKLSSVM.
Stage 3 (combiner creation and prediction). When the computational results of the individual member predictors in the second stage are acquired, they are aggregated into an ensemble result by another special MKLSSVM. Similarly, to create the optimal MKLSSVM, PSO is applied again.
Here, is the mapping function determined by the special combiner MKLSSVM; thus, the final prediction output of the PPLLE model can be given as below:
4. Case Study
Due to different gas path component degradations such as fouling, erosion, corrosion, and foreign object damage, the performance of an aero engine will decline over the service time [23]. A lot of gas path performance parameters are often used in health monitoring of aero engine from different angles and levels, such as exhaust gas temperature (EGT), fuel flow (FF), and low pressure fan speed (N1). Among these performance parameters, EGT is considered as one of the most crucial working performance parameters of aero engine, which is measured to represent outlet temperature of combustor chamber in practice. When other conditions remain the same, the higher the EGT is, the more serious the performance degradation of aero engine is [4]. EGT gradually rises when the working life of aero engine increases, if the EGT value reaches or exceeds the scheduled threshold provided by the original equipment manufacturer, then the aero engine needs to be arranged for maintenance timely.
In this study, we select EGT as the AEPP representative to predict by using the proposed PPLLE model, and it is worth mentioning that other similar parameters can also be predicted in the same way.
4.1. Data Description and Samples Reconstruction
In this study, the EGT data come from the real flight recorders of the cruise state of a certain type of aero engine, and the sampling interval is 5 flight cycles. The data series consists of 148 EGT datasets, covering the period from February 2013 to September 2014. To increase the quality of the prediction results, some abnormal samples have been discarded from the original data series. The observed EGT data is shown in Figure 2.
For the observed EGT data series , according to (2), (3), and (4), the delay time is set as 1 and embedding dimension is obtained by computing. Thus, is taken as the input vector , and () is used as the corresponding expected value, so we can get the reconstructed data samples . The data samples are used as training subset to train each individual LSSVM of the ensemble model, and the samples are chosen as testing subset to validate the ensemble model. The onestep ahead prediction used in this paper is explained as in Figure 3. After the ensemble model has been trained, vector is entered into 4 individual predictors (SKLSSVM predictors) to compute their predicted values , respectively. Then, these predicted values are aggregated into an ensemble result by using a combination predictor (MKLSSVM predictor). Hence, the final predicted value is obtained. In this way, from to 148, all the final predicted values to can be got in turn.
4.2. Evaluation Indices
Mean absolute percentage error (MAPE), mean absolute error (MAE), mean squared error (MSE), and Theil’s Inequality Coefficient (TIC) are used to evaluate the prediction ability of the prediction model:where and are the observed values and corresponding prediction values, respectively.
4.3. Model Parameters Setting
In the modeling process of LSSVMs, the parameters of PSO are set as follows: , , , , , and . By using the PSO, the corresponding optimal parameters of LSSVM_{2}LSSVM_{5} are obtained and listed in Table 1. An appropriate individual member number of the ensemble model is able to achieve a balance between the prediction efficiency and the prediction ability [17]. In this study, the member number is set as 5.

4.4. Results and Discussion
Figure 4 illustrates the prediction results for the EGT testing dataset by PPLLE model and corresponding observed EGT value. The black symbol represents the observed value, and the red symbol expresses the prediction value. From Figure 4, we can find that the rise and fall trends of the two curves are approximately the same, and only the individual points have some higher gaps of the size, which means EGT is predicted with good accuracy on the testing data samples as a whole. There are two causes that may explain the gaps between the observed values and prediction values. Firstly, it is difficult to give a thorough consideration to extract the EGT characteristics when determining the model input data. Secondly, due to the influence of subjective factors, it is impossible to eliminate all the outliers properly.
In contrast, the single LSSVM model proposed by Tian et al. [15], RBFchaos model proposed by Zhang et al. [24], and PPLL (PSRPSOLSSVMLSSV ensemble) model are built. The kernel function and parameters of the single LSSVM model are the same as those of the LSSVM_{5} model listed in Table 1. The RBFchaos model aggregated chaos characteristics and RBF neural networks (here, the input layer, hidden layer, and output layer of RBF neural network are set as 5, 11, and 1, resp.). The difference between the PPLLE model and PPLL model lies in that the latter uses an RBFLSSVM (i.e., SKLSSVM) as the combiner.
In Table 2, the MAPE, MAE, MSE, and TIC values of the PPLLE, PPLL, RBFchaos, and single LSSVM models on the testing dataset are listed. It shows that the PPLLE model performs the best among the four modes with MAPE of 0.51, compared with those of 0.62, 0.85, and 1.10 by the PPLL, RBFchaos, and single LSSVM models, respectively. The MAE of PPLLE, PPLL, RBFchaos, and single LSSVM models are 3.67, 4.48, 6.16, and 7.99, respectively, which demonstrates the prediction accuracy of the proposed model. PPLLE model predicts the EGT with MSE of 14.04, better than PPLL, RBFchaos, and single LSSVM models with those of 22.48, 49.70, and 75.66, respectively. Besides, it should be pointed out that the TIC of PPLLE is 0.00258, which is quite acceptable compared with those of the other 3 models. A strong support is also exhibited by Figure 5, where the curve of PPLLE model intuitively shows the good prediction accuracy and excellent ability in tracking the observed EGT compared to the other 3 models.

Figures 6(a)–6(d) show a detailed profile of relative percentage error (RPE) between the observed values and prediction values of different models on the EGT testing data samples. It illustrates that the PPLLE model has an outstanding approximation ability with the RPE ranging from −0.7% to 0.9%; the RPE ranging around [−1.4%, 2.9%] in Figure 6(d) shows that the single LSSVM has the worst performance. RPE distribution range of PPLL model is better than that of RBFchaos model, which are exhibited by Figures 6(c) and 6(d). Comparison results of Figure 6 also prove the effectiveness of our proposed approach. Some of the main reasons why the PPLLE model is superior to others can be summarized as follows: the PPLLE ensemble model based on kernel diverse principle eliminates the possible inherent biases of single LSSVM and makes full use of the advantages of individual member LSSVMs; the PSR extracts the chaotic feature of the original data source and reconstructs data samples, which elucidates the input characteristic for the PPLLE model; the PSO ensure that each individual LSSVM achieves the best performance; the particular ensemble strategy of PPLLE employs an MKLSSVM and further enhances the prediction ability of the ensemble model.
(a) RPE of PPLLE
(b) RPE of PPLL
(c) RPE of RBFchaos
(d) RPE of single LSSVM
5. Conclusions
Designing a high accuracy and robust model for AEPP prediction is quite challenging, since AEPP data is nonlinear, chaotic, and smallsample, and the traditional single prediction model may have some inherent biases. To solve this problem and to realize high prediction accuracy level, a new LSSVM ensemble model based on PSR and PSO is presented and applied to AEPP prediction in this paper.
For the presented PPLLE prediction model, individual member LSSVMs based on kernel diverse principle eliminate the inherent biases of single LSSVM and make full use of the advantages of them as much as possible. PSR is applied to reconstruct data samples, which alleviates the influence of the chaotic feature of the original data source to the PPLLE model. PSO is used to guarantee that each individual LSSVM achieves the best performance. The particular ensemble strategy employs an MKLSSVM combiner, as the MKF integrates the advantages of global kernel function and local kernel function, and it offsets some shortages of both; this ensemble strategy further enhances the prediction ability of the ensemble model.
EGT is selected as the representative health monitoring parameter of aero engine for validating the effectiveness of the proposed PPLLE model. For comparison, the PPLL, RBFchaos, and single LSSVM models are also developed and evaluated. The PPLLE predicts EGT with MAPE of 0.51%, better than the PPLL, RBFchaos, and single LSSVM models with those of 0.62%, 0.85%, and 1.10%, respectively. Similarly, the PPLLE predicts EGT with TIC of 0.00258, better than the PPLL, RBFchaos, and single LSSVM models with those of 0.00327, 0.00485, and 0.00598, respectively. In addition, MAE and MSE indices also confirm that the presented model gives improved prediction accuracy. In a word, the above four evaluation indices consistently demonstrate that the PPLLE model is more suitable for AEPP prediction problem, and the PPLLE model can meet the actual demand of engineering application. Moreover, comparing results imply that this ensemble model has a promising application in other similar engineering areas where the data have complex nonlinear chaos relationships.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
References
 L.F. Yang and I. Ioachim, “Adaptive estimation of aircraft flight parameters for engine health monitoring system,” Journal of Aircraft, vol. 39, no. 3, pp. 404–408, 2002. View at: Publisher Site  Google Scholar
 J. Hong, L. Han, X. Miao et al., “Fuzzy logic inference for predicting aeroengine bearing gradlife,” in Proceedings of the 9th International Flins Conference on Computational Intelligence: Foundations and Applications, vol. 4, pp. 367–373, Chengdu, China, August 2010. View at: Google Scholar
 G. You and N. Wang, “Aeroengine condition monitoring based on Kalman filter theory,” Advanced Materials Research, vol. 490–495, no. 4, pp. 176–181, 2012. View at: Google Scholar
 N.B. Zhao, J.L. Yang, S.Y. Li, and Y.W. Sun, “A GM (1, 1) Markov chainbased aeroengine performance degradation forecast approach using exhaust gas temperature,” Mathematical Problems in Engineering, vol. 2014, Article ID 832851, 11 pages, 2014. View at: Publisher Site  Google Scholar
 Z. G. Liu, Z. J. Cai, and X. M. Tan, “Forecasting research of aeroengine rotate speed signal based on ARMA model,” Procedia Engineering, vol. 15, pp. 115–121, 2011. View at: Publisher Site  Google Scholar
 Y.X. Song, K.X. Zhang, and Y.S. Shi, “Research on aeroengine performance parameters forecast based on multiple linear regression forecasting method,” Journal of Aerospace Power, vol. 24, no. 2, pp. 427–431, 2009 (Chinese). View at: Google Scholar
 C. Chatfield, The Analysis of Time Series: An Introduction, Chapman & Hall/CRC, Boca Raton, Fla, USA, 6th edition, 2003. View at: MathSciNet
 S. G. Luan, S. S. Zhong, and Y. Li, “Hybrid recurrent process neural network for aero engine condition monitoring,” Neural Network World, vol. 18, no. 2, pp. 133–145, 2008. View at: Google Scholar
 D. Gang and S. S. Zhong, “Aircraft engine lubricating oil monitoring by process neural network,” Neural Network World, vol. 16, no. 1, pp. 15–24, 2006. View at: Google Scholar
 C. Zhang and N. Wang, “Aeroengine condition monitoring based on support vector machine,” Physics Procedia, vol. 24, pp. 1546–1552, 2012. View at: Publisher Site  Google Scholar
 X.Y. Fu and S.S. Zhong, “Aeroengine turbine exhaust gas temperature prediction using process support vector machines,” in Advances in Neural Networks—SNN 2013, vol. 7951 of Lecture Notes in Computer Science, pp. 300–310, Springer, Berlin, Germany, 2013. View at: Publisher Site  Google Scholar
 J. A. K. Suykens and J. Vandewalle, “Least squares support vector machine classifiers,” Neural Processing Letters, vol. 9, no. 3, pp. 293–300, 1999. View at: Publisher Site  Google Scholar
 J. A. K. Suykens, T. Van Gestel, J. De Brabanter, B. De Moor, and J. Vandewalle, Least Squares Support Vector Machines, World Scientific, Singapore, 2002.
 X. Yan and N. A. Chowdhury, “Midterm electricity market clearing price forecasting: a hybrid LSSVM and ARMAX approach,” International Journal of Electrical Power & Energy Systems, vol. 53, no. 1, pp. 20–26, 2013. View at: Publisher Site  Google Scholar
 Z.D. Tian, X.W. Gao, and T. Shi, “Combination kernel function least squares support vector machine for chaotic time series prediction,” Acta Physica Sinica, vol. 63, no. 16, Article ID 160508, 2014 (Chinese). View at: Publisher Site  Google Scholar
 O. Cagcag Yolcu, “A hybrid fuzzy time series approach based on fuzzy clustering and artificial neural network with single multiplicative neuron model,” Mathematical Problems in Engineering, vol. 2013, Article ID 560472, 9 pages, 2013. View at: Publisher Site  Google Scholar
 L. Yu, W. Y. Yue, S. Y. Wang, and K. K. Lai, “Support vector machine based multiagent ensemble learning for credit risk evaluation,” Expert Systems with Applications, vol. 37, no. 2, pp. 1351–1360, 2010. View at: Publisher Site  Google Scholar
 Y. Lv, J. Liu, T. Yang, and D. Zeng, “A novel least squares support vector machine ensemble model for ${\mathrm{N}\mathrm{O}}_{x}$ emission prediction of a coalfired boiler,” Energy, vol. 55, pp. 319–329, 2013. View at: Publisher Site  Google Scholar
 Q. Zhang, P. W.T. Tse, X. Wan, and G. Xu, “Remaining useful life estimation for mechanical systems based on similarity of phase space trajectory,” Expert Systems with Applications, vol. 42, no. 5, pp. 2353–2360, 2015. View at: Publisher Site  Google Scholar
 B. F. Feeny and G. Lin, “Fractional derivatives applied to phasespace reconstructions,” Nonlinear Dynamics, vol. 38, no. 1–4, pp. 85–99, 2004. View at: Publisher Site  Google Scholar  MathSciNet
 L. Su and C. Li, “Local prediction of chaotic time series based on polynomial coefficient autoregressive model,” Mathematical Problems in Engineering, vol. 2015, Article ID 901807, 14 pages, 2015. View at: Publisher Site  Google Scholar
 Y. B. Yuan and X. H. Yuan, “An improved PSO approach to shortterm economic dispatch of cascaded hydropower plants,” Kybernetes, vol. 39, no. 8, pp. 1359–1365, 2010. View at: Publisher Site  Google Scholar
 L. Wang, Y. G. Li, M. F. Abdul Ghafir, and A. Swingler, “A Rough Setbased gas turbine fault classification approach using enhanced fault signatures,” Proceedings of the Institution of Mechanical Engineers, vol. 225, no. 8, pp. 1052–1065, 2011. View at: Publisher Site  Google Scholar
 Z. Y. Zhang, T. Wang, and X. G. Liu, “Melt index prediction by aggregated RBF neural networks trained with chaotic theory,” Neurocomputing, vol. 131, pp. 368–376, 2014. View at: Publisher Site  Google Scholar
Copyright
Copyright © 2016 Dangdang Du et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.