Abstract

For predicting the conversion velocity of the vinyl chloride monomer (VCM) in the polymerization process of polyvinylchloride (PVC), an improved Group Method of Data Handling- (GMDH-) type neural network soft-sensor model is proposed. After analyzing the technique of PVC manufacturing process, the auxiliary variables for setting up the soft-sensor model are selected and the experimental data are normalized. Because the internal standard of the original GMDH-type neural cannot solve the problem of multiple-collinearity problem and the useful variables tend to be prematurely eliminated in the modeling process, a hybrid method combining the regression analysis method and the least squares method is proposed to solve the multiple-collinearity problem. On the same time, by adopting some optimization experiences in genetic algorithm (GA), the generational crossover combination variables method is proposed to solve the shortcoming of useful variable being eliminated prematurely. The simulation results show that the proposed soft-sensor model can significantly improve the prediction accuracy of economic and technical indicators in the PVC polymerization process and can meet the real time control requirements of polymerization reactor production process.

1. Introduction

Polyvinyl chloride (PVC) is one of the most widely used polymers in the world, and it is also one of the first polymers to realize industrial production [1]. The production of polyvinyl chloride (PVC) is a kind of typical batch chemical production process by using vinyl chloride monomer (VCM) as raw material and using the suspension polymerization process to produce polyvinyl chloride (PVC) resin. The quality index of PVC is closely related to its processing process, but it is easily influenced by many factors, such as raw materials, additives, and technique parameters. Among them, the conversion rate and conversion velocity of VCM are the most important factors affecting the quality of PVC. The different VCM conversion has a certain impact on the molecular weight of PVC resin, thermal stability, porosity, the residues of VCM, the absorptivity of plasticizers, and processing liquidity [2]. The conversion rate of VCM is a key factor affecting the thermal aging property of the resin. The higher the conversion rate, the higher the yield of PVC resin. When the conversion rate reaches 80%, if the proportion of the unstable structure in the polymer PVC continues to increase, the thermal aging property of PVC resin will decreased. However, due to the limitation of the industrial field conditions and lack of mature measurement equipment, the conversion rate and conversion velocity of vinyl chloride in the actual production process are very difficult to obtain in real time, so it is also difficult to achieve the direct and efficient closed-loop control [3]. So it is very important to establish the soft-sensor model of the conversion rate and the conversion velocity of VCM.

Group method of data handling (GMDH) is a family of inductive algorithms for computer-based mathematical modeling of multiparametric datasets that features fully automatic structural and parametric optimization of models [4]. Inductive GMDH algorithms give possibility of finding automatically interrelations in data, to select an optimal structure of model or network, and to increase the accuracy of existing algorithms [5]. GMDH is a heuristic self-organization method and its essence is parameter estimation. Based on the dynamic analysis of the target process, the heuristic method is used to search the model structure, and then the model parameters are estimated according to the model results. The best GMDH model is indicated by the minimum of the external criterion characteristic. Multilayered procedure is equivalent to the artificial neural network (ANN) with polynomial activation function of neurons. Therefore, the algorithm with such an approach usually referred to as GMDH-type neural network (NN) or Polynomial Neural Network [6]. GMDH-type neural network is also a feed-forward network; the key is that it has three advantages over other forward neural networks [7]. () It can obtain the explicit function analytic expression of the model. That is to say it solves the problem that the model structure can be used to reveal the interaction and dependence among all variables, which cannot be achieved by the traditional neural network in the past. () The modeling process of the model is self-organized without any initial assumptions. Because the algorithm is based on the data driven to find the input items that have a substantial impact on the explanatory variables. () It has the optimal complexity and high precision prediction. It reduces the impact of small samples or bigger noise on the system and ensures the system’s generalization ability.

GMDH-type neural network has been applied in a great variety of areas for data mining and knowledge discovery, forecasting and systems modeling, optimization, and pattern recognition. GMDH neural network predictive model combining Harmony Search (HS) algorithm was proposed to predict the pullout capacity of suction caissons in clay [8]. Neurofuzzy method was combined with GMDH network so as to establish the NF-GMDH forecast model to predict the local scour depth around pile groups under clear-water conditions [9]. GMDH-type NN was used to realize the short-term prediction and the prediction accuracy was very stable [10]. The wavelet transformation (WT) algorithm was introduced into GMDH-type NN. The time series of significant wave height (SWH) were decomposed into some subseries using WT and then the decomposed time series were imported to the GMDH NN model to forecast the SWH in different time periods [11]. The particle swarm optimization (PSO) algorithm and Neurofuzzy theory were introduced into the GMDH network to realize the NF-GMDH-PSO algorithm, which was used to predict the longitudinal dispersion coefficient of the river. In this paper, an improved Group Method of Data Handling- (GMDH-) type neural network soft-sensor model is proposed for predicting the conversion velocity of the Vinyl Chloride Monomer (VCM) in the polymerization process of polyvinylchloride (PVC). The paper is organized as follows. In Section 2, the technique flowchart of the PVC polymerization process is introduced. The GMDH-type neural network is described in Section 3. In Section 4, the improved GMDH-type neural network soft-sensor model is introduced in detail. The simulation experiments and results analysis are discussed in Section 5. The conclusion illustrates the last part.

2. Polymerization Production Process of PVC

In the resin industrial production industry, the following four kinds of polymerization patterns are generally used: suspension polymerization, noumenon polymerization including gas phase polymerization, emulsion polymerization including microsuspension polymerization, and solution polymerization [1]. The suspension polymerization production technology is the main production mode of PVC resin because it is easy to adjust the product variety, the production process is simple, and it is easy to be controlled and realize the mass production. The typical PVC polymerization process is shown in Figure 1 [3].

PVC is polymerized by VCM. The general production process of the PVC resin based on suspension method is firstly to clean the polymerization reactor, which includes the cleaning before and after nurikabe; then the vinyl chloride monomer, water and suspending agent, and antioxidant are added in the polymerization kettle. These materials form a suspension in the polymerization reactor under strong agitation. The PVC monomer was polymerized into PVC particles at the elevated temperature and with the addition of initiator. When the polymerization proceeds to a certain extent, these particles will form PVC slurry. This is the polymerization process of polyvinyl chloride. In addition, the PVC production process also includes monomer recovery, PVC slurry stripping, PVC drying, and the packaging of the products. The production flow chart of polymerization kettle is shown in Figure 2.

PVC polymerization process is a typical batch process. In the polymerization process of PVC, all kinds of raw materials and auxiliary agents are put into the reactor. They are fully and evenly dispersed under the function of stirring. Then, the cooling water is ventilated to the clip set of the reaction kettle and baffle plate constantly in order to remove homopolymer. When the conversion rate of VCM reaches a certain value, the reaction terminates, that is to say the finished products are obtained. The degree of polymerization decreases with the increase of temperature, and the degree of polymerization is only related to the reaction temperature of VCM. Ultimately, the accuracy of the conversion velocity prediction model directly affects the quality of the product and the type of polyvinyl chloride. According to the characteristics of the polymerization process, 10 process variables related to the conventional rate and velocity of VCM are identified as the secondary variables of the soft-sensor model, which are listed in Table 1 [2].

3. GMDH Neural Network

The basic idea of GMDH is described as follows. The black box analysis method is used to establish the relationship between input and output. Then the function of the network model is expressed by the description of the relationships among the elements in the network. The establishment of GMDH-type neural network is a process of continuously producing active neurons. Then the external criteria are adopted to screen neurons; the quality of the retained neurons is generally superior to that of the discarded neurons (although some “better” neurons may be prematurely excluded). To combine the retained outstanding neurons is to screen out the better neurons as the neurons of the next layer until that the optimal model is selected.

3.1. Principle and Network Structure of GMDH-Type NN

The algorithm flowchart to generate GMDH-type NN is shown in Figure 3. The algorithm procedure of the standard GMDH-type NN is described as follows.

() Divide the training set and test set. The samples data setis divided into the training setand the testing set. The number of samples, whereis the total number of samples,is the number of training samples, andis the number of testing samples.

() Select the reference function to establish the relationship between the input variables and the output variables. In general, the discrete form of Volterra function or Kolmogorov-Gabor function is used as the reference functions

() Determine the external criteria.(a) Prediction error sum of squares (PESS)(b) An information criterion (AIC)(c) Average relative error (ARE)

In the above three criteria,is the estimated output value of the intermediate model on theth sample,is the actual output value of the intermediate model on theth sample, C is a constant, is a tunable parameter, andis the number of the observed samples.

() Generate variables of the initial layer. Each item in the selected reference function is used as the initial input variable of the algorithm. If the K-G polynomial has been selected, it is shown as follows, whereare the variables of the input data,

Thus, the input variables of the network can be obtained as follows:

() The intermediate model of the first layer is generated by the internal standard. The standard GMDH adopts the least square method. Self-organization strategy is used to generate an intermediate modelof the first layer. The parameters () of are estimated on the training set A according to the inner criterion.

() Select the intermediate model. On the testing data set, the intermediate model of the upper layer is selected by using the external criterion determined in Step (). Thenmodels with small external criterion values are selected as the input variables of the next layer. This method is carried out continuously to screen the variables until the optimal results are obtained.

() Generate the optimal model. Firstly, the termination rule is established. Normally, the model with the lowest external criterion value or the external criterion value which no longer decreases is selected as the optimal model. The complexity of the model produced in Step () is increasing, so the final optimal model can be determined according to the selected termination rule.

3.2. Advantages and Disadvantages of the Standard GMDH-Type NN

In this paper, the standard GMDH-type NN and the classic back propagation (BP) neural network are compared on the following factors: the structure of the network, the connection pattern of neurons, the self-organizing, the estimating method of parameters, the rules of using samples, the interpretability of the model, the convergence performance of the model, and the complexity of the program. The compared results are shown in Table 2.

It can be seen from Table 2 that the most obvious advantages of GMDH are that there is a clear expression, faster convergence rate and ease of controlling the estimated parameters, and so on. But seen from the network structure in GMDH modeling process, GMDH-type NN has certain shortcomings described as follows. () It can be seen from the GMDH structure chart, after screening the first layer variables, that the unselected variables will be discarded forever. Because these variables are only filtered once, there are some “useful variables” which will inevitably be prematurely eliminated, which will thereby reduce the quality of the model. () For the selection of initial variables, if too many initial variables are selected, the system will become too complex and difficult to understand. But if you choose too few variables, it will lead to some “useful variables” being eliminated in advance. So it is important to choose the initial variable. () The internal criterion that the standard GMDH-type NN generates the intermediate model is the least squares estimation method. Because of the limitation of the sample data, the traditional least squares method cannot estimate the coefficient of the regression coefficients when the multiple-collinearity relationship is generated between the regressions, which will reduce the accuracy and reliability of the model constructed by the traditional GMDH-type NN.

4. Improved GMDH-Type Neural Network

In view of the shortcomings of the traditional GMDH-type NN mentioned in Section 3.2, the paper will propose an improve GMDH-type NN to overcome these three shortcomings so as to improve its generalization ability. As the main point of this improvement is to improve the internal criterion, the stepwise regression analysis is introduced to eliminate the multiple-collinearity. Therefore, in order to facilitate the expression, the algorithm is called F-GMDH.

4.1. Structure of the Proposed Soft-Sensor Model

Ten variables described in Section 2 are set as the input variables and the conversion velocity and conversion rate of VCM are the output variables. The improved GMDH-type neural network is used to fit the nonlinear relationshipbetween input and output so as to establish the soft-sensor model of VCM conversion velocity and conversion rate, whose model structure is shown in Figure 4.

4.2. Division Method of Samples Set in Standard GMDH-Type NN
4.2.1. Cross Division Method

The standard GMHD-type NN generally divides the samples into the training data set and testing data set with the manual pattern, but this classification method relies on the personal experience to determine which part of the data is suitable for training data sets and which part of the data for the test data set. At the same time, the intermediate models generated by the different partitioning methods are very different. Such a division strategy is contrary to the idea of self-organization and self-evolution of the GMDH. In order to solve this problem, this paper puts forward the cross division method, which is described as follows.

() For a given set of samples, randomly generate a positive integer,  .

() The training set and test set are denoted asand.

() Inspired by the crossover operator in genetic algorithm (GA), at the same time, in order to allow the training set and testing set to be full crossed, in this paper, the data in the two data sets are crossed one by one. Generate a positive integer,  ; extract the datafrom the training set into the testing set. At the same time, generate a positive integer,  , extract the datafrom the testing set, and put it into the training set. and are obtained until all the original data in the two data sets are exchanged, whereis the number of exchanges. Then, the intermediate model is generated on the training setaccording to the inner criterion and the searching process for the optimization is carried out by using the testing set.

4.2.2. Description of OLS-Frisch Algorithm

When there are multiple-collinearity relationships among the regression coefficients, the traditional least squares method cannot estimate the coefficients of the regression coefficients. This paper introduces a stepwise regression method to eliminate the multiple-collinearity relationship among these variables. Firstly, the standard least squares method is used to establish the model by using one set of variables. Then variables are fed into the model one by one. Each introduction of a variable will results in carrying out a significant test for all variables in the model and extracting no significant variables. The recycling process is gradually carried out (variable introduction-variable elimination-variable introduction) until all the variables in the model are not significant. In order to facilitate the expression, the algorithm is called OLS-Frisch method.

Step 1. The least square method is used to establish the model, whose specific calculation process is described as follows [12, 13].
There arepairs of observational data,  , about the variableand. Supposewhere A represents the regression matrix,is the desired output,is the default mutual independent residual vector, andis the regression coefficient waiting to be solved and needing to be updated.
The training error can be expressed as follows:where,is a known function. It can be seen that E is a multivariate function related to. The training error E should be minimized in the training sample under 2-norm sense. So letto obtain. The solving process is described as follows.
Supposeto obtainSupposeis a n-matrix,,,().to obtainAccording to the above equations, the values ofare obtained, which are the required coefficients. Then the first regression equation is established.

Step 2. The initial correlation coefficient matrixis obtained from the above sample data set.

Step 3. The stepwise optimization of the variables is to establish the optimal regression equation.

() According to the modeling method described in Step to establish a regression model:,  , where. The test value F is selected, which is used to introduce and eliminate the variables. The principle of the introduction is to make the partial regression square sum largest, which is also called the largest variance contribution. Because the greater the partial regression square sum, the better the effect of the regression equation.

The selection method is a direct method. The direct method is to obtain the maximum partial regression square sum from. That is to say find out the relationship frombased on the inverse compact transformation method. The following results can be obtained from the above transformation relation:where,represents the diagonal elements in matrixandis the diagonal elements of the inverse matrix of. According to the above deduction, the partial regression square sum ofcan be expressed as follows:

It can be seen from the above deduction thatcan be obtained from. So there is the following relationship:

The introduction process of variables is described as follows.

() Determine the variables which will be introduced. The direct method is used to calculate all partial regression square sum, and then the largestis selected as the first introduced variable; that is to say,  .

() In this paper, the F testing method was used to test the variables. Firstly, the critical valueis determined. The size of the value is mainly related to the reliability and degree of freedom, so the selection value must be moderate. If the value is too large, it will cause the introduction of too few variables so as to make the established model not practical. In this paper, . If you choose six variables and each group of data has 361 samples, there is. Thus, whereis the corresponding parameter of the F distribution table. When the calculated F is greater than the critical valuein F distribution table under the corresponding significant level() and the degree of freedom, . That is to say when F is greater than, there are significant differences between the two groups of variances; when F is less than, there is no significant difference between the two groups of variances.

()can be obtained by solving the inverse compact transformation method on. The process of eliminating variables is described as follows. If it is the first introduction of a variable, you need not eliminate the variable; if the variable is introduced in the Nth sequence, then perform the step of eliminating the variable. The method of eliminating variables is also F testing method.

() Introduce new variables. Repeat the above Steps until the optimal model is established.

4.2.3. Intelligent Variable Selection Method (Generational Crossover Method)

As mentioned in Section 3.2, the variables screening method adopted by the standard GMDH is very rough, which is easy to make some “useful variables” prematurely eliminated and the reliability and accuracy of the final obtained model not ideal. In view of the above problems, this paper puts forward the intelligent variables selection method, which is mainly reflected in the following two aspects. () The reasonable choice of the initial variables because it can make the “quality” of the initial variables be guaranteed. () Initialize the reasonable rules for the reservation of the neurons in the middle layer so as to ensure that the “useful variables” will not be prematurely eliminated.

() Intelligent selection of the number of initial variables: in the division process of the sample set, there are a lot of group data available for training. So in the training of the first data set, the variable is selected from one to N. Then all the models are compared and a model with the best current performance is selected. So the number of initial variables corresponding to this modelis used as the initial variable number in the training of the remaining data set. In order to make the initial variable number more flexible, the number of initial variables is defined as,  andis less than the maximum number of variables. By adopting this method, a reasonable number of initial variables can be chosen and the method is easier to be programed.

() The retained principle of intermediate variables: the retention of intermediate variables has influence on the results of the multilayer iteration, which is described as follows. Suppose a group of independent variablesare the identified object. The output of a linear system is y. The input variablesconsist of the vector space. Then the projection of the output y onis, and, . So the expression of y on the vector spaceis described as follows:

If,, wheremeans “better than.”

According to the modeling characteristics of the GMDH algorithm and the principle of orthogonal projection, the necessary and sufficient conditions for the optimal approximation are described as follows.

() The necessary condition for the final calculation result with the optimal expression is that the number of intermediate variablesin the layer satisfies the following relationship:

() The sufficient condition for the final calculation result with the optimal expression is that the final result of the operation is that the number of intermediate variablesin the layer satisfies the following relationship:

The necessary condition is obtained under ideal conditions, and the sufficient condition is obtained under the worst condition. Therefore, if we want to get the optimal model, the above necessary and sufficient conditions need be satisfied and the selection method is designed to obtain the intermediate variables. In view of this problem, the idea of the method proposed in this paper is described as follows.

Because there are too many useful information in the initial variables, the initial variables are used to establish the model as much as possible so as to avoid the useful information prematurely eliminated, while the rest variables in each layer eliminated from the whole variables may have useful information. So, the main idea of this paper is to generate the first initial variable layer model; then the retained variables from the first layer and the initial variables are carried out in the crossover operation to generate the second layer model. Then the eliminated variables from the first layer and the retained variables from the second layer are carried out in the crossover operation to generate a third layer model. On this basis, if there are three consecutive times the variable is not selected, the variable will be completely removed, and so on, until the formation of the optimal model. Its structure is shown in Figure 5.

In the combination of variables, the crossover strategy is used to combine two sets of variables. The method to generate the intermediate model by using the random combination pattern is to replace the enumeration method of the standard GMDH algorithm to generate the intermediate model by using the variables pairwise combinatorial method. This can reduce the number of intermediate models in order to improve the efficiency of the model, reduce the complexity of the model, and effectively eliminate the redundant elements of the model.

4.2.4. Algorithm Flowchart of F-GMDH Method

Step 1. In order to make data have the same quantity rank in the training process for the GMDH-type NN, the normalized method is adopted to deal with the input and output data of neural network.

Step 2. Divide the sample set. The training set and testing set are divided by using the partitioning method mentioned in Section 4.2.1.

Step 3. Select the transfer function. The followed transfer function used in this paper is described as follows:

Step 4. Generate the intermediate model in the first layer. In this paper, the OLS-Frisch method is used as the internal standard to train the intermediate model in the first layer.

Step 5. Screen the intermediate model. The method mentioned in Section 4.2 is used to combine the initial variables, the eliminated variables, and the retained variables in each layer. Then the next layer model is generated based on the provided external criteria.

Step 6. Repeat Step and Step until the optimal model is obtained. The termination condition is that the external criterion value is not reduced or the iteration number reaches a certain given value.

5. Simulation Experiments and Results Analysis

In this paper, a chemical group company’s 40 thousand tons/year PVC production unit of the polymerization process is selected as the research object, where the VCM is the raw material and the production of PVC resin is based on the suspension polymerization technique. In the process of polymerization, the conversion rate of VCM is an important parameter. The conversion rate of VCM has a great influence on the quality of PVC resin products, such as the molecular weight, porosity, plasticizer absorption rate, VCM residues, and thermal stability. In conclusion, it directly affects the quality of PVC products and the economic and technological efficiency of enterprises. Therefore, it is very important to predict and control the conversion rate and conversion velocity.

In this paper, the simulation experiments data are divided into training set and test set. The definition of model performance index is listed in Table 3, which includes the root mean square error (RMSE), the square sum of error (SSE), the maximum positive error (MPE), and the maximum negative error (MNE). The prediction results of F-GMDH-type NN model NN model and the standard GMDH-type NN model are shown in Figures 610 and Table 4. The predictive results of VCM conversion rate under BPNN and GMDH-type NN are shown in Figure 6. The predictive error of VCM conversion rate is shown in Figure 7. It can be clearly seen that the prediction accuracy of the BP network model is not better than the prediction accuracy of the standard GMDH-type NN model. The predictive results of VCM conversion rate under GMDH-type NN and F-GMDH-type NN are shown in Figure 8 and the predictive errors of VCM conversion rate are shown in Figure 9. The performance comparison results of VCM conversion rate under different soft-sensor models are listed in Table 4. It can be seen from the simulation results that the F-GMDH-type NN model has a higher precision than the standard GMDH-type NN model.

Figures 10 and 11 are the comparison results of the prediction results of the conversion velocity of VCM under different soft-sensor models. The predictive results of VCM conversion velocity under BPNN, GMDH-type NN, and F-GMDH-type NN are shown in Figure 10. The predictive errors of VCM conversion velocity under BPNN, GMDH-type NN, and F-GMDH-type NN are shown in Figure 11. The performance comparison results of VCM conversion velocity under different soft-sensor models are listed in Table 5.

It can be seen from Figure 11 that the accuracy of F-GMDH-type NN model in the prediction of the conversion velocity of VCM is better than the standard GMDH-type NN model and BP NN model. The only shortcoming is that the training time of the proposed F-GMDH-type NN model is long.

6. Conclusions

In order to verify the advantages of the improved GMDH network (F-GMDH), the prediction performance experiments are carried out between the original GMDH neural network and the BP neural network, and the simulation results show that the original GMDH network is better than the BP neural network. Then the original GMDH neural network is compared with the F-GMDH neural network. The simulation results show that the prediction accuracy of F-GMDH neural network is better than the other two neural networks. In addition, the time required for the F-GMDH neural network model is much less than other models. So the simulation results show that the soft-sensor model based on the proposed F-GMDH-type neural network has high prediction accuracy.

Competing Interests

The authors declare no conflict of interests.

Authors’ Contributions

Most of Wei-zhen Sun’s contributions were in the data collection, analysis, algorithm simulation, and the draft writing. Jie-sheng Wang’s contributions were in the concept, design, interpretation, and comments on the manuscript. Shu-zhi Gao’s contribution was in the data collection and analysis of the manuscript.

Acknowledgments

This work is partially supported by the Project by National Natural Science Foundation of China (Grant no. 21576127), the Program for Liaoning Excellent Talents in University (Grant no. LR2014008), the Project by Liaoning Provincial Natural Science Foundation of China (Grant no. 2014020177), the Program for Research Special Foundation of University of Science and Technology of Liaoning (Grant no. 2015TD04), and the Opening Project of National Financial Security and System Equipment Engineering Research Center (Grant no. USTLKFGJ201502).