Abstract

Classification model of support vector machine (SVM) overcomes the problem of a big number of samples. But the kernel parameter and the punishment factor have great influence on the quality of SVM model. Particle swarm optimization (PSO) is an evolutionary search algorithm based on the swarm intelligence, which is suitable for parameter optimization. Accordingly, a nonlinear inertia convergence classification model (NICCM) is proposed after the nonlinear inertia convergence (NICPSO) is developed in this paper. The velocity of NICPSO is firstly defined as the weighted velocity of the inertia PSO, and the inertia factor is selected to be a nonlinear function. NICPSO is used to optimize the kernel parameter and a punishment factor of SVM. Then, NICCM classifier is trained by using the optical punishment factor and the optical kernel parameter that comes from the optimal particle. Finally, NICCM is applied to the classification of the normal state and fault states of online power cable. It is experimentally proved that the iteration number for the proposed NICPSO to reach the optimal position decreases from 15 to 5 compared with PSO; the training duration is decreased by 0.0052 s and the recognition precision is increased by 4.12% compared with SVM.

1. Introduction

Power cables play an extremely important role in industrial production and modern life. At present, it is difficult for people to accept a bank system chaos or a wrong airport management system because of the power cable faults. In order to decrease and avoid the economic loss, the correct state classification of the online power cable is very necessary.

Nowadays, the commonly used fault diagnosis methods of power cable are electrical bridge method and electrical impulse method. Both of these methods are offline methods [1]. Obviously, these offline methods cannot satisfy the requirement.

In theory, the entropy of the zero-sequence components [2] of 3-phase voltages and 3-phase currents of the online power cable was used to extract the fault feature. In addition, wavelet transform is useful for the feature extraction of the early fault of the online cable, and then the voltages and currents are detected to get the subtle singular points by utilizing wavelet transform [3]. Furthermore, the artificial neural network [4] was also used to build the state classification model of the power cable because it can realize any nonlinear mapping. But artificial neural network needs a large number of samples and the training process may go to the local minimum point. The classification model of support vector machine (SVM) overcomes the disadvantage of big samples by obeying the rule of the minimum structural risk [5]. But the kernel parameter and the punishment factor have great influence on the quality of SVM model.

Particle swarm optimization (PSO) is an evolutionary search algorithm [6] based on the swarm intelligence, which is suitable for parameter optimization. Therefore, the combination of PSO and SVM can find the optimal kernel parameter and the punishment factor of SVM and obtain a high quality of SVM classification model.

This paper is organized as follows. After the introduction in Section 1, SVM and PSO related to this study are given in Section 2. It includes the conventional SVM and the traditional PSO as well as the specific PSO with a convergence factor and an inertia factor. Then, the nonlinear inertia convergence PSO (NICPSO) and the nonlinear inertia convergence classification model (NICCM) are proposed in Section 3. The experiments are implemented using NICPSO and NICCM to classify the normal state and several fault states of online power cable in Section 4. Finally, the conclusion is obtained in Section 5.

2. SVM and PSO

The basic idea of NICCM is that SVM is taken as the classification model and PSO is taken to optimize the important parameters of the punishment factor and the kernel parameter of SVM. The principles of SVM and PSO are reviewed below.

2.1. SVM

The principle of SVM is to find an optimal classification hyperplane which separates as much as possible patterns of two classes to the correct classes. Meanwhile, the hyperplane ensures the maximum distance between the two classes of separable samples [7]. If is the minimum distance from the classification hyperplane to the positive sample set, and is the minimum distance from the classification hyperplane to the negative sample set, then the margin of the classification hyperplane is “”. The linear SVM is to find the separation hyperplane with maximum margin. Namely, all the training samples should satisfy the following constraints of : or the equivalent constraint where is the th training sample and is the class label of the training sample ; and are the parameters of the classification hyperplane; represents the transposition of a vector; is the number of samples.

If a training sample satisfies (2), then it is a support vector. The change of a support vector impacts the change of the margin and the solution of problem.

Linear support vector machine is a maximization problem in the view of (2). Equivalently, it is a minimum problem of in the condition of (2); namely, it is an optimization problem with the constraint where is the th training sample and is the class label of the training sample and and are the parameters of the classification hyperplane. This is a typical convex quadratic programming problem.

For the linear separable classification problem of two-class, the minimum structural risk problem can be described by the conditional quadratic optimization problem. SVM perfectly solves the two-class classification problem by finding the solution of a convex quadratic optimization problem [8].

For the indivisible linear problem of two-class, SVM reaches the optimal result by constructing a classification hyperplane with a soft margin. According to Mercer kernel expansion theorem, a sample space can be mapped to a higher dimensional feature space in which the linear learning machine can be used to solve the nonlinear classification problem. The solution of inseparable linear two-class problem makes SVM become one of the formal two-class classifiers with two values [9].

Given a sample set , . is an input vector; denotes the corresponding desired output vector; is the sample number. The Lagrange multiplier is introduced to construct the optimal classification function of separable linear samples; then where is the sign function. The class of pattern is determined by the sign of the brackets.

For the case of imperfect separable linear samples, the loss introduced by classification error should be considered. The relax factor ,  , is used to (2). There is

Then the classification hyperplane with a soft margin is determined by the optimization problem where and are, respectively, the weights and constant of classification hyperplane; is the punishment factor of wrong classification which adjusts the balance between the confidence range and the experienced error. The bigger implies the smaller experienced error, and the smaller means the bigger classification margin. In addition, is a positive constant, and the bigger means the more serious punishment.

To solve (6), Lagrange multiplier and Lagrange function are used to convert the problem (6) to a quadratic programming problem

Most of the practical problems are nonlinear. Then kernel function is introduced to convert an inseparable linear problem in a lower dimensional space to a separable linear problem in a higher dimensional space. In this case, the quadratic programming problem is

So the nonlinear SVM classifier is where is the sign function and is a kernel function. The commonly used kernel functions are linear function, polynomial function, Sigmoid function, and Gauss redial basis kernel function which is used in this paper. Gauss redial basis kernel function is given as follows: where and are, respectively, the th and th sample; is the width of redial basis function.

The punishment factor controls the punishment degree by which the sample with over error is penalized [10]. Besides, the generalization ability of SVM is determined by the kernel parameter [11]. That is to say, these two parameters have great influence on the quality of SVM classification model.

2.2. PSO

The basic idea of PSO is to take every particle with a random initialization as a possible solution of optimization problem, and the quality of the particle is determined by a predefined fitness function. Each particle moves in the possible solution space, and the direction and the distance are determined by a velocity variable. Particles generally follow the current optimal particle, and the optimal solution will be obtained finally through each generation of searching [12]. If is the most optimal position which has been experienced by the individual particle, and is the most optimal position which has been experienced by the whole particle swarm, then particle [13] updates velocity and position according to the following equations during the iterations: where is the iteration times; and are random numbers in range of and they will keep the diversity of the swarm; and are learning factors, and they make the particle have the ability to summarize itself and learn from the excellent individual in the whole particle swarm, and finally the particle closes to the optimal positioning in its history and in the whole particle swarm history.

To improve the movement velocity of particle and enhance the local searching ability, PSO with convergence factor is commonly used. The velocity and position are as follows [14]: where is the number of iteration.

The convergence factor and its parameter can control the velocity. The optimal position, , is experienced so far by the th individual particle. The optimal position, , is experienced so far by the whole particle swarm [12, 13].

To keep the balance between the global searching and the local searching, and decrease the iteration number for optimal solution, the inertia PSO (IPSO) is usually used. The velocity and position of IPSO can be described as follows [15]: where is inertia factor which plays a role of a tradeoff between global optimization and local optimization.

The probability of the global optimization can be increased with a fewer iteration number by applying the IPSO.

3. NICCM Classifier

As we know from the above section, the punishment factor and the kernel parameter should be optimized to ensure the quality of SVM classification model. That is to say, the parameter selection processes of SVM are the optimal searching processes, and each point in the searching space is a potential solution of the optimal model [16, 17].

3.1. Definition of NICPSO

In order to make use of the advantages of the PSO with inertia factor and the PSO with convergence factor, these two PSOs are fused to construct the NICPSO. The NICPSO velocity of a particle is defined as the weighted velocity of the inertia PSO: where is the convergence factor, is the nonlinear inertia factor, and is the control factor.

Besides, the nonlinear inertia factor is selected to be where is the number of iterations and is the control factor which control the smoothness degree of the w-t curve. The inertia factor varies with the iteration time in Figure 1. These - curves move from the right upper to the left lower when .

3.2. NICCM Classifier

Based on the above discussion, the NICPSO can be written as where is the nonlinear inertia factor of NICPSO, is the convergence factor, is the speed of particle, and is the position of particle.

To optimize the punishment factor and the kernel parameter , the proposed NICPSO and the NICCM are used. As we know, these two parameters have great influence on the quality of SVM. The NICCM is given as where and are, respectively, the th pattern and its class label, is the th Lagrange multiplier, is the constant of the classifier, and is the number of samples.

The flow chart of NICCM is given in Figure 2.

The NICCM can be described as follows.

Step 1. Initialize the parameters of PSO.

Step 2. Evaluate every particle.

Step 3. Update the velocity and the position of each particle.

Step 4. Keep the particles as the optimal particles if the searching result is satisfied. Otherwise, go back to Step 2.

Step 5. Train the NICCM classifier using the optical punishment factor and the optical kernel parameter that comes from the optimal particle.

Step 6. Keep the trained NICCM as the classification model and end the training process if the performance is satisfied. Otherwise go back to Step 5.

Some parameters are set in Step 1. They are the size of PSO, the accelerators , the maximum velocity , and the maximum position . The particle is constructed by a punishment factor and a kernel parameter. Namely, the particle is a vector ().

For the particle evaluation, the fitness value of each particle is calculated according to fitness function. The fitness value of each particle is compared with the best one of all its experienced positions. If the comparison result is better, then this position is taken as the best local one instead of the old best local one. Furthermore, if a fitness value is better than the best one among all the particles, then this position is taken as the best global one instead of the old best global one.

4. Simulated Experiments

In this section, we recognize the normal state and the fault states of an online power cable by using NICPSO and NICCM. The fault states include the short circuit between phases (SCBP), the three-phase shirt circuit (TPSC), and the normal state (NS).

4.1. Experimental Data

In the system model of the online cable in Figure 3, the two power supplies are ideal, and they provide voltage of 6000 KV. The module of three-phase fault is used to produce various types of fault states. The module of three-phase V-I measurement is connected with the power line in front of the fault module to measure the transient fault signals. The sampling frequency is 10 MHz. The 1024 samples are collected after the initial traveling waves arrived. The fault starts at the 0.03th second and ends at the 0.07th second. The total length of the power cable line is 60 km.

The patterns of the states of online power cable system can be constructed by two kinds of features of phase entropy and amplitude entropy. We select the phase entropy and the amplitude entropy as follows: where is the energy function of the th frequency band at scale of the zero-component of 3-phase currents and is the coefficient of the th frequency band of wavelet packet decomposition [16].

The experimental sample numbers are given in Table 1 and partial sample data are shown in Table 2. The classes of the data include the SCBP, the TPSC, and the NS. The first dimension represents the phase entropy , and the second dimension represents the amplitude entropy in the 2-dimensional vectors of Table 2.

The distribution of the sample data is shown in Figure 4. The red symbols “” represent the training data of SCBP and the blue symbols “” are the training data of TPSC and the purple symbols “” are the training data of NS. Meanwhile, the red symbols “” are the test data of SCBP, the blue symbols “” are the test data of TPSC, and purple symbols “” are the test data of NS.

4.2. NICCM Experiment

The position changes and velocity changes of NICPSO are shown in Figures 5, 6, 7, 8, 9, 10, and 11 with different convergence factor and different inertia factor .

The changes of position and velocity are, respectively, in Figures 5 and 6 for , . Furthermore, the changes are in Figures 7 and 8 for , . Figures 9 and 10 are for , . Figures 11 and 12 are for , .

From Figures 5 and 11, it can be seen that the iteration numbers of position decrease from 15 times to 5 times for NICPSO to compare with SVM.

From Figures 6 and 12, it can be seen that the iteration numbers of velocity are the same for SVM and NICPSO. The value of both SVM and NICPSO is 12.

The classification model of SVM is trained and shown in Figure 13. Obviously, all the training samples of NS, SCBP, and TPSC are classified perfectly.

The test result of SVM model is shown in Figure 14. Training duration is 0.0575 s, and the recognition precision is 87.55%.

The classification model of NICCM is trained and shown in Figure 15. Obviously, all the training samples of NS, SCBP, and TPSC are classified correctly.

The test result of NICCM model is shown in Figure 16. Training duration is 0.0523 s, and it decreases by 0.0052 s compared with SVM. Furthermore, the test results illustrate that all the samples of NS are classified correctly, and all the faulty samples of SCBP and TPSC are confirmed as faulty state. The shortcoming is that 2 samples of SCBT are recognized as TPSC, and 1 sample cannot be confirmed whether it belongs to SCBT or TPSC. The recognition precision is 91.67%, and it increases by 4.12% compared with SVM.

4.3. Evaluation

The results obtained from the above experiments are shown in Tables 3 and 4.

5. Conclusion

In this paper, the NICPSO is developed to optimize the punishment factor and the kernel parameter of SVM, and then the NICCM is proposed to implement classification. Finally, the NICCM is applied to classify the normal state and fault states of online power cable. It is experimentally proved that the iteration number for the proposed NICPSO to reach the optimal position decreases from 15 to 5 compared with the conventional PSO. All the samples of normal state are classified correctly, and all the samples of fault state are confirmed by using NICCM. Compared with SVM, the training duration of NICCM is decreased by 0.0052 s. Meanwhile, the recognition precision of NICCM is increased by 4.12%.

NICCM is not only suitable for the classification of the normal state and fault states of online power cable but also suitable for other classifications with a small number of samples and a quick recognition process.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Acknowledgments

This research is sponsored by Scientific Research Foundation for Returned Scholars, Ministry of Education of China ([2011]508), and Natural Science Foundation of Shaanxi Province (2011JM8005).