Abstract

A novel evolutionary extreme learning machine (ELM) based on improved quantum-behaved particle swarm optimization (IQPSO) for radar target classification is presented in this paper. Quantum-behaved particle swarm optimization (QPSO) has been used in ELM to solve the problem that ELM needs more hidden nodes than conventional tuning-based learning algorithms due to the random set of input weights and hidden biases. But the method for calculating the characteristic length of Delta potential well of QPSO may reduce the global search ability of the algorithm. To solve this issue, a new method to calculate the characteristic length of Delta potential well is proposed in this paper. Experimental results based on the benchmark functions validate the better performance of IQPSO against QPSO in most cases. The novel algorithm is also evaluated by using real-world datasets and radar data; the experimental results indicate that the proposed algorithm is more effective than BP, SVM, ELM, QPSO-ELM, and so on, in terms of real-time performance and accuracy.

1. Introduction

Radar target classification technology is of great significance in both military and civil aspects [1, 2]. At present, the commonly used classification methods include Bayesian [3], Dempster-Shafer (D-S) theory [4], decision tree [5], support vector machine (SVM) [6], and back propagation (BP) neural network [7]. Although these methods can obtain good classification accuracy, their common problem is that the real-time performance is not high. ELM [8], as a new learning algorithm for single-hidden-layer feedforward neural networks (SLFNs), has attracted great concerns from various fields for its fast learning speed, such as traffic sign recognition [9], face recognition [10], human action recognition [11], and image analysis [12]. The idea of the algorithm is to generate the input weights and hidden biases randomly and then train the network by solving the norm least-squares solution of the output weights [13]. ELM not only has faster learning speed than traditional learning methods, but also has good generalization performance in many applications [14]. In order to further improve the ELM algorithm, the researchers put forward many improved algorithms. The fully complex ELM (C-ELM) is proposed in [15], which extends the ELM algorithm from the real domain to the complex domain. Considering that when new data is received, many training methods use the past data together with the new data and perform a retraining, which consumes lots of time, an online sequential ELM (OS-ELM) is proposed in [16], which can learn the training data one by one or chunk by chunk and discard the data for which the training has already been done. To get the better predicting performance, a new adaptive ensemble model of ELM (Ada-ELM) is proposed in [17], which can adjust the ensemble weights automatically. Considering that the performance of ELM is affected by hidden layer nodes and the number of hidden layer nodes is difficult to determine, the incremental ELM (I_ELM) [18], pruned ELM (P_ELM) [19], and self-adaptive ELM (SaELM) [20] have been proposed. Note that traditional ELM only utilizes labeled data to carry out the supervised learning task; [21] applied manifold regularization (MR) to ELM to exploit unlabeled data in the ELM model. Moreover, [22] proposed sparse Bayesian ELM (SBELM) which has the advantages of the two algorithms.

However, due to the random selection of input weights and hidden biases, ELM tends to need a large number of hidden nodes for better generalization, which may increase the complexity of the network. In order to solve this problem, an improved ELM method based on particle swarm optimization (PSO) is proposed in [23]. This method can resolve the drawbacks of ELM. But, at the same time, we also know that PSO easily has premature convergence and has low robustness due to the fact that its global search ability relies on the up-limit of velocity [24]. To discourage premature convergence, [25] proposed comprehensive learning particle swarm optimizer (CLPSO), which utilized all other particles’ historical best information to update a particle’s velocity. In [26], Sun et al. introduce quantum theory into PSO and put forward a quantum-behaved PSO (QPSO) algorithm, which outperforms PSO in search ability and has fewer parameters to control. QPSO and its improved model have been applied in many ways [27, 28]. The enhanced weighted quantum PSO (EWQPSO) has been developed to perform the design of the supershaped lens antennas yielding optimal antenna performance [27] and the random local optimized QPSO (RLQPSO) has been used in fast threshold image segmentation [28]. In [2931], QPSO was applied to ELM to improve the algorithm performance. However, as with other evolutionary algorithms, QPSO also has the problem of premature convergence [32]. References [33, 34] proposed QPSO with random mean best position and weighted mean best position, respectively. Reference [35] introduced a novel search strategy with a selection operation into QPSO. In the modified QPSO (MQPSO), the global best position is substituted by a personal best position of a randomly selected particle. Although QPSO algorithm has better global convergence than PSO, this method calculates the characteristic length of Delta potential well only according to the mean best position, which will reduce the global search ability of the algorithm. In order to overcome this problem, this paper presents a new method to calculate the characteristic length of Delta potential well. Then the improved QPSO (IQPSO) algorithm is used to optimize the weights and biases of ELM.

The rest of this paper is organized as follows: Section 2 introduces the relevant theoretical knowledge of ELM and QPSO. In Section 3, we present the improved formula for calculating the characteristic length of Delta potential well of QPSO and the application of IQPSO in parameter optimization of ELM. Experimental results are analyzed in Section 4, and Section 5 summarizes the paper.

2.1. Extreme Learning Machine

Given a set of training datasets , where , , and , is an -dimensional input vector and is the expected output. The output function of ELM with hidden nodes is represented as follows:where is the weight vector of input nodes to hidden nodes and is the bias of hidden node, is the weight vector between hidden nodes and the output nodes, is the activation function of hidden layer, and is the output vector.

If the SLFNs with hidden nodes can approximate the samples with zero error, we know that (1) can be converted to the following formula:

The above equations can be written aswhere

So training the SLFNs corresponds to finding the norm least-squares solution , which can be shown as follows:where is the Moore-Penrose generalized inverse of hidden layer output matrix .

Then, according to KKT theorem, (6) can be expressed aswhere is the unit matrix and is the regularization coefficient.

Thus, the learning steps of the ELM can be summarized as follows.

Step 1. Determine the structure of neural networks and set random values to the input weights and the hidden layer biases .

Step 2. Calculate the hidden layer output matrix according to (4).

Step 3. Calculate the output weight vector according to (6).

2.2. Quantum-Behaved Particle Swarm Optimization

Given the particle swarm where represents the swarm size, denotes the dimension of the search space. At each iteration, the position of the particle can be expressed as . The personal best position can be expressed as and the global best position can be expressed as . The personal best position and the global best position are updated using the following formula:where is the fitness function.

In [30], it can be seen that the position of the particle can be updated by using (9)where and are two random numbers uniformly distributed in the interval (0,1) and is the characteristic length of Delta potential well, and its value is directly related to the convergence speed and search ability of the algorithm. In the traditional QPSO, is determined bywhere is the mean best position of the swarm and . is the contraction-expansion coefficient, which can be adjusted to control the convergence speed of the algorithm. There are two ways to set : fixed parameter and linear reduction. If the linear reduction method is adopted, then , where is the maximum number of iterations and is the current number of iterations.

3. Proposed Approach

3.1. Improved Quantum-Behaved Particle Swarm Optimization

The method of (11) has the following problems: since is a relatively stable value, the search information of the whole swarm cannot be effectively utilized. Especially when most particles fall into local optima, only a few particles are distributed in other regions; the will also tend to the local optimum. As a result, most particles only perform local search. This will reduce the global search ability of the algorithm. To overcome this problem, this paper presents a new method to calculate .where is the randomly selected particle in the swarm and , is the average fitness of the swarm, and .

Compared with (11), (12) takes into account the particles distributed in other regions. It can be seen from (12) that even if most particles fall into local optima, the particles still have moderate probability of jumping out of local optima due to the particles distributed in other regions. The improved method is able to improve the global search ability of QPSO.

The flow chart of IQPSO is shown in Figure 1.

3.2. ELM Based on IQPSO Algorithm

We know from Section 2 that ELM has the advantages of fast learning speed and easy implementation. However, ELM tends to require a large number of hidden nodes to get better performance, which complicates the network structure. In addition, the hidden layer output weights are determined by the random input weights and biases; as a result, the final output may be instable. This paper proposed IQPSO-ELM algorithm to solve the shortcoming of ELM.

The IQPSO-ELM can be divided into three parts: initial ELM, trained ELM, and test ELM as shown in Figure 2. Each part of the algorithm is summarized as follows.(1)Initial ELM: we first obtain training samples, validation samples, and test samples. Then set the number of input nodes of ELM equal to the dimension of input data and set the number of output nodes equal to the number of sample classes. Of course, we also need to set hidden layer nodes, as well as the swarm size and the maximum number of iterations(2)Trained ELM: in this process, appropriate input weights and hidden layer biases are obtained by IQPSO to train ELM. First, the particle dimension is calculated according to the formula , where and represent the number of input nodes and hidden layer nodes, respectively. Then the position of particle is randomly initialized according to the particle dimension and swarm size . can be written as . Finally, according to IQPSO introduced in Section 3.1, we can get the global best position , which can be converted to the input weights and hidden layer biases of ELM. What needs to be explained is that the fitness of IQPSO is represented by the correct classification rate.(3)Test ELM: test samples are used to evaluate the effectiveness of the proposed method.

4. Experimental Result and Discussion

In this section, we will verify the effectiveness of the proposed algorithm. The experiments are performed on the Intel(R) Core(TM) 3.60 GHz CPU, with 8 GB of RAM, and Matlab R2013a environment.

4.1. The Performance of IQPSO on Benchmark Functions

In order to demonstrate the effectiveness of IQPSO, four benchmark functions (see Table 1) are selected for the experiment. The performance of IQPSO is compared with QPSO [26] and MQPSO [35]. We set different swarm size for the four benchmark functions with different dimensions. values are 20, 40, and 80. The maximum number of iterations is set as 1000, 1500, and 2000 corresponding to the dimensions 10, 20, and 30 for the four benchmark functions, respectively. The value of contraction-expansion coefficient decreases from 1.0 to 0.5 linearly. The mean values of the best fitness values for 50 runs of each function are recorded in Tables 25, respectively.

According to Tables 25, the results show that the IQPSO works better than QPSO and MQPSO on Sphere function. IQPSO can also generate better results than the other methods in most cases on Rosenbrock function. On Ackley function, we know that the proposed method outperforms the QPSO but is not as good as MQPSO when the swarm size is 80. In addition, on Rastrigin function, the IQPSO is superior to QPSO and MQPSO except when the swarm size is 40 and dimension is 10. Generally speaking, the results show that the IQPSO has better global search ability than QPSO and MQPSO.

Figure 3 shows the convergence process of QPSO, MQPSO, and IQPSO on the four benchmark functions when the swarm size is 20, the dimension is 30, and the number of iterations is about 2000. It can be seen from the figures that the convergence speed of the proposed method is much faster than that of the other two methods. Then we know that the proposed method has strong ability of global optimization and could generate better solutions.

4.2. The Classification Performance of IQPSO-ELM on Real-World Datasets

In this section, some real-world datasets such as Diabetes, Liver-disorders, and Auto-MPG are used to test the proposed algorithm. What needs to be explained is that Diabetes dataset is a dataset about diabetes, which includes patient physiological data and disease progression after one year; the dataset is divided into 2 classes, with a total of 768 samples; each sample has 8 kinds of attribute values, that is, the number of features. Liver-disorders dataset is a medical research database donated by Richard S. Forsyth. There are 345 samples in this dataset, each sample has 6 attribute values, and the dataset is divided into 2 categories. Auto-MPG dataset, which concerns the city-cycle fuel consumption, is taken from the StatLib library which is maintained at Carnegie Mellon University. There are 398 samples in Auto-MPG dataset and each sample has 7 attribute values. The Auto-MPG dataset is divided into 3 classes. The detailed description of the three datasets is listed in Table 6. It is important to note that the training samples, test samples, and validation samples are randomly generated according to the number listed in Table 6. Figure 4 shows the performance comparison of five algorithms, that is, ELM, PSO-ELM, QPSO-ELM, MQPSO-ELM, and IQPSO-ELM. We should know that the number of hidden nodes of the five algorithms is 10 and the maximum number of iterations of evolutionary ELM algorithms is 100. In order to avoid accidental results, the algorithms run 50 times, respectively, and then calculate the average value of results.

From Figure 4, we know that, in the aspect of training and testing accuracy, the proposed method can achieve higher accuracy on the given three datasets, which shows the effectiveness of the proposed method. Figure 5 shows the testing accuracies of ELM with varying hidden nodes on the three datasets. As can be seen from Figure 5, with the increase of the number of hidden nodes, the testing accuracy of ELM has been improved. For the Diabetes dataset, when the hidden layer node is set to 30, the test accuracy is up to 0.7598, for the Liver-disorders dataset, when the number of hidden nodes is 95, the test accuracy is 0.6601, and for the Auto-MPG dataset, when the number of hidden layer nodes is 95, the test accuracy is up to 0.7123. Although ELM can greatly improve the classification speed, in order to achieve better performance, ELM needs more hidden nodes which may increase the complexity of the network. The proposed method can use a simple network to obtain a good classification result.

4.3. The Classification Performance of IQPSO-ELM on Radar Data

In this section, simulated data and darkroom measured data are utilized to verify the validity of IQPSO-ELM. The performance comparison of BP, SVM, ELM, and evolutionary ELM including PSO-ELM, QPSO-ELM, MQPSO-ELM, and IQPSO-ELM in this experiment is given in Tables 8 and 9. It should be noted that the number of hidden nodes of evolutionary ELM is set to 10, the swarm size is 20, and the maximum number of iterations is about 100. The number of hidden nodes of BP is set to 10, the activation functions are “logsig” and “purelin,” the training function is “trainlm,” and the learning rate is 0.01. The kernel function of SVM is “guass” and the penalty factor is 0.2. The number of hidden nodes of ELM is 10 and the hidden layer activation function is “sigmoid.”

4.3.1. Simulated Data

Radar target classification technology is of great significance in both military and civil aspects. The classification accuracy and real-time performance are particularly important. Previously widely used classification methods, such as BP and SVM, have the problems of low classification accuracy or long time consumption. It is hard for them to be excellent in two aspects of the real-time performance and accuracy. To solve this issue, we propose IQPSO-ELM algorithm, which not only makes full use of the advantages of fast learning speed and good generalization performance of ELM, but also utilizes the improved QPSO to obtain the appropriate input weights and hidden layer biases. The proposed method is able to solve the above problem very well.

In this section, we utilize the simulated data to validate the proposed algorithm. The simulated radar targets include conical target, cylindrical target, cone-like target, and spherical target. The size of each target model is shown in Table 7. It is assumed that the radar observation time is from 100 to 600 s. The selected target features include RCS mean, RCS variance, scatter centers, and Micro-Motion period. According to the prior knowledge and the related work [1, 2, 36], the data is simulated by a certain random error added to a set of truth values. In the simulated data, the RCS mean value of the target is set to −5 dB, −4 dB, −4.5 dB, and −2 dB, respectively, and the relative error is 0.3, the RCS variance is set to 0.5, 0.8, 1.0, and 0.3, respectively, and the relative error is 0.1, the Micro-Motion period is set to 5 s, 4 s, 3 s, and 1 s respectively, and the relative error is 0.3, and the number of scattering centers is set to 2, 2, 3, and 1, respectively, and the relative error is 1. The simulated data is shown in Figure 6.

From Table 8, we know that the BP, SVM, ELM, and the evolutionary ELM including PSO-ELM, QPSO-ELM, MQPSO-ELM, and IQPSO-ELM algorithms have good effect on radar target classification. In the training phase, the ELM takes the least time and shows its fast learning ability. The evolutionary ELM algorithms take more time than other algorithms; that is because the evolutionary ELM algorithms need to spend some time on optimizing the input weights and hidden biases. In the testing phase, although the accuracy of the proposed algorithm is slightly lower than that of SVM, the test time is greatly reduced. Compared with ELM, although the test times of the two algorithms are not much different, the accuracy of proposed method is higher. From Figure 7, we know that, in order to achieve better classification performance, ELM needs more hidden layer nodes, which results in complex network structure. The experimental results show that the proposed algorithm can meet the requirements of real-time performance and accuracy in radar target classification.

4.3.2. Darkroom Measured Data

In this section, darkroom measured data is used to verity the validity of the proposed method. The radar targets include the conical target, cylindrical target, and cone-like target. The features of radar targets are shown in Figure 8, including RCS mean, RCS variance, Micro-Motion period, and Micro-Motion amplitude. Compared with the simulated data, due to the scene complexity, the darkroom measured data has the characteristics of large dynamic range and low stability.

According to Table 9, the evolutionary ELM methods take more time in the training process; that is because the evolutionary ELM algorithms need to spend some time on optimizing the input weights and hidden biases. In the testing process, the accuracy of the proposed method is improved compared with BP and ELM; although not as good as SVM, the proposed method takes less time than SVM. It can be seen from Figure 9 that, in order to get better classification effect, ELM needs more hidden nodes, which makes the network structure more complex. Therefore, considering the two aspects of the real-time and accuracy, the IQPSO-ELM method is better.

5. Conclusions

In this paper, we proposed a novel evolutionary extreme learning machine based on improved quantum-behaved particle swarm optimization. Further, we introduced it to radar target classification. The proposed method not only makes full use of the advantages of fast learning speed and good generalization performance of ELM, but also utilizes the improved QPSO to obtain the appropriate input weights and hidden layer biases, which is able to solve the problem that the ELM needs more hidden nodes to get better classification performance. The performance of the IQPSO method is evaluated on the well-known benchmark functions. The experimental results show that the proposed method can not only achieve the best solutions, but also converge to the optimal solution faster than other methods. Moreover, the results of the experiment on the real-world datasets show that the proposed IQPSO-ELM method can achieve good performance. Finally, some experiments on radar target classification verify the effectiveness of IQPSO-ELM. Although the classification accuracy of the proposed method is a little lower than that of SVM, it runs much faster than SVM. The experimental results show that the proposed method is more cost-efficient than other traditional radar target classification methods. We are absolutely convinced that the work presented in this paper is extremely significant for radar target classification. We will further optimize our proposed method and expand the scope of its application in the near future.

Acronyms

ELM:Extreme learning machine
PSO:Particle swarm optimization
QPSO:Quantum-behaved particle swarm optimization
IQPSO:Improved quantum-behaved particle swarm optimization
D-S:Dempster-Shafer
SVM:Support vector machine
BP:Back propagation
SLFNs:Single-hidden-layer feedforward neural networks
C-ELM:Complex ELM
OS-ELM:Online sequential ELM
Ada-ELM:Adaptive ensemble model of ELM
I_ELM:Incremental ELM
P_ELM:Pruned ELM
SaELM:Self-adaptive ELM
MR:Manifold regularization
SBELM:Sparse Bayesian ELM
CLPSO:Comprehensive learning particle swarm optimizer
EWQPSO:Enhanced weighted quantum PSO
RLQPSO:Random local optimized QPSO
MQPSO:Modified QPSO.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

The work in this paper has been supported by the National Science Foundation of China (61422114 and 61501481) and the Natural Science Fund for Distinguished Young Scholars of Hunan Province under Grant no. 2015JJ1003.