Research Article | Open Access
Qingsong Xu, "A Comparison Study of Extreme Learning Machine and Least Squares Support Vector Machine for Structural Impact Localization", Mathematical Problems in Engineering, vol. 2014, Article ID 906732, 8 pages, 2014. https://doi.org/10.1155/2014/906732
A Comparison Study of Extreme Learning Machine and Least Squares Support Vector Machine for Structural Impact Localization
Extreme learning machine (ELM) is a learning algorithm for single-hidden layer feedforward neural network dedicated to an extremely fast learning. However, the performance of ELM in structural impact localization is unknown yet. In this paper, a comparison study of ELM with least squares support vector machine (LSSVM) is presented for the application on impact localization of a plate structure with surface-mounted piezoelectric sensors. Both basic and kernel-based ELM regression models have been developed for the location prediction. Comparative studies of the basic ELM, kernel-based ELM, and LSSVM models are carried out. Results show that the kernel-based ELM requires the shortest learning time and it is capable of producing suboptimal localization accuracy among the three models. Hence, ELM paves a promising way in structural impact detection.
Plate structure is able to provide some merits in terms of light weight, high strength, and design flexibility. Hence, plate structures have been widely employed in aerospace, civil, and mechanical engineering. Even though the structures are designed to tolerate some damage, maintenance is needed to be performed to guarantee the structure integrity for safety reasons. One usual origin of structural damage arises from the impact exerted on the structures. Hence, it is important to detect the location and magnitude of the impact force [1, 2]. Such a detection process is beneficial to the subsequent maintenance work on the plate structures.
In the literature, various techniques have been developed to facilitate the impact localization of plate structures. Generally, these methods can be classified into active and passive approaches [3, 4]. This paper is focused on the passive approach owing to its nice efficiency. The passive method is implemented using the sensors attached to structures . System modeling and data-driven techniques are two popular means in passive detection of structural damage. The former approach suffers from a major shortcoming; that is, an accurate system model is difficult to be established for complex structures, and the model error will be transferred as explicit localization errors. In contrast, the data-driven technique relies on the information of input-output data only, while it does not require the knowledge of system model. Hence, it is easier to realize and exhibits some advantages over the system modeling approach.
Data-driven techniques based on artificial intelligence (AI) are prevalent in the past two decades . In particular, neural networks (NNs) have been popularly adopted in previous works to establish a regression model [7, 8]. However, it has been recognized that there is no universal method to determine an optimal NN structure in terms of the number of hidden layers and number of neurons in each layer . Moreover, NN exhibits the risks of overfitting and sinking into local optima. Alternatively, support vector machine (SVM) is a promising tool to localize the impact on structures accurately [10, 11]. As compared with NN, SVM owns the major advantages of global optimization and higher generalization capability. Least squares support vector machine (LSSVM) simplifies traditional SVM by introducing equality constraints instead of inequality constraints . The potential of LSSVM in impact localization of plate structures has been demonstrated in the previous work .
For any AI-based data-driven algorithms, a practical concern is the learning time. Generally, an NN requires considerable learning time to generate an accurate regression model, which blocks its applications in rapid implementation scenarios. To overcome this issue, extreme learning machine (ELM) has been recently proposed to realize an extremely fast learning [13, 14]. ELM is a learning algorithm for single-hidden layer feedforward neural network. In theory, this algorithm is able to provide good generalization capability at extremely fast learning speed. Comparative studies of benchmark function approximation problems reveal that ELM can learn thousands of times faster than conventional NN and can produce good generalization performance in most cases.
ELM has been widely applied to solve different engineering problem, such as nonlinear hysteresis modeling . However, the research on impact localization using ELM is rare in the literature. In this paper, the ELM is introduced into the domain of structural impact localization. Experimental study is carried out using LSSVM and ELM for the impact location detection of a plate structure. Comparative investigations in terms of learning time and localization accuracy are carried out in detail. It is shown that ELM paves a promising way in the domain of structural impact localization, because it provides the fastest learning time along with suboptimal localization accuracy.
The remainder of the paper is organized as follows. Section 2 presents the data processing technique that is employed in this work. Section 3 gives a brief review of the LSSVM and ELM algorithms. The experimental setup and data acquisition are described in Section 4. Section 5 conducts a collection of experimental studies using LSSVM and ELM approaches along with a detailed comparative study. Section 6 summarizes this paper.
2. Data Processing
In this research, the impacts are exerted on an aluminum structural plate. A set of four piezoelectric (PZT) sensors are mounted on the surface of the plate to acquire the induced acoustic waves. The sensor outputs are acquired with an NI PCI-4472 card using a sampling rate of 100 kHz. In order to reduce the dada dimension, useful features are extracted from the time-history signal. Specifically, the magnitude of the maximum response and its corresponding time are employed because these two features are the best among the known features .
To extract the two features from each sensor output, the Hilbert transform is employed to solve the envelope of each sensor signal. Then, the two features of each sensor output are extracted. More details can be found in the literature .
The four sensors provide a total of eight features for the modeling of impact location. To generate a better modeling accuracy, it is suggested to normalize the input data into the range between −1 and 1 . However, in this research, preliminary testing reveals that the normalization to zero mean and unit variance produces a better prediction result for the impact location problem. Hence, each of the eight signals is normalized through the transformation: where and represent the mean and variance of the signal, respectively.
3. LSSVM and ELM Modeling
In this section, the employed LSSVM and ELM algorithms are briefly described.
Generally, given an input data, LSSVM maps it into a high-dimensional feature space and constructs a linear regression function therein. In this research, the LSSVM as described in  is employed. Moreover, the Bayesian inference is used to tune the hyperparameter in an optimal manner. More details about its implementation can be referred to in the literature [5, 17].
3.2. Basic ELM
Given arbitrary samples (, ) with and , a standard single-layer feedforward neural network with hidden neurons and activation function can be mathematically described as where is the weight vector connecting inputs and the th hidden neurons, is the weight vector connecting the th hidden neurons and output neurons, is the threshold of the th hidden node, and is the ELM output for the th data point.
This network can approximate these samples with zero error . That is, there exist , , and to meet The above equations can be expressed into a compact form: where In ELM, the hidden-node parameters (, ) remain fixed once randomly generated. To train the network is simply equivalent to find a least square solution of the linear system (4); that is, The smallest norm least square solution to (4) is where is the Moore-Penrose generalized inverse of the matrix .
Therefore, given a training set , the basic ELM algorithm is very straightforward as stated below .
Step 1. Randomly generate hidden-node parameters (, ), .
Step 2. Calculate the hidden layer output matrix .
Step 3. Calculate the output weight vector as .
3.3. Kernel-Based ELM
In the case that the hidden layer feature mapping is unknown to the user, a kernel matrix of ELM can be defined as follows: The output function of ELM becomes In this kernel-based ELM, the hidden layer feature mapping needs not to be known to the user, while the kernel is given to the user. In addition, the number of hidden nodes needs not to be specified.
Hence, given a training set , the kernel ELM algorithm is further simplified as a single learning step implementation as follows .
Step 1. Calculate the output function:
4. Experimental Setup and Data Acquisition
4.1. Experimental Setup
The experimental setup consists of a clamped aluminum plate of size 490 mm × 390 mm, four piezoelectric sensors glued on the plate surface, a data acquisition card plugged in a personal computer, and an instrumented hammer. The experimental setup is shown in Figure 1.
The four sensors are mounted on the plate surface by conductive glue. The distances between sensors and the nearest edges are 170 mm (horizontally) and 100 mm (vertically), respectively. The dimension of the plate and the location of the glued sensors are shown in Figure 2.
For data acquisition, an NI PCI-4472 data acquisition card was plugged in a PCI slot of a personal computer. The time-varying data were gathered by a developed LabVIEW program. Preliminary experiment shows that when the hammer knocks on the plate, the voltages collected by sensors increase rapidly initially and then decay to zero rapidly. During this transient response period, the sensor output voltage signals were recorded and saved in data files. The data acquisition rate is set as 100 kHz.
4.2. Data Acquisition
Through experimental impact testing, two sets of impact data were acquired with the aforementioned experimental setup. The first set includes a regular grid of 63 impacts, as shown in Figure 3. It is used for the training of the LSSVM and ELM models. Another set of 90 random impacts, as shown in Figure 4, is employed for the model testing.
The data acquisition and impact localization were executed on a personal computer (PC) equipped with 2.83 GHz Quad CPU and 8 GB RAM running Windows 7 (64-bit) operating system.
4.3. Performance Index
To quantify the localization accuracy, an error function is employed to provide a measure of the accuracy. Specifically, to visualize the results, the averaged errors of and coordinates are calculated, and the ratio between the product of them and the plate area is considered as the evaluation index: where denotes the area of the plate structure, and is the ratio between the error area and the plate area. It is found that the lower the index is, the better the detection accuracy for the impact location will be.
5. Results and Discussion
5.1. LSSVM Results
Using the approach as described previously, the LSSVM model is trained with the training set as shown in Figure 3 and then tested with the testing set as illustrated in Figure 4. Moreover, two LSSVM models are generated to predict the and coordinates, respectively. To get a reliable result, the LSSVM model is trained by 30 times. The training error of a typical result is shown in Figure 5, which leads to a low training error index %.
Using the two trained LSSVM models, the testing results for and coordinates are shown in Figure 6(a). In addition, the discrepancies between the predicted and actual and coordinates are illustrated in Figure 6(b). It is calculated that a testing error index of 0.7358% is resulted for the 90 sets of testing data samples.
5.2. Basic ELM Results
Concerning the basic ELM algorithm, the activation function is selected as the sigmoid function. Two ELM models are established for the prediction of the and coordinates, respectively.
As the number of hidden nodes increases, the training and testing accuracy of the basic ELM models for the and coordinate predictions are depicted in Figures 7(a) and 7(b), respectively. It is found that when the number , which is close to the number (63) of training samples, the training error becomes zero, whereas the testing error is greatly increased. To generate stable results for the two basic ELM models, the number of hidden nodes is selected as .
In addition, the training of each model is conducted by 30 times. The training errors of a typical result are illustrated in Figure 8. It is found that the ELM models produce a zero training error index approximately. This indicates that the basic ELM exhibits a very nice function approximation capability.
Using the trained basic ELM models, the testing results are generated as shown in Figure 9. For a clear visualization, the testing errors of the two coordinates are depicted in Figures 9(a) and 9(b), respectively. It is calculated that the ELM models create a large testing error index of 3.8980%.
5.3. Kernel ELM Results
Regarding the kernel ELM model, the radial basis function (RBF) kernel is selected owing to its popularity. Two kernel ELM models are developed for the detection of and coordinates of the impact. By selecting the regularization coefficient , preliminary results indicate that the training and testing performances of ELM models are dependent on the RBF parameters.
As the RBF parameter increases, the training and testing accuracies for the and coordinates are shown in Figures 10(a) and 10(b), respectively. It is seen that the training errors increase with the rising of the RBF parameter. Whereas the testing errors arrive at the minimum values when the RBF parameters are 65 and 20 for and coordinates, respectively. Hence, to produce better testing results, the RBF parameters of the two ELM models are selected as 65 and 20 for the prediction of and coordinates, respectively.
By running each model 30 times, it is found that each time leads to the same result. The training errors of the two kernel ELM models are illustrated in Figure 11, which results in a training error index of 2.0154%. Using the two trained kernel ELM models, the testing results for the two coordinates are obtained, as shown in Figure 12(a). The corresponding testing errors of the and coordinate predictions are shown in Figure 12(b). The kernel ELM models produce the testing error index of 0.7746%.
5.4. Discussion on Localization Results
For a clear comparison of the detection performance of the LSSVM, basic ELM (BELM), and kernel ELM (KELM) models, the results are tabulated in Table 1. The training time means the time required by each model for the model training for a single coordinate prediction. The testing time is the execution time needed by each model to predict a single coordinate of the impact.
By comparing the training and testing time, it is evident that the KELM requires the minimum time for both training and testing processes. Whereas the LSSVM needs the maximum time for the model training and BELM calls for the longest execution time for the impact localization.
In addition, given the training samples, it is deduced from Table 1 that the BELM is able to produce the minimum training error, which is equal to zero approximately. In contrast, the KELM creates the maximum training error. Therefore, the BELM provides an excellent function approximation capability.
Moreover, from the testing accuracy of the three models, it is found that the LSSVM is able to deliver the most accurate localization result. It is further observed that the BELM gives the lowest accuracy of the impact localization although it produces the best training error. By comparing the testing accuracy of KELM and LSSVM, it is seen that the KELM localization error is only 5.3% higher than that of LSSVM.
In summary, the KELM produces the suboptimal localization accuracy among the three models, while it requires the shortest training and testing execution time. Therefore, the KELM is more suitable for the scenarios where the execution time is the most critical requirement, whereas the LSSVM is more appropriate to the situations when the localization accuracy is most desirable.
This paper presents a comparison study of LSSVM and ELM dedicated to impact localization of a plate structure with surface-mounted PZT sensors. By extracting useful features through Hilbert transform of each sensor output, regression models are established to predict the location of impact exposed by a hammer. Both basic ELM and kernel-based ELM algorithms are employed for comparison investigation with respect to LSSVM approach. Experimental results reveal that the basic ELM is able to provide a nice function approximation capability. The kernel ELM requires the minimum execution time for training and testing processes, while LSSVM is able to produce the most accurate localization result. In addition, kernel ELM gives the suboptimal localization accuracy. Therefore, the kernel ELM is more suitable for the scenarios where the execution time is the most critical requirement, whereas the LSSVM is more appropriate to the situations when the localization accuracy is most desirable.
In the future, more advanced ELM algorithm will be attempted to further improve its localization accuracy while maintaining the smallest execution time. Moreover, its application in real impact localization of aircraft structures will be carried out in the future work.
Conflict of Interests
The author declares that there is no conflict of interests regarding the publication of this paper.
This work was supported by the Macao Science and Technology Development Fund under Grant 070/2012/A3 and the Research Committee of the University of Macau under Grants MYRG083(Y1-L2)-FST12-XQS and MYRG078(Y1-L2)-FST13-XQS.
- Q. Wang and X. Deng, “Damage detection with spatial wavelets,” International Journal of Solids and Structures, vol. 36, no. 23, pp. 3443–3468, 1999.
- J. Haywood, P. T. Coverley, W. J. Staszewski, and K. Worden, “An automatic impact monitor for a composite panel employing smart sensor technology,” Smart Materials and Structures, vol. 14, no. 1, pp. 265–271, 2005.
- V. Giurgiutiu, “Tuned Lamb wave excitation and detection with piezoelectric wafer active sensors for structural health monitoring,” Journal of Intelligent Material Systems and Structures, vol. 16, no. 4, pp. 291–305, 2005.
- S. F. Masri, A. W. Smyth, A. G. Chassiakos, T. K. Caughey, and N. F. Hunter, “Application of neural networks for detection of changes in nonlinear systems,” Journal of Engineering Mechanics, vol. 126, no. 7, pp. 666–676, 2000.
- Q. Xu, “Impact detection and location for a plate structure using least squares support vector machines,” Structural Health Monitoring, vol. 13, no. 1, pp. 5–18, 2014.
- L. E. Mujica, J. Vehí, W. Staszewski, and K. Worden, “Impact damage detection in aircraft composites using knowledge-based reasoning,” Structural Health Monitoring, vol. 7, no. 3, pp. 215–230, 2008.
- R. T. Jones, J. S. Sirkis, and E. J. Friebele, “Detection of impact location and magnitude for isotropic plates using neural networks,” Journal of Intelligent Material Systems and Structures, vol. 8, no. 1, pp. 90–99, 1997.
- K. Worden and W. J. Staszewski, “Impact location and quantification on a composite panel using neural networks and a genetic algorithm,” Strain, vol. 36, no. 2, pp. 61–70, 2000.
- J. R. LeClerc, K. Worden, W. J. Staszewski, and J. Haywood, “Impact detection in an aircraft composite panel: a neural-network approach,” Journal of Sound and Vibration, vol. 299, no. 3, pp. 672–682, 2007.
- J. A. K. Suykens, “Support vector machines: A nonlinear modelling and control perspective,” European Journal of Control, vol. 7, no. 2-3, pp. 311–327, 2001.
- H. Fu and Q. Xu, “Locating impact on structural plate using principal component analysis and support vector machines,” Mathematical Problems in Engineering, vol. 2013, Article ID 352149, 8 pages, 2013.
- Q. Xu and P. Wong, “Hysteresis modeling and compensation of a piezostage using least squares support vector machines,” Mechatronics, vol. 21, no. 7, pp. 1239–1251, 2011.
- G. Huang, Q. Zhu, and C. Siew, “Extreme learning machine: Theory and applications,” Neurocomputing, vol. 70, no. 1–3, pp. 489–501, 2006.
- G.-B. Huang, D. H. Wang, and Y. Lan, “Extreme learning machines: a survey,” International Journal of Machine Learning and Cybernetics, vol. 2, no. 2, pp. 107–122, 2011.
- R. Dong and Y. Tan, “Nonlinear modeling of hysteresis in piezoelectric actuators,” in Proceedings of the 9th IEEE International Conference on Control and Automation (ICCA '11), pp. 1250–1254, December 2011.
- S. T. Quek, P. S. Tua, and Q. Wang, “Detecting anomalies in beams and plate based on the Hilbert-Huang transform of real signals,” Smart Materials and Structures, vol. 12, no. 3, pp. 447–460, 2003.
- J. A. K. Suykens, T. Van Gestel, J. De Brabanter, B. De Moor, and J. Vandewalle, Least Squares Support Vector Machines, World Scientific, Singapore, 2002.
Copyright © 2014 Qingsong Xu. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.