Research Article | Open Access
Sparse Representation Based SAR Vehicle Recognition along with Aspect Angle
As a method of representing the test sample with few training samples from an overcomplete dictionary, sparse representation classification (SRC) has attracted much attention in synthetic aperture radar (SAR) automatic target recognition (ATR) recently. In this paper, we develop a novel SAR vehicle recognition method based on sparse representation classification along with aspect information (SRCA), in which the correlation between the vehicle’s aspect angle and the sparse representation vector is exploited. The detailed procedure presented in this paper can be summarized as follows. Initially, the sparse representation vector of a test sample is solved by sparse representation algorithm with a principle component analysis (PCA) feature-based dictionary. Then, the coefficient vector is projected onto a sparser one within a certain range of the vehicle’s aspect angle. Finally, the vehicle is classified into a certain category that minimizes the reconstruction error with the novel sparse representation vector. Extensive experiments are conducted on the moving and stationary target acquisition and recognition (MSTAR) dataset and the results demonstrate that the proposed method performs robustly under the variations of depression angle and target configurations, as well as incomplete observation.
In the recent years, sparse representation has attracted much attention in the fields of signal representation, compress sensing, and classification. The sparse representation classification (SRC) algorithm, which is proposed by Wright et al. , has boosted the classification method on many subjects such as face recognition , hyperspectral image classification , and synthetic aperture radar (SAR) automatic target recognition (ATR) [4–6].
In particular, with the development of SAR imaging techniques including high resolution and multipolarization, much effort has been devoted to SAR ATR. The moving and stationary target acquisition and recognition (MSTAR) dataset , which collects the SAR images of typical vehicles under the conditions of various radar grazing angles and target aspect angles, is a benchmark for the development and evaluation of recognition algorithms. One of the most popular methods for MSTAR classification is template matching. Ross et al.  decomposed the train samples of each class into 36 templates with an aspect range of 10° and then classified the test samples based on the distance measurement compared with the templates. Ravichandran and Casasent  proposed the minimum noise and correlation energy (MINACE) filter method to achieve an optimal classification result. The learning vector quantization (LVQ) method , which acquires the train samples with learning, is another template matching method. Besides the previous template matching methods that directly applied on image pixels, the feature-based template matching methods improve the performance. Ramamoorthy and Casasent  extract the rotation invariant Fourier features and propose a feature space trajectory (FST) classifier. Mishra and Mulgrew  investigate the classification of MSTAR targets based on principle component analysis (PCA). Yang et al.  summarize and compare the various classifiers for MSTAR target classification.
Through representing the test sample as the combination of training samples, sparse representation classification, which can be considered as a generalization of the LVQ, determines the class of the test sample based on the resulted sparsest coefficients. The sparse coefficients contain discriminatory information of the samples in low-dimensional subspace and are robust to noise and occlusion as well as incomplete observation . However, the target presents diverse appearance and heavy occlusion on SAR image according to the target’s aspect angle. The sparse coefficients that lie on a large difference of aspect angle lead to classification errors. In this paper, we propose a novel SAR vehicle classification method based on SRC along with aspect angle (SRCA). The method first estimates the aspect angle of all the samples and solves the sparse representation vector with a dictionary that consists of the PCA features of all the training samples. Then, we project the sparse coefficient vector onto a subspace that is around the test sample’s aspect angle. Finally, the method assigns the test sample to a certain category, which minimizes the reconstruction error with the novel sparse representation vector. We validate our proposed method by testing on a subset of the MSTAR dataset. It is shown that the proposed method is superior to the methods of linear SVM, kernel SVM, and the original SRC.
2. Sparse Representation Based Classification
In this section, we give a brief review on the sparse representation and the classification strategy, that is, how to represent a test sample as the combination of training samples from a dictionary  and determine the class based on the sparse representation vector. Note that the test sample represents a vehicle or other objects such as a face. In this letter, we focus on vehicles.
2.1. Sparse Representation 
Suppose that there are distinct classes of vehicles and the labeled training samples for the classes are known a priori; then, our objective is to correctly determine the class of a new test vehicle sample. Assume that the number of training samples for the th class is and that the dimension of each sample is , and denote by the th training sample for class . Then, the matrix for class containing all training samples is given by
Accordingly, all training samples for the classes are concatenated into a dictionary matrix ; that is, where denotes the total number of training samples. With the dictionary at hand, we consider an observed new test sample denoted by . If this sample belongs to class , then it can be well approximated by a linear combination of the training samples in the th class; that is, in which the scalar is the weighted coefficient associated with the th training sample of class to reconstruct the sample . Correspondingly, the linear representation of using all training samples in the dictionary can be denoted by with being a sparse weighted coefficient vector whose entries are all zero except those associated with the th class.
With a sufficiently large number of samples for each class, the coefficient vector is expected to be very sparse. Based on the recent development in the theory of sparse representation and compressive sensing, the solution of can be recovered via solving the following -norm minimization problem [1, 2]: where denotes the -norm of , which sums up the absolute values of all entries in . Moreover, the equality constraint in (5) can be relaxed to allow noise; that is, problem (5) can be relaxed as where is the allowed error tolerance. Problems (5) and (6) can be recast as linear programs (LP) and second-order cone programs (SOCP), respectively. Thus, they are both convex and can be solved by existing convex optimization software . It is worth noting that the complexity of solving the SOCP is , where and are the dimension of the sample and the total number of training samples, respectively (c.f. (1), (2)).
2.2. Sparse Representation Based Classification
With the sparsest coefficient at hand, the SRC method determines the class of the test sample in the following. Ideally, if all the nonzero entries in the estimate are associated with one single class , then we can easily determine the th class that the test sample belongs to. However, due to modeling error and noise, small nonzero entries associated with multiple other classes may exist. To tackle this challenge, the classification strategy in  is as follows to harness the subspace structure of for classification.
For each class , let be the characteristic function that selects the coefficients associated with the th class. That is, for any vector , is a new vector whose nonzero entries are only the entries in associated with class . Then, we can reconstruct the test sample as and recognize as the class that has the minimum residual between and ; that is,
In Figure 1, we present an example for sparse representation based classification method. In this illustration, we use the first 3 classes (SN_9563 from BMP2, SN_C71 from BTR70, and SN_132 from T72) from the MSTAR database. Detailed description of the MSTAR database is given in Section 4. The dictionary used in the experiment consists of the training data from the 3 classes. The sparse representation coefficient vector for 3 test images from each class is shown in Figure 1, recovered by solving (6) using the algorithm described in . As shown in Figure 1, for each test image, the recovered sparse coefficient vector has most of its nonzero elements concentrated at the ground-truth class, and the resulting residual error for the same class is minimum. Therefore, the class of the test image is determined by the sparse representation coefficient vector.
3. SRC along with Aspect Angle
In SAR images, even the same target presents different appearances with the variation of aspect angle. In this section, the aspect information is evaluated for the classification of vehicles in SAR image. Based on the analysis of the correlation of the test image with the train images of various aspects, the sparse representation vector is mapped onto a local aspect range and the algorithm of SRC along with aspect angle is proposed.
3.1. Correlation Analysis
The correlation between two images reflects the similarity of them. A higher correlation coefficient means the two target images are likely to come from the same class. Based on the correlation coefficient, the template matching method has been widely adopted in SAR ATR . By calculating the correlation of the test image with the training images of various aspects, we here evaluate the essentiality of introducing the aspect information for the vehicle classification in SAR images.
Given two images, the correlation coefficient is calculated as follows : in which is the test image, is the train image, and are the offsets on the direction of range and azimuth separately, and the shift aligns the target area in the image. The numerator in (8) is a convolution procedure, which can be achieved efficiently by multiplications in Fourier domain.
Figure 2 illustrates the correlation coefficients of 3 different test images with the train images from distinct classes. In general, the correlation coefficients of the test image with the same class of train samples are larger than with the other two classes. In particular, the test image presents high correlation with the train samples within a local aspect range, as indicated by the rectangle in Figure 2. Therefore, the aspect information is expected to be utilized in the SAR ATR.
3.2. Mapping the Sparse Representation Vector onto a Local Aspect Range
The vehicles in SAR images are aspect sensitive and the test sample is more likely represented by the train sample whose aspect angle is close to the test sample’s. The conclusion is preliminarily validated by the correlation coefficients in Figure 2. Moreover, we present a sparse representation vector that leads to incorrect result of classification in Figure 3. The ground truth class of the test sample is SN_C71. When the sparse coefficients on the whole aspect space are adopted, the reconstruction residual error is the least for the class of SN_9563 instead of SN_C71. If we concentrate the sparse coefficient vector on a local range of aspect that is around the aspect of the test sample, the resulting residual error of the class of SN_C71 is the minimum and an improved classification result is observed.
Motivated by the above observations and analysis, we propose the SRC method along with aspect angle. For each class , we redefine a characteristic function that selects the coefficients associated with the th class and a certain range of the test sample’s aspect angle . Then, similar with the original SRC, the class of the test sample is determined with the minimum residual:
It should be noticed that the aspect information can also be introduced to the SRC by other alternative ways, such as constructing the dictionary with the train samples of certain aspect or taking the aspect angle as one of the rows of the dictionary. However, the first one requires a large number of train samples of certain aspect to construct the overcomplete dictionary, and the second one is limited by the different dimensions of the aspect and other atoms in the dictionary. Therefore, we intuitively map the sparse coefficient vector onto a local range of aspect and calculate the residual error with the tailored sparse vector. The effectiveness of the proposed method will be further validated in Section 4.
3.3. Sparse Representation Based Classification with Aspect Angle
The proposed SAR vehicle classification method consists of three modules: preprocessing, including cropping the image size, principle component analysis (PCA) feature extraction, and estimating the aspect angle; sparse representation of the test sample with a constructed dictionary; and determining the class of the test sample based on the sparse representation vector and aspect angle. The overall procedure of the proposed sparse representation classification along with aspect angle (SRCA) method is summarized in Figure 4.
In the first module, the aspect angle of the vehicle in SAR image is estimated through image processing techniques. Firstly, the target area is separated from the background with segmentation methods . Then, the minimized enclose rectangle (MER) is calculated and therefore the aspect angle is estimated. In this procedure, there exists an uncertainty of 180° of the estimated aspect angle. We eliminate the ambiguity by assuming the tailstock section to be the border section with the highest mean RCS value. In the following two processing modules, the sparse representation vector is achieved by resolving (6), and the reconstruction residual error is calculated with (9).
4. Experiment Results
In this section, we evaluate the performance of the proposed method using MSTAR public database, which is a standard dataset for evaluating SAR ATR algorithms, and collected in 1995 and 1996 by the Sandia National Laboratory X-band (9.6 GHz) HH-polarization SAR sensor with the resolution of 0.3 m × 0.3 m. One subset of the MSTAR data consists of three classes of vehicles, that is, the BMP2, BTR70, and T72, with several configuration variations for each class. The vehicles are imaged in spotlight mode at 15° and 17° depression angles over 360° of aspect angles. The capacity of the subset is illustrated in Table 1. Since the original image dimension is very high (), we crop the data to and use the PCA method [12, 19] for feature extraction to reduce their dimensionality. Other feature extraction methods such as downsampling, Gaussian random projection , and Manifold learning  are also applicable to the SRC. For comparison, we compare with several state-of-the-art classification methods: linear SVM, kernel SVM (KSVM) with radial basis function (RBF) kernel, and SRC. For both linear SVM and KSVM, the LIBSVM package  is adopted. The radius for the RBF of KSVM is empirically set as , and the tolerance error in (6) is set as .
|Note: 1the samples corresponding to the numbers in brackets are not used in training or testing, unless notified.|
In the sequel, we carry out several experiments. Firstly, we evaluate the performance of the proposed method under different adopted range of aspect and feature dimensionalities. We then examine the robustness of the proposed method with respect to the variations of depression angle and target configurations. Finally, we evaluate the proposed algorithm under the condition of incomplete observation.
4.1. Performance on Different Adopted Range of Aspect and Dimensionalities
In this experiment, we use the first serial number targets from each class, that is, SN_9563 for BMP2, SN_C71 for BTR70, and SN_132 for T72, for algorithm evaluation and comparison. The training samples are captured at depression angle of 17° and the testing samples are captured at depression angle of 15°.
In our first experiment, we evaluate the recognition accuracy of the proposed SRCA method via different range of aspect for the different feature dimensions. The performance curves in Figure 5(a) illustrate that the recognition accuracy of the SRCA method varies with the adopted aspect range. At the beginning, the accuracy increases as fast as the adopted aspect range increases and retains a high level for several aspect range. However, when the aspect range keeps on increasing, the performance presents some degradation, which validates the effectiveness of carrying the classification on a certain range of aspect. When the feature dimension is , the proposed method achieves best performance under the condition that the aspect range is = 17°.
In the following experiment, we compare the performance of different algorithms when feature dimension changes. The corresponding results are summarized in Table 2 and a graphical plot is given in Figure 5(b) for visualization. As can be seen from Table 2 and Figure 5(b), the SRCA and SRC methods outperform the SVM methods by a notable margin. As the feature dimension increases, the proposed SRCA method achieves saturation faster than the other methods. Even when the performance of SRC represents little degradation, the performance of SRCA is still desirable. This once again verifies the effectiveness of introducing the aspect information for SAR ATR.
4.2. Depression Angle Invariance
For the real-world tasks, the invariance to depression angle is crucial to the successful application of a recognition algorithm. In this subsection, we evaluate the invariance to depression angle for the four algorithms. There are two different depression angles for the first 3 classes of MSTAR, that is, 17° and 15°. In the previous experiment, we have taken the samples captured on the depression angle of 17° for training and the samples captured on the depression angle of 15° for testing. In this experiment, we exchange the testing and training samples. As can be seen from Table 3, all the methods perform some degradation when the samples of 15° depression angle are used for training, which illustrates that the depression angle is important for the recognition task. However, the proposed SRCA method is still superior to the other methods.
4.3. Configuration Invariance
In this subsection, we examine the invariance of different algorithms under different configurations, which is a desirable property of an algorithm for SAR ATR applications. As shown in Table 1, the BMP2 and T72 both have different configurations of images captured from different variants of the same vehicle type. We compare the results of different algorithms according to the following settings: for training, the images from SN_9563 for BMP2, SN_C71 for BTR70, and SN_132 for T72 at the depression angle of 17° are used. For testing, the configuration of SN_C71 for BTR70 is used and 3 different testing sets for the other classes are used at the depression angle of 15°: invariant: SN_9563 for BMP2 and SN_132 for T72; mixed: all the images of BMP2 and T72 from all the 3 variants; variant: carrying out a test on SN_9566 and SN_C21 for BMP2 and SN_812 and SN_S7 for T72.The classification results are summarized in Table 4. For the invariant case, the proposed SRCA method achieves a highest classification rate of 99.83%, which is much better than all the other methods. When testing the dataset with different configurations (“variant”), the proposed method can still achieve a recognition rate of 87.37%. In particular, for the configuration variants of BMP2, the degradation is acceptable and is better than the other methods. The results in this subsection further validate the effectiveness of the proposed method.
4.4. Incomplete Observation Invariance
In the real-world tasks, the targets are not observed under all conditions, such as every aspect angles, radar frequencies, and grazing angles. The incomplete observation proposes challenges to the recognition algorithms. We evaluate the robustness of proposed SRCA method under the condition of incomplete observation. In this experiment, the training samples captured at the depression angle of 17° are selected randomly with a certain percentage to construct the training set, and the samples captured at the depression angle of 15° are tested. The performances of different methods are compared in Figure 6. The SRC based methods perform better than the SVM based methods by a notable margin. When the percentage is small, the absence of majority of training samples degrades the performance of SRCA. As the percentage increases, the performance of SRCA outperforms the SRC method and performs the best among the four methods. Another worthwhile point to note is that the performance of SRCA method improves with the increase of the adopted range of aspect angle, as shown in Figure 6 ( = 31° compared to = 17°).
In this paper, we propose a SAR vehicle recognition method based on sparse representation classification along with aspect angle. The method projects the sparse coefficient vector onto a subspace that is within a certain range of aspect angle around the estimated aspect angle of the test sample and then determines the class label according to the reconstruction residuals. The rationality of the idea lies in that the vehicles on SAR image are sensitive to its aspect angle and they are much more likely represented by the training samples with similar aspect angles. The proposed SRCA method is compared with the linear SVM, KSVM, and SRC methods by carrying extensive experiments on the MSTAR database. The results validate that the proposed SRCA method is robust to the variation of depression angles and target configurations, as well as the incomplete observation of training samples. Despite the effectiveness of the proposed method, much development needs to be further considered in the future work, including the learning of a more compact dictionary from the training data and the fast and effective solution of the sparse representation vector.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
This work is partially supported by the National Natural Science Foundation of China under Grant no. 61372163.
- J. Wright, A. Y. Yang, A. Ganesh, S. S. Sastry, and Y. Ma, “Robust face recognition via sparse representation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 31, no. 2, pp. 210–227, 2009.
- J.-B. Huang and M.-H. Yang, “Fast sparse representation with prototypes,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR '10), pp. 3618–3625, San Francisco, Calif, USA, June 2010.
- Y. Chen, N. M. Nasrabadi, and T. D. Tran, “Hyperspectral image classification using dictionary-based sparse representation,” IEEE Transactions on Geoscience and Remote Sensing, vol. 49, no. 10, pp. 3973–3985, 2011.
- H. Zhang, N. M. Nasrabadi, Y. Zhang, and T. S. Huang, “Multi-view automatic target recognition using joint sparse representation,” IEEE Transations on Aerospace and Electronic System, vol. 48, no. 3, pp. 2481–2497, 2012.
- C. Lin, B. Wang, X. Zhao, and M. Pang, “Optimizing kernel PCA using sparse representation-based classifier for MSTAR SAR image target recognition,” Mathematical Problems in Engineering, vol. 2013, Article ID 847062, 10 pages, 2013.
- X. Xing, K. Ji, H. Zou, W. Chen et al., “Ship classification in TerraSAR-X images with feature space based sparse representation,” IEEE Geoscience and Remote Sensing Letters, vol. 10, no. 6, pp. 1562–1566, 2013.
- “Moving and Stationary Target Acquisition and Recognition (MSTAR) Public Release Data,” https://www.sdms.afrl.af.mil/datasets/matar/.
- T. Ross, S. Worrell, V. Velten, J. Mossing, and M. Bryant, “Standard SAR ATR evaluation experiments using the MSTAR public release data set,” in Algorithms for Synthetic Aperture Radar Imagery V, Proceedings of SPIE, pp. 566–573, Orlando, Fla, USA, April 1998.
- G. Ravichandran and D. P. Casasent, “Minimum noise and correlation energy optical correlation filter,” Applied Optics, vol. 31, no. 11, pp. 1823–1833, 1992.
- A. M. P. Marinelli, L. M. Kaplan, and N. M. Nasrabadi, “SAR ATR using a modified learning vector quantization algorithm,” in Proceedings of the Algorithms for Synthetic Aperture Radar Imagery VI, Proceedins of SPIE, pp. 343–354, April 1999.
- L. G. Ramamoorthy and D. P. Casasent, “Classification and rejection of MSTAR data,” in Optical Pattern Recognition XV, Proceedings of SPIE, pp. 265–276, April 2004.
- A. K. Mishra and B. Mulgrew, “Radar signal classification using PCA-based features,” in Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP '06), Toulouse, France, May 2006.
- Y. Yang, Y. Qiu, and C. Lu, “Automatic target classification experiments on the MSTAR SAR images,” in Proceedings of the 6th International Conference on Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing and 1st ACIS International Workshop on Self-Assembling Wireless Networks, pp. 2–7, May 2005.
- R. G. Baraniuk, V. Cevher, M. F. Duarte, and C. Hegde, “Model-based compressive sensing,” IEEE Transactions on Information Theory, vol. 56, no. 4, pp. 1982–2001, 2010.
- E. Candes and J. Romberg, “L1-magic, Recovery of sparse signals via convex programming,” 2005, http://statweb.stanford.edu/~candes/l1magic/.
- P. L. Douville, “Measured and predicted synthetic aperture radar target comparison,” IEEE Transactions on Aerospace and Electronic Systems, vol. 38, no. 1, pp. 25–37, 2002.
- O. Guillaume, R. Loïcand, and M. Philippe, “Correlation and similarity measures for SAR image matching,” in SAR Image Analysis, Modeling, and Techniques VI, F. Posa, Ed., Proceedings of SPIE, Barcelona, Spain, 2004.
- L. M. Delves, R. Wilkinson, C. J. Oliver, and R. G. White, “Comparing the performance of SAR image segmentaion algorithms,” International Journal of Remote Sensing, vol. 13, no. 11, pp. 2121–2149, 1992.
- M. A. Turk and A. P. Pentland, “Face recognition using eigenfaces,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 586–591, June 1991.
- Z. Zhang, J. Wang, and H. Zha, “Adaptive manifold learning,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 34, no. 2, pp. 253–265, 2012.
- C. C. Chang and C. J. Lin, “LIBSVM: a library for support vector machines,” 2001, http://www.csie.ntu.edu.tw/~cjlin/libsvm.
Copyright © 2014 Xiangwei Xing et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.