Computational Intelligence and Neuroscience

Computational Intelligence and Neuroscience / 2017 / Article
Special Issue

Recent Developments in Deep Learning for Engineering Applications

View this Special Issue

Research Article | Open Access

Volume 2017 |Article ID 7186120 | 18 pages | https://doi.org/10.1155/2017/7186120

Automatic Target Recognition Strategy for Synthetic Aperture Radar Images Based on Combined Discrimination Trees

Academic Editor: Elio Masciari
Received13 Mar 2017
Revised21 Sep 2017
Accepted08 Oct 2017
Published29 Nov 2017

Abstract

A strategy is introduced for achieving high accuracy in synthetic aperture radar (SAR) automatic target recognition (ATR) tasks. Initially, a novel pose rectification process and an image normalization process are sequentially introduced to produce images with less variations prior to the feature processing stage. Then, feature sets that have a wealth of texture and edge information are extracted with the utilization of wavelet coefficients, where more effective and compact feature sets are acquired by reducing the redundancy and dimensionality of the extracted feature set. Finally, a group of discrimination trees are learned and combined into a final classifier in the framework of Real-AdaBoost. The proposed method is evaluated with the public release database for moving and stationary target acquisition and recognition (MSTAR). Several comparative studies are conducted to evaluate the effectiveness of the proposed algorithm. Experimental results show the distinctive superiority of the proposed method under both standard operating conditions (SOCs) and extended operating conditions (EOCs). Moreover, our additional tests suggest that good recognition accuracy can be achieved even with limited number of training images as long as these are captured with appropriately incremental sample step in target poses.

1. Introduction

Synthetic aperture radar (SAR) is a valuable technique for remote sensing and monitoring applications. Automatic target recognition (ATR) of SAR images is one of the most challenging SAR applications [1]. A typical SAR ATR system recognizes tactical ground targets of interests, that is, tanks, howitzers, and armoured vehicles, which is essential for identifying friends and foes and prerequisite for precision strikes.

SAR ATR involves a sequence of processes, such as some type of preprocessing, feature extraction, classifier construction, and finally target classification. The preprocessing stage may involve multiple types of processing that aims at facilitating the efficiency of image interpretation and analysis in the subsequent stages, for example, by suppressing the clutter reflections that obscure the contrast between the target of interest and the clutter. Moreover, SAR images are resized, shifted, and rotated to predefined standards. The so-called resizing is normally implemented by cropping out part of the image. The shifting and rotating processes are also known as image registration and pose rectification, respectively [2, 3].

Feature extraction is another essential stage which extracts effective discriminant features for improving recognition accuracy. Several features have already been exploited in SAR ATR [410]. Based on the consideration that tactical ground targets usually have a rectangular shape with different widths and lengths, geometric features are commonly used in SAR ATR. Zernike moments (ZMs) are employed in [6], taking advantage of their linear transformation invariance properties and robustness to the presence of noise. In [7], features are extracted based on pseudo-Zernike moments (pZm), which have merits such as the invariance properties, the independent property, and much lower sensitivity to noise in comparison with the ZMs. In [11], multiple geometric features are produced from calculating the axis projection of a target shape blob rotated clockwise with certain increment about the centre of the target. Then, the redundancy of the learned feature set is eliminated by keeping the rank of the covariance matrix of the new feature set the same as that of the entire data set. However, the geometric features of the target of interest in SAR images are difficult to measure precisely due to the cluttered background and variations in poses and depression angles. Therefore, the recognition accuracy is not guaranteed. The polar mapping method, which is frequently used in ISAR image classification, is modified and used in [3] to address the SAR ATR problem. The original images are converted from the original 2D spatial domain (range and cross-range) to images in the polar coordinate domain (radius and angle) to produce polar-mapped images. The polar-mapped images are similar to the images that are mapped from the same target even in different poses. For that reason, the commonly used pose estimator is not necessarily needed for polar-mapped images. However, the performance of the polar mapping method depends highly on the determination of the reference central point for coordinate transformation, which is not a simple task especially for SAR images captured under various clutter environments.

Certain features are not feasible to be directly applied to classification due to their high dimensionality [1219]. In [12], a compact representation feature, the monogenic signal, is employed for SAR ATR, where the high dimensional problem is circumvented by uniform downsampling, normalization, and concatenation of the monogenic components. Feature dimensionality reduction methods for SAR ATR based on manifold learning theory are also studied in recent years [1317]. In [16], each sample is given a weight, which is called the sample discriminant coefficient (SDC), relating to its similarity to neighbouring samples, and then the SDC is combined with the Local Discriminant Embedding (LDE) method for producing redundancy-reduced features. Similarly, in [17], the so-called neighbourhood geometric centre scaling embedding (NGCSE) method is proposed, where geometric centre scaling is introduced into the neighbourhoods such that the samples are provided with clear clustering directions. However, the performance of most of the nonlinear dimensionality reduction methods relies heavily on the parameter selection of the neighbourhood, which is still an open problem.

The nearest neighbour classifier is one of the most used classifiers, where the extracted features are directly fed into the classifier to achieve the classification results [16]. Sparse representation based classification (SRC) is recently developed and exploited in SAR ATR, where the feature vectors of the testing samples are coded as sparse linear combinations of the feature vectors of the training samples, and the target with the minimum residual energy is recognized [12, 20]. Methods such as Support Vector Machines (SVM), Neural Networks (NN), and adaptive boosting (AdaBoost) are all vastly exploited in SAR ATR [2, 5, 2123]. Various choices of base learners can be combined with the AdaBoost algorithm to solve the SAR ATR problem [2]. As explained in the Hughes phenomenon (also known as the curse of dimensionality), the difficulty of constructing classifier models becomes more prominent especially when the feature set is high in dimensionality while the number of the training data is limited (a fact in SAR ATR). However, the combination of the AdaBoost and graphical models is empirically proven in [24] to demonstrate good performance even when the training data is limited in number.

The SAR images are known for their indistinct appearances, variations in target appearances, and small number of available training samples. These problems must be properly addressed to achieve good recognition results for ATR tasks. To this end, a SAR ATR scheme is introduced as illustrated in Figure 1. Firstly, an initial processing stage is applied to facilitate the efficiency of feature extraction in the subsequent stages. More specifically, aiming at reducing the impact of variations in SAR images caused by variational echo energy and target poses, an image energy normalization process and a pose rectification process are applied sequentially. The construction of effective feature sets for ATR tasks is of crucial importance for achieving reliable recognition results. Therefore, it is suggested to extract a rich feature set that is formed by combining various types of discrimination features and then construct a more compact feature set by eliminating the redundancy of the rich feature set. We have decided to employ wavelet-based features. A rich feature set is firstly formed by combining the decomposed wavelet subband features, for example, the low-frequency information in LL subband coefficients and the high-frequency information in both LH and HL subband coefficients, where the HH subband is not involved since it is not stable feature in SAR images [25]. The involved coefficients actually depict the combination of texture features and horizontal and vertical edge features. After this, a compact low dimensional feature set which comprises features which retain most of the variance is constructed by employing the Principle Component Analysis (PCA) technique [26]. The relationship among features is statistically learned in a discriminative fashion rather than a generative fashion. Specifically, instead of using the true distribution, which is usually unknown for most of the time, the empirical estimates are learned in a discriminative fashion by maximizing the -divergence. Therefore, although the learned models may have low consistency with the real model of target classes due to limited amount of training data, high discrimination ability can still be achieved. Then, a final classifier is constructed by combining several discriminative tree based classifiers with the Real-AdaBoost framework [27]. To evaluate the performance of the proposed method, the moving and stationary target acquisition and recognition (MSTAR) public release data set is involved. Experimental results demonstrate that the proposed method outperforms several widely cited methods under both standard operating conditions (SOCs) and extended operating conditions (EOCs).

Variation reduction techniques that facilitate the efficiency of feature extraction are introduced in Section 2. The feature extraction and processing techniques are introduced in Section 3. The recognition scheme is detailed in Section 4. Experimental results using the MSTAR public database are shown in Section 5, followed by our conclusions in Section 6.

2. Variation Reduction Techniques

2.1. Image Energy Normalization

The echo strength of SAR is strongly affected by, for example, the range distance between the imaging target and its corresponding radar and several other reasons; therefore the average amplitude of image pixels in different image chips may be different even for the same target [28]. To mitigate the potential influence of amplitude variations in subsequent features extraction, the image energy normalization process needs to be applied. Let and denote the number of pixels in range and cross-range dimension for a given SAR image chip. The SAR image chip can be denoted as , where and are the dimension of range and cross-range, respectively. The energy normalized image pixel can be described aswhere and is the minimum and maximum value among all pixels of , respectively, and is calculated as The benefit of employing the image energy normalization process is provided in Section 5.1.

2.2. Pose Rectification

Pose rectification is beneficial for improving the accuracy of SAR ATR and can be achieved by rotating the given images according to the pose of target of interests. However, targets with partial defected contour shapes that are caused by the shadow effect may suffer from poor pose estimation accuracies. This section introduces a pose estimation method that is based on the exploration of targets’ geometrical information for achieving higher estimation accuracy.

Several methods have been proposed for achieving higher accuracy in pose estimation. The methods proposed in [29, 30] are based on maximizing the mutual information with multilayer perceptron (MLP). Although a low estimation error is achieved, these methods are computationally expensive and require a long training time. The method proposed in [31] is based on the 2D continuous wavelet transform (CWT), where the orientation that maximizes the angular energy is considered as the estimated pose. However, this method is based on the assumption that the target of interest is already placed in the image centre, which is difficult to achieve especially for SAR images with indistinct targets.

In fact, the tactical ground targets show rectangular shaped boundaries, which can be used for pose estimation. Therefore, methods based on the analysis of the geometrical information of target of interests have been proposed. The methods proposed in [2, 32] are based on finding the encapsulating box of the target of interest, where the basic assumption is that the edges of the estimated box should be tangent to the rectangular shaped target boundaries. However, this is not always true with incomplete target shape boundaries due to the shadow effects in SAR images. Moreover, the least squares linear fit based methods estimate the centreline of the target of interest, where the slope of the centreline is considered as the target pose. However, for similar reasons, the shadow effect in SAR may produce images with defected target, which can affect the corresponding pose estimation results. As discussed, the encapsulating box based methods have failed to achieve the optimum estimation result due to the defect targets in SAR images. However, as will be introduced, the Radon transform based method can achieve better estimation result in such scenarios [33]. Therefore, better estimation accuracy can be achieved by employing these two methods in a well-designed fashion. Firstly, the target of interest is segmented from the SAR image, and the rectangle that has the minimum perimeter around the segmented target is considered as the minimum bounding rectangle (MBR) [34]. Then, the completeness of the target of interest can be evaluated. In the case of a target with complete contour shape in the SAR image, the MBR estimated result is considered as the final result. Otherwise, the Radon transform is conducted and its estimation is used as the final result.

2.2.1. Estimation for Targets with Complete Contour Shapes

Tactical targets in SAR images have randomly distributed poses ranging from to (the target pose is defined as the angle between the target’s longer edge and the horizontal image axis). Tactical targets in SAR images show rectangular-like shapes. Figure 2 shows the segmented SAR chips, where the target poses can be estimated according to the inclination angle of its MBR. As introduced in [34], the rectangle that has the shortest perimeter enclosing a convex polygon has at least one side collinear with one of the convex edges. The MBR can be efficiently calculated as follows:Step 1. Estimate the centroid of the target of interest.Step 2. Compute the convex polygon of the target of interest.Step 3. Compute and store the edge orientations of the convex polygon.Step 4. Rotate a bounding rectangle according to the stored edge orientations until a full rotation is done.Step 4.1. Find a fitted rectangle.Step 4.2. Store the perimeter of the fitted rectangle.Step 4.3. Rotate the rectangle.Step 5. Return the rectangle corresponding to the minimum perimeter.

2.2.2. Estimation for Targets with Incomplete Contour Shapes

Due to the imaging principle of SAR, partial part of the target of interest is not radiated by radar beam, and therefore the imaged target shows incomplete boundary shape. However, the long edge of the target of interest is always well imaged, as shown in Figure 3. In fact, the Radon transform (RT) can be used for long edge detection. Therefore, for SAR images with targets that show incomplete contour shapes, the RT based estimation can achieve higher accuracy. The application of the RT on a target image limited by a set of angles can be considered as calculating the projection of the target along given angles. The calculated projection result is the sum of pixel numbers in each single direction, where a line can be found in the corresponding target image according to the peak of the projection result [33]. Define as the projection at angle with distant to the image centroid, and the RT is implemented as follows:where is the Dirac delta function. The parameters and determine the projection direction, where the projection is repeated from . Note that a pixel in the RT transform is divided into four subpixels such that accurate projection result can be achieved, where the projection contribution is calculated according to the position of the subpixel that hits the projection bin.

2.2.3. Degree of Overlapping Rectangle

In fact, for any given image, the completeness of the target in SAR images can be automatically calculated. As introduced in Section 2.2.1, the calculated MBR has at least one edge overlap with the target boundary. Therefore, in the case of a complete target, one long edge of the target of interest will overlap with that of its corresponding MBR. In the case of a target with partial defect, the diagonal line of the target of interest may overlap with a long edge of its corresponding MBR with few pixels. Let denote number of pixels of the two MBR long edges, and let denote the number of target pixels that overlap with the two MBR long edges. The completeness of the target in SAR images can be evaluated as follows: the target is firstly dilated, and then the degree of overlapping rectangle is calculated as , and finally the completeness of the target boundary is evaluated according to the calculated degree of overlapping rectangle. After dilation, since the difference between the complete contour shape and defected contour shape is large, the proposed method is not sensitive to the selected threshold employed for evaluating the degree of overlapping. Overall, as shown in Figure 4, the target pose is estimated using the MBR based method or the RT based method depending on the evaluation result of the degree of overlapping rectangle, and several estimation results are shown in Figure 5.

3. Feature Extraction and Processing Techniques

3.1. Rich Feature Set Extraction

Feature extraction is of crucial importance to the overall performance of the entire ATR system. It is ideally preferable to extract features that have characteristics of high discrimination ability (or, in other words, high interclass variation) and high tolerance to target translation. These feature characteristics can be achieved by efficiently employing the wavelet decomposition technique. As depicted in Figure 6, the texture features are reflected in LL and the horizontal and vertical edge feature are reflected in LH and HL, respectively. HH is actually a combination of features reflected in LH and HL. Furthermore, the translation invariant features can be extracted by sequentially further decomposing the previously decomposed image to a much coarser resolution. The idea behind the translation invariant features is that each decomposition process throws away the exact positional information of certain feature that exists in a specific area. More specifically, as illustrated in Figure 7, a pixel point in a newly decomposed image implicitly reflects the presence of certain feature(s) in a corresponding entire local region in the original image.

Several wavelet families have been proposed with the shape and duration of the mother wavelets being the main differences among them. The number of vanishing moments (order number) is used as an indication of the wavelets’ smoothness and the frequency response flatness of the wavelet filters. It is suggested that we employ one fixed mother wavelet for the entire recognition scheme. A wavelets’ comparison test is conducted in [5], where 7 mother wavelets with variations in order numbers are compared, according to minimum distance. To determine the most appropriate mother wavelet, in this paper, we compare 7 mother wavelets with more variations in order numbers with the maximum margin criterion (MMC) [35]. Specifically, we compare discrete Meyer wavelet, Biorthogonal wavelets (orders 1.1, 1.3, 1.5, 2.2, 2.4, 2.6, 2.8, 3.1, 3.3, 3.5, 3.7, 3.9, 4.4, 5.5, and 6.8), Coiflets (orders 1, 2, 3, 4, and 5), Haar wavelet, Daubechies wavelets (orders 2, 3, 4, 7, 10, 25, and 45), Reverse biorthogonal wavelets (orders 1.1, 1.3, 1.5, 2.2, 2.4, 2.6, 2.8, 3.1, 3.5, 3.7, 3.9, 4.4, 5.5, and 6.8), and Symlets (orders 2, 4, 8, and 16).

MMC finds the mother wavelet that maximizes the average margin between classes. This is achieved by comparing the difference between the average within-class distance and the average between-class distance . The mother wavelet that achieves the maximum difference is the best selection. Suppose we have classes , each class with samples and therefore, samples in total. Let denote the th sample in the th class, let be the centroid of the th class, and let be the centroid of the training set. The average within-class distance and the average between-class distance can be denoted as

The comparison of the discrimination performance of the mentioned wavelets is illustrated in Figure 8. It is noted that the Reverse biorthogonal wavelet 3.1 achieves the highest value, an observation which indicates that it has the highest discrimination ability among these wavelets. Therefore, the Reverse biorthogonal wavelet 3.1 is selected as the default mother wavelet for feature extraction in SAR images.

The above process yields large sets of features which exhibit a high variability as far as the quality of the discriminative information that they convey is concerned. To achieve the truly effective features, the PCA is used, which achieves comparable result to both 2D-PCA and two-stage 2D-PCA when they are employed for SAR feature compression purposes, as analysed in [26]. Moreover, the PCA is much more efficient as far as both computation time and storage space are concerned. The implementation of the PCA is introduced as follows:Given. Data . Number of principal components .Step 1. Subtract the mean of variables from .Step 2. Solve the Singular Value Decomposition (SVD) of .Step 3. The dimensionality reduced feature set is calculated with the first column of as .

3.2. Learn Statistical Relationship among Features

Since access to data arising from true distributions is often not available, the learned models based separately on positive/negative samples are usually not accurate enough for classification. In fact, the discriminative methods construct models from both the positively and negatively labelled samples in a discriminative fashion. Since the final objective is classification, even if the learned distributions may not converge to the true distributions, the constructed discriminative models tend to have better discrimination performance than the generative models [36, 37].

In binary classification case, which can be naturally extended to the more general -ary classification case, for a given labelled training set , where represents the sample label, each pair ( is normally a finite set of integer values as ). Supposing we have two models and that can describe the true distribution of and , the log-likelihood ratio test is known to be the optimal test (under both the Neyman-Pearson and Bayesian settings [38])where is the threshold [38].

In most cases, it is impossible to have access to the true conditional distributions and . Approximations and are normally built to learn the unknown distribution from the labelled training set . Therefore, the log-likelihood ratio test can be rewritten as

The recently proposed method named discriminative tree estimates the multivariate distributions and jointly from both the positively and negatively labelled samples in the training set of Tan et al. [37]. This method is based on the assumption that the learned distribution is Markov with respect to an undirected graph , where represents the vertex set and represents the set of all unordered pairs of vertexes. The mentioned Markov conforms to the local Markov property where represents the set of neighbour nodes of and for any set .

A tree structured distribution that is Markov with respect to an undirected graph can be factorized as follows [39]:where represents the marginal of the random variable and represents the pairwise marginal of the pair .

Based on this, for a given distribution , the projection of onto some tree distribution is defined as follows:

We digress here to introduce the method for constructing models in generative fashion and then provide the method for constructing models in discriminative fashion. The generative methods attempt to construct a model that is the same as the underling model of the classification target. The widely researched generative method, namely, the Chow-Liu algorithm [40], employs the KL-divergence as the measure of the differences between two probability distributions and . The optimization in the Chow-Liu algorithm is therefore defined as where states that is a tree structured distribution over the same alphabet as . It is shown by Chow and Liu that this optimization problem can be solved by using a maximum weight spanning tree (MWST) algorithm (e.g., Kruskal’s [41]) where the mutual information is used to represent the edge weights between pairs of variables.

In contrast, the recently proposed discriminative method employs the -divergence as the measure of the separation between two probability distributions and . The -divergence is defined as follows [42]:

The optimization problem reduces to two tractable MWST problems for maximizing the tree approximate -divergence over the two tree structured-distributions and for known empirical distributions and , which is defined as where

It is noted that, as described in [37], (13) can be decoupled into two independent optimization problems:

These can be solved by the MWST algorithm

Overall, the procedure of the learning of the discriminative tree is summarized in the following steps [37]:Given. Training set .Step 1. Estimate the pairwise statistics and for all edges .Step 2. Calculate edge weights and for all edges .Step 3. Find the optimal tree structures with the given edge weights.Step 4. Set and to be the projection of onto and onto , respectively.Step 5. Classify the test sample using the learned distributions and in a likelihood ratio test .

Since the classification result is finally determined by the numerical result of the log-likelihood ratio test, we choose to employ one fixed threshold 0 for the entire training process. This is because likelihoods larger than 0 indicate higher probability of belonging to . Similarly, likelihoods smaller than 0 indicate high probability of belonging to .

4. Recognition Scheme

The main aim of classifier construction in ATR is to convert a wealth of training data into useful knowledge for classification by learning. However, a classifier learned from massive amounts of high varying data is not guaranteed to achieve good performance in classification and may yield large feature dimensions. Therefore, it is of great importance to find effective representations for the targets of interest to be used for constructing the classifiers.

Extracted features might comprise large sets of features which at a glance might be worth of exploiting but turn out to be too “messy” and high in redundancy. In fact, the learning process of the classifiers could be enormously benefited from a feature dimensionality reduction process after the acquisition of the extracted features as previously discussed. The redundancy-reduced features can be used for learning classifiers, where efficient classifiers and better classification accuracy results can be achieved. Therefore, it is suggested to enlarge the quantity of the potential features but then eliminate the existing redundancy, reduce the dimensionality of the enlarged feature set, and finally exploit the preserved features for classification, which comprise the characteristics of proper combination of both quality and quantity. In the proposed recognition scheme, features are extracted with wavelet decomposition, but then the dimension of the feature set is reduced to provide a feature set rich in discriminative information but with limited dimensionality and less redundancy. To make the most of the extracted features, tree structured classifiers are learned in discriminative fashion based on the statistical information provided by the training data of the target classes. In the learned classifiers, the feature nodes are connected as a spanning tree, where each node is connected to another node which has the maximum relevance. Moreover, the relevance between feature nodes can be accordingly calculated. Finally, classifiers are combined using the Real-AdaBoost algorithm to construct the final classifier that has high classification accuracy and is less prone to overfitting, where the recently proposed discriminative trees are involved as the base classifiers. A generic sequence of steps of the proposed scheme is illustrated in Figure 9.

4.1. Construct a Strong Classifier

Efforts have been constantly made to construct a classifier with high classification accuracy and strong generalization ability (the later meaning that performance of the classifier learned from a given training dataset will still be good when the classifier is exposed to unseen data) [43]. Employing ensemble learning methods is one of the solutions. Ensemble learning methods construct and combine a set of base classifiers instead of constructing and using one single classifier learned from the training dataset. Base classifiers can be generated from a training dataset with the use of any learning algorithm (e.g., decision tree, graphical models, and neural networks).

AdaBoost is one of the ensemble methods that have achieved great success in diverse domains [27, 4347]. The general idea of the AdaBoost is to constantly update the distribution of the training data such that the learning of the base classifiers in each iteration focuses more on the wrongly labelled samples by the previous learned base classifiers. Real-Adaboost is a variant of the AdaBoost which has been empirically proved to have better performance than ordinary AdaBoost (Discrete-AdaBoost) [27, 37, 44]. Specifically, for a given training dataset , each sample is assigned with an initial weight , where is the number of training samples. A base classifier is learned in each iteration such that , where a larger absolute value in indicates higher confidence. Then, the samples wrongly labelled by are increased in weights such that the constructed classifiers in the following iterations can focus on the misclassified samples. Finally, the combined classifier resulting after iterations iswhere is the sign function that if and otherwise and is the coefficient calculated in each iteration for minimizing the weighted training error. Overall, the Real-AdaBoost algorithm trains a set of base classifiers sequentially and combines them to a strong classifier, where the current learned base classifiers focus more on the wrongly labelled samples by the previous base classifiers.

The ensemble process of the Real-AdaBoost is iterated with the rearrangement of the training set distribution while the learning method of the base classifiers is not changed. For the learning of the base classifier in each iteration , the group of redundancy and dimensionality reduced wavelet features are employed and fed to learn the discriminative trees for classifier construction. By employing the learning method introduced in Section 3.2, a pair of discriminative trees is constructed to provide an estimation of the classification result. Specifically, the pair of discriminative trees constitutes a base classifier for the Real-AdaBoost , where , and and denotes the learned discriminative tree models at the th iteration of the Real-AdaBoost. After iterations, pairs of discriminative trees are learned and combined to construct a stronger classifier with better approximation of the classification result which can be written as Viola and Jones [45]where and .

For the iterative updating of the training set distribution, the misclassified samples are reassigned with larger weights and the correctly classified samples are reassigned with smaller weights compared to their previous weights. Regarding the weight distribution updating problem, simply reduplicating the samples with higher weights is time and computation inefficient. This is because as the number of iterations increases, the wrongly labelled samples would be much less in number but have much larger weights. Therefore, the final training set is fixed in size and constructed in random sampling fashion, where samples of the original training set are chosen according to the updated distribution weights. The entire classifier construction scheme is summarized as below:Given. Training dataset . Number of iterations .Step 1. Wavelet feature extraction from the given training dataset .Step 2. Redundancy and dimensionality reduction for the extracted features.Step 3. Initialization of the distribution weights, for all .Step 4. Classifier construction(1) for do(2) Learn the pair of discriminative trees , from the weighted empirical distributions and .(3) Get the base classifier .(4) Calculate the coefficient (5) Update the weighted empirical distribution:where is the normalization factor (to ensure that will be a distribution).(6) end forStep 5. Output the final classifier with coefficients .

4.2. Multiclass Classification

The One-vs.-One (OvO) and One-vs.-All (OvA) are the two most popular strategies for the extension of a two-class classification (binary classification) problem to a multiclass classification (multinomial classification) [48]. For a class problem, the OvO strategy trains binary classifiers, each of which classifies a pair of classes selected from the original training set. For the classification of the unseen samples, the samples are fed and tested in all classifiers by employing a voting scheme where the class which achieves the highest number of positive predictions would be considered as the final prediction. The OvA strategy trains one classifier for every class where the samples of the target class are considered as positive samples and all of the rest of the samples as negative samples. At predication stage, the unseen sample is assigned with the label of class if its corresponding classifier produces the highest likelihood score.

5. Experimental Results

In this section, the performance of the proposed scheme is evaluated and compared with several established methods. The widely used SAR ATR experimental validation and comparison benchmark moving and stationary target acquisition and recognition (MSTAR) public release database is employed for performance evaluation [4951]. The MSTAR database consists of 10 vehicle classes, which are collected by X-band SAR with 1-ft by 1-ft resolution, including BMP2, BTR70, T72, BTR60, 2S1, BRDM2, D7, T62, ZILI131, ZSU234, and SLICY. The collection of target images is captured under various depression angles and aspect angles, which are suitable for testing the SAR ATR methods with targets under various operating conditions.

There are two categories of operating conditions in the MSTAR database: the standard operating conditions (SOCs) and the extended operating conditions (EOCs) [50]. The targets captured under SOCs are listed in Table 1 including information about vehicle types, number of chip images, serial numbers, and depression angles. It is worth noting that the EOCs are much more difficult for SAR ATR than the SOCs. In EOC-1, the depression angles are larger in variation where the training images are captured under 15° and the testing images are captured under 30°, as shown in Table 2. In EOC-2, the training and testing set have various versions of T72 with different serial numbers, as shown in Table 3.


Training set
VehicleNumber of imagesSerial numberDepression angle

BMP26999563, 9566, C2117°
BTR70233C7117°
T72699132, 812, S717°
BTR60256k10yt753217°
2S1299B0117°
BRDM2299E7117°
D729992v1301517°
T62299A5117°
ZILI131299E1217°
ZSU234299D0817°

Testing set
VehicleNumber of imagesSerial numberDepression angle

BMP25879563, 9566, C2115°
BTR70196C7115°
T72588132, 812, S715°
BTR60196k10yt753215°
2S1274B0115°
BRDM2274E7115°
D727492v1301515°
T62274A5115°
ZILI131274E1215°
ZSU234274D0815°


Training set
VehicleNumber of imagesSerial numberDepression angle

2S1274B0115°
BRDM2274E7115°
ZSU234274D0815°
T72274A6415°

Testing set
VehicleNumber of imagesSerial numberDepression angle

2S1288B0130°
BRDM2288E7130°
ZSU234288D0830°
T72288A6430°


Training set
VehicleNumber of imagesSerial number Depression angle

BMP2233C2117°
BRDM2298E7117°
BTR70233C7117°
T7223313217°

Testing set
VehicleNumber of imagesSerial numberDepression angles

T72419S715° and 17°
T72572A3215° and 17°
T72573A6215° and 17°
T72573A6315° and 17°
T72573A6415° and 17°

We experiment with both two-level and three-level two-dimensional wavelet decomposition with respect to the Reverse biorthogonal wavelet (the selected mother wavelet as introduced in Section 3.1). In the following, wavelet 768 () and wavelet 192 () are used to denote the two-level and three-level wavelet decomposition, respectively. The stopping criterion of the Real-AdaBoost is set to 400 iterations. The segmentation of the target of interest is implemented with the MRF model based method, where the potential class number is 2, the expectation is 0.4, and the maximum iteration number is 50. The segmented target is dilated with a disk-shaped template with radius 3. The degree of overlapping rectangle is 0.5 indicating that an appropriate MBR must have more than 50% long edge overlapping pixels in terms of the target of interest. Moreover, the proposed method is implemented using Matlab R2013a and tested on a computer with 1.8 GHz CPU and 4 GB RAM. Regarding the computation complexity, for a classifier trained for classifying 10 targets in OvO fashion, the processing time for one single sample takes less than 0.02 s, including the processes of extraction and compressing of features and recognition of targets.

Before applying the proposed method to SAR ATR and comparing with other methods, it is necessary to test the proposed method in conjunction with several important processes, including image energy normalization, feature extraction, extension of two-class to multiclass classification, and pose rectification. These four tests are conducted in Sections 5.1 to 5.4, and the comparisons of recognition accuracy performance with other methods are provided in Section 5.5.

5.1. Image Energy Normalization

The significance of image energy normalization in SAR ATR is tested in this section, where the performance of the proposed scheme is tested with or without image energy normalization processing. The dataset includes all 10 classes captured under SOCs as listed in Table 1. The wavelet 192 is used for feature extraction.

It is noticed in Figure 10 that the involvement of normalization before feature extraction is beneficial for improving classification accuracy. In fact, as the dimension of feature vectors employed for classification grows, the advantage of image energy normalization diminishes. This is because a larger training feature set provides more information for classification, where the classifier is empowered with more discrimination ability by exploiting the provided information. However, the classification with normalization achieves good classification accuracy (around 96%) even when the feature vector dimension is much lower, yielding an accuracy which is almost the same as the accuracy achieved with higher feature dimensions. Therefore, it is still suggested to employ image energy normalization for preprocessing, especially for classifiers constructed from training feature sets of lower dimensionality. In the following, the image energy normalization process is employed as a standard default processing step.

5.2. Extension to Multiclass

We compare the OvO and OvA strategies on the same training set (all 10 classes under SOCs) to test their performance on the SAR ATR problem. It is noted in Figure 11 that the OvO strategy appears to be outperforming the OvA strategy marginally in the SAR ATR problem. The marginal differences in recognition accuracy lie in the unbalance of the training set, where the OvA strategy employs the positive sample classes that are much less in quantity than the negative sample classes. In fact, the advantage of the OvA strategy is that it is less in computation and time complexity, where the OvO constructs 45 classifiers and the OvA constructs 10 classifiers for a 10-class problem, respectively. Since the aim of this paper is to provide a SAR ATR scheme with high recognition accuracy, OvO is employed as the default strategy for solving the multiclass problem.

5.3. Feature Extraction

In this section, we compare the performance of feature extraction using the wavelet 192 (three-level wavelet decomposition) and the wavelet 768 (two-level wavelet decomposition). All 10 classes captured under SOCs are employed for both training and testing. As illustrated in Figure 12, these two curves coupled with each other. The wavelet 192 outperforms the wavelet 768 when feature vectors possess lower dimensions. However, this situation changes as the dimension of feature vectors grows to 40. Moreover, the best classification result (97.46%) is achieved by the use of wavelet 768 with feature dimension of 70. This is because the wavelet 768 provides more features for discrimination and the proposed ATR scheme constructs and combines several discriminative tree classifiers that make the most of the discriminative information inherent to the extracted features.

5.4. Pose Rectification
5.4.1. Pose Estimation

To test the performance of the proposed pose estimation method, the estimation results of the proposed methods are compared to the ground truth of target poses (the azimuth information provided in the MSTAR database). The correctness of the estimation results is evaluated with the so-called mean absolute difference (MAD), which is calculated as . The MAD reveals the actual estimation error about the ground truth in comparison to the mean error (ME), since it prevents the offset of the positive and negative errors. Moreover, the performance of the proposed method is evaluated and compared with several widely cited methods, such as the least square method (LSM) based estimation, the Hough transform (HT) based method, the MBR based method, and the Radon transform (RT) based method.

All 10 targets captured with different depression angles and target poses are involved to test the robustness of the proposed method over depression angle variations. The evaluation results for the 10 targets at depression angles 17° and 15° are listed in Tables 4 and 5, respectively. All serial number variants of BMP2 and T72 in MSTAR dataset are involved to test the robustness over variation in serial numbers. The evaluation results of the data captured at depression angles 17° and 15° are listed in Tables 6 and 7, respectively. It is noted that the MBR based method has achieved much lower estimation error in comparison to other methods. However, the performance of the MBR based method has estimation error higher than 10° in several tests. More specifically, the MBR based method achieves the highest estimation error 15.32° for the BRDM2 captured at depression angle of 15°. In comparison with these methods, the proposed method achieves the lowest estimation error in all tests (lower than 10°).


Vehicle BTR602S1BRDM2D7T62ZIL131ZSU234BMP2BTR70T72

LSM7.0511.0515.4811.9310.4710.7915.429.399.3310.22
HT9.1711.2612.1418.8614.6912.3415.1610.728.0912.64
MBR6.107.4514.666.277.1811.878.236.799.384.99
RT7.239.5510.3819.3213.329.2718.409.547.4913.73
Proposed method4.855.848.706.279.398.027.835.436.674.51


VehicleBTR602S1BRDM2D7T62ZIL131ZSU234BMP2BTR70T72

LSM5.5210.6315.2110.099.208.6714.077.567.006.87
HT7.0010.1610.9616.7213.4910.1314.329.026.8710.38
MBR5.116.6415.326.406.9310.238.815.127.013.53
RT4.769.639.3817.3612.347.0817.318.275.8510.61
Proposed method3.815.429.046.326.105.897.674.455.133.38


Vehicle typeBMP2T72
132812s795639566c21

LSM17.159.5813.1213.5313.2212.06
HT18.8714.8116.2312.9613.0413.76
MBR7.975.506.4011.909.968.71
RT21.4615.1517.6211.4713.6512.25
Proposed method7.935.425.798.477.946.96


Vehicle typeBMP2T72
132812s795639566c21

LSM12.589.3010.7411.2610.969.71
HT15.9812.3613.329.449.6311.58
MBR7.754.854.538.576.936.57
RT17.1611.3113.629.3810.3110.61
Proposed method7.214.604.346.965.635.71

Furthermore, the average estimation error of the above tests is illustrated as bar figure in Figures 13 and 14, such that a much more distinct comparison can be observed. Similarly, the proposed method is compared to the least square method (LSM) based estimation, the HT based method, the MBR based method, and the RT based method. It is noted that, for most of these methods, higher estimation error is achieved in the serial number variation test. Compared with these methods, the proposed method achieves robust and accurate estimation results in both tests. Specifically, the proposed method achieves the lowest average estimation error in all tests (lower than 8°).

5.4.2. Pose Rectification

The various target poses introduce great variations into the SAR images. It has been experimentally proven in several researches that rotating images in certain directions or introducing rotationally invariant features is beneficial for improving classification accuracy [2, 3]. To this end, we rotate the image according to the target poses in the SAR images, which is named as pose rectification. In this section, we test the performance of the proposed scheme with or without pose rectification using the same training and testing set (all 10 classes under SOC). The SAR images are rotated anticlockwise according to their poses. As can be seen from Figure 15, the classification with pose rectification universally outperforms the classification without pose rectification. It is also noted that the best classification accuracy (99.3%) is achieved by the wavelet 768 with feature dimension of 75. These results meet our expectation that the classification can benefit from eliminating the pose variations in SAR images. Specifically, the rectification of poses provides target images for classification with fewer variations.

5.4.3. Outlier Rejection Performance

To evaluate the outlier rejection performance of the proposed method, a varying threshold for the log-likelihood test, which is introduced in Section 3.2, was incorporated to provide the ROC curve. BTR70, BMP2, and T72 listed in Table 1 are involved for classifier training and the SLICY set is involved as confusers with 1168 image chips, that is, 210 chips captured at 15°, 298 chips captured at 16°, 386 chips captured at 17°, and 274 chips captured at 29°. As shown in Figure 16, at the probability of detection , the probability of false alarm for feature dimension of 75 is . It is clear that the proposed method is robust at rejecting confuser targets.

5.5. ATR Performance Comparisons

The effectiveness of the proposed ATR scheme is tested in this section. Several widely cited methods are involved for performance comparison, for example, the Extended Maximum Average Correlation Height Filter (EMACH) [53], the Support Vector Machine (SVM) classifier with Gaussian kernel [52], feature fusion via AdaBoost with neural networks as the base classifiers [2], and the Iterative Graph Thickening (IGT) approach [24]. The best result obtained from the proposed method is used to compare with other methods.

Table 8 lists the performance comparison of the mentioned methods under SOCs. It is noted that the proposed method has achieved significant improvements in classification accuracy in comparison with other methods. A majority of the classes are correctly classified with 100% accuracy and the rest have higher than 98%, which is also much higher than other methods. Moreover, the superiority of the proposed method is strengthened by the fact that the average (99.3%) is much higher than the second highest average (84.8%).


Confusion matrix of EMACH, the method proposed in [52], and the proposed method
VehicleBMP2BTR70T72BTR602S1BRDM2D7T62ZILI131ZSU234

BMP290/90/1002/2/04/3/01/1/01/1/00/2/00/0/01/0/00/1/01/0/0
BTR702/3/093/90/1001/3/00/0/01/0/01/2/00/0/02/0/00/0/00/2/0
T722/2/00/1/096/93/1000/3/01/0/00/0/00/0/00/0/01/1/00/0/0
BTR600/2/01/2/00/1/295/92/981/0/00/0/03/3/00/0/00/0/00/0/0
2S15/5/06/3/04/2/12/0/074/81/993/3/01/2/02/3/01/0/02/1/0
BRDM23/6/16/8/03/2/00/1/01/0/084/79/992/0/00/3/00/0/01/1/0
D72/0/13/0/02/0/01/0/00/1/00/0/085/98/993/0/02/0/02/1/0
T621/1/01/0/01/0/01/1/04/0/10/0/00/0/086/91/994/4/02/3/0
ZILI1312/2/00/1/01/0/02/0/00/0/00/0/00/0/04/0/088/95/1003/2/0
ZSU2341/0/00/1/04/0/12/3/00/0/00/0/00/1/01/0/00/3/092/92/99

Confusion matrix of the method proposed in [2], IGT, and the proposed method
VehicleBMP2BTR70T72BTR602S1BRDM2D7T62ZILI131ZSU234

BMP292/95/1002/1/02/1/00/0/01/1/02/1/00/0/00/0/01/1/00/0/0
BTR703/2/093/94/1000/0/00/0/00/00/2/2/00/0/00/0/00/0/02/2/0
T722/2/01/1/096/96/1001/1/00/0/00/0/00/0/00/0/00/0/00/0/0
BTR602/1/00/0/02/1/293/97/980/0/00/0/03/1/00/0/00/0/00/0/0
2S13/3/04/4/01/1/10/0/087/89/992/0/00/0/02/1/00/0/01/2/0
BRDM25/2/13/1/02/4/00/0/00/0/085/90/995/2/00/0/00/0/00/1/0
D70/0/10/0/00/0/00/0/01/1/00/0/098/99/990/0/00/0/01/0/0
T621/1/00/0/00/0/00/0/00/0/10/0/00/0/093/95/993/3/03/1/0
ZILI1312/2/02/2/00/1/00/0/00/0/00/0/00/0/00/0/094/95/1002/0/0
ZSU2341/1/00/0/00/0/11/1/00/0/00/0/00/0/00/0/02/2/096/96/99

The average of these 5 methods are sequentially 88.3%, 90.1%, 92.7%, 94.6%, and 99.3%.

Four distinct target classes are involved in the following test: EOC-1, including 2S1, BRDM2, T72, and ZSU234, as listed in Table 2. All of these four classes are involved in training and testing stages. The only difference is that the training and testing set are captured under depression angles 15° and 30°, respectively. The increase in depression angle variations introduces a bigger challenge to the classification problem. It is noted in Table 9 that the classification accuracy of the most of the mentioned methods is lower than 88% under EOC-1, where the superiority of the proposed method (higher than 96%) is obvious. Furthermore, the average classification accuracy of the proposed method is 97.5% which is much higher than the other listed methods.


Vehicle 2S1 BRDM2 T72 ZSU234

2S167/74/77/78/9915/8/5/6/012/9/11/9/16/9/7/7/0
BRDM217/12/15/15/257/66/73/76/9719/9/5/6/17/13/7/3/0
T727/17/11/10/09/6/9/9/066/73/75/78/9618/4/5/3/4
ZSU23410/7/4/5/17/5/6/5/02/3/2/2/181/85/88/88/98

The average of these 5 methods are sequentially 67.75%, 74.5%, 78.25%, 80.0%, and 97.5%.

The training dataset under EOC-2 is composed of four different target classes, BMP2, BRDM2, BTR70, and T72, as summarized in Table 2. This test aims at testing the performance of the SAR ATR algorithms with significant different in serial numbers and configurations. The testing set has only the T72 family with five different serial numbers and the training set is composed of all these four mentioned classes. In addition, the training set was obtained at 17° while the testing set was obtained at depression angles of both 15° and 17° as shown in Table 3. Table 10 lists the performance comparison of the mentioned methods under EOC-2. The improvement in classification accuracy is substantial since the average of the proposed method is 96.9% which is much higher than the second highest 84.8%.


VehicleBMP2BRDM2 BTR70T72

T72_S74/5/4/5/08/4/6/4/26/4/2/3/182/87/88/88/97
T72_A329/7/5/6/05/5/8/3/05/2/3/2/081/86/84/89/99
T72_A628/7/6/5/06/5/5/4/13/6/4/4/383/84/85/87/97
T72_A6313/15/11/7/06/6/9/5/111/3/4/7/270/76/76/81/97
T72_A6416/9/10/11/04/5/5/4/212/13/9/6/368/73/76/79/94

The average of these 5 methods are sequentially 76.8%, 81.2%, 81.8%, 84.8%, and 96.9%.
5.6. Performance Comparison of Variations in Target Poses

As analysed in Section 5.4, the involvement of pose rectification is beneficial for improving the classification performance. In fact, a single target will exhibit different appearances when it is captured under various poses. In this section, we conduct an experiment to test the influence of the appearance differences introduced by the pose variations. The experimental database is almost the same as the data listed in Table 1 except that only one single serial number of each target is involved for training and testing (C21 for BMP2 and S7 for T72). The images for training are selected from the training database with different sample steps of target poses (varied from 1° to 7°), where 51 images are selected for the training of each target. For example, the poses for Step are , the poses for Step are , and the poses for Step are . Additionally, we have also investigated the possibility of training classifiers using training databases with different sizes, for example, 51 training images, 60 training images, 71 training images, and 85 training images. Features are extracted with wavelet 768 and reduced to dimensionality of 55. The results are illustrated in Figure 17.

It is noted in Figure 17 that as the incremental step of poses increases, the achieved classification accuracy grows too. More specifically, a much higher of is achieved by employing 51 training images with incremental Step in pose, in comparison with a of achieved by employing 85 training images with incremental Step in pose. The principle behind this observation is that the training datasets formed with small pose variation steps can provide less target signal information and thus, their content is not sufficient enough to cover the different appearances of the targets captured under various poses. In contrast, a much more complete training dataset can be formed when the involved images are captured with larger pose variations. The experimental results in Figure 17 show that the best classification performance 90.3% is achieved when training with 51 images captured using incremental Step and 93.0% is achieved when training with 85 images captured using incremental Step . Moreover, it is quite straightforward to find that better classification performance is always achieved when training with relatively larger number of training images. It is worth pointing out that only a small number of images are involved in the training stage rather than several hundreds of them as used in the previous tests. This is a promising result which implies that a good classification result can be achieved even with much less number of training images, as long as they are captured with appropriate incremental step.

6. Conclusion

In this paper, we presented a systematic scheme for the SAR ATR task. The proposed scheme involves three main stages: preprocessing, feature extraction and processing, and classifier construction. The effectiveness of involving several preprocessing approaches (e.g., the image energy normalization and the pose rectification processes) is analysed and empirically verified. The results suggest that the involvement of these preprocessing steps is beneficial for improving the classification accuracy. Moreover, we proposed to expand the feature set to provide more information for discrimination and then eliminate the redundancy and dimensionality of the extended feature set to form a more compact and efficient feature set. Finally, the discriminative trees are learned as the base classifiers and combined to construct a strong classifier by using the Real-AdaBoost algorithm. The proposed method is evaluated with the MSTAR dataset under various operating conditions. Experimental results demonstrate that the proposed method outperforms traditional methods, for example, EMACH, SVM, NN, and IGT. The advantages of the proposed method give credit to the reduction of variations in target images, the improvement of feature efficiency, the elimination of redundancy in feature sets, and the excellent generalization capability of the combined strong classifier. Moreover, we have tested the classification performance of the classifiers trained with different combinations of target poses. Experimental results show that a classifier trained with training images covering large variations of target poses can produce good classification result even with limited number of training images.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This work was supported by the EU H2020 TERPSICHORE project “Transforming Intangible Folkloric Performing Arts into Tangible Choreographic Digital Objects” under the Grant Agreement 691218.

References

  1. S. Ochilov and D. A. Clausi, “Operational SAR sea-ice image classification,” IEEE Transactions on Geoscience and Remote Sensing, vol. 50, no. 11, pp. 4397–4408, 2012. View at: Publisher Site | Google Scholar
  2. Y. Sun, Z. P. Liu, S. Todorovic, and J. N. Li, “Adaptive boosting for SAR automatic target recognition,” IEEE Transactions on Aerospace and Electronic Systems, vol. 43, no. 1, pp. 112–125, 2007. View at: Publisher Site | Google Scholar
  3. J.-I. Park and K.-T. Kim, “Modified polar mapping classifier for SAR automatic target recognition,” IEEE Transactions on Aerospace and Electronic Systems, vol. 50, no. 2, pp. 1092–1107, 2014. View at: Publisher Site | Google Scholar
  4. Ö. Aytekin, M. Koc, and I. Ulusoy, “Local primitive pattern for the classification of SAR images,” IEEE Transactions on Geoscience and Remote Sensing, vol. 51, no. 4, pp. 2431–2441, 2013. View at: Publisher Site | Google Scholar
  5. N. M. Sandirasegaram, “Spot SAR ATR using wavelet features and neural network classifier,” Tech. Rep., Report. DTIC Document, 2005. View at: Google Scholar
  6. M. Amoon and G.-A. Rezai-rad, “Automatic target recognition of synthetic aperture radar (SAR) images based on optimal selection of Zernike moments features,” IET Computer Vision, vol. 8, no. 2, pp. 77–85, 2014. View at: Publisher Site | Google Scholar
  7. C. Clemente, L. Pallotta, I. Proudler, A. de Maio, J. J. Soraghan, and A. Farina, “Pseudo-Zernike-based multi-pass automatic target recognition from multi-channel synthetic aperture radar,” IET Radar, Sonar & Navigation, vol. 9, no. 4, pp. 457–466, 2015. View at: Publisher Site | Google Scholar
  8. C. Zhu, H. Zhou, R. Wang, and J. Guo, “A novel hierarchical method of ship detection from spaceborne optical image based on shape and texture features,” IEEE Transactions on Geoscience and Remote Sensing, vol. 48, no. 9, pp. 3446–3456, 2010. View at: Publisher Site | Google Scholar
  9. Z. Jianxiong, S. Zhiguang, C. Xiao, and F. Qiang, “Automatic target recognition of SAR images based on global scattering center model,” IEEE Transactions on Geoscience and Remote Sensing, vol. 49, no. 10, pp. 3713–3729, 2011. View at: Publisher Site | Google Scholar
  10. E. Giusti, M. Martorella, and A. Capria, “Polarimetrically-persistent-scatterer-based automatic target recognition,” IEEE Transactions on Geoscience and Remote Sensing, vol. 49, no. 11, pp. 4588–4599, 2011. View at: Publisher Site | Google Scholar
  11. J.-I. Park, S.-H. Park, and K.-T. Kim, “New discrimination features for SAR automatic target recognition,” IEEE Geoscience and Remote Sensing Letters, vol. 10, no. 3, pp. 476–480, 2013. View at: Publisher Site | Google Scholar
  12. G. Dong, N. Wang, and G. Kuang, “Sparse Representation of Monogenic Signal: With Application to Target Recognition in SAR Images,” IEEE Signal Processing Letters, vol. 21, no. 8, pp. 952–956, 2014. View at: Publisher Site | Google Scholar
  13. H. Chen, H. Chang, and T. Liu, “Local discriminant embedding and its variants,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR '05), vol. 2, pp. 846–853, June 2005. View at: Publisher Site | Google Scholar
  14. F. Dornaika and A. Bosaghzadeh, “Exponential local discriminant embedding and its application to face recognition,” IEEE Transactions on Cybernetics, vol. 43, no. 3, pp. 921–934, 2013. View at: Publisher Site | Google Scholar
  15. X. Fang, Y. Xu, X. Li, Z. Fan, H. Liu, and Y. Chen, “Locality and similarity preserving embedding for feature selection,” Neurocomputing, vol. 128, pp. 304–315, 2014. View at: Publisher Site | Google Scholar
  16. X. Liu, Y. L. Huang, J. F. Pei, and J. Y. Yang, “Sample discriminant analysis for SAR ATR,” IEEE Geoscience and Remote Sensing Letters, vol. 11, no. 12, pp. 2120–2124, 2014. View at: Publisher Site | Google Scholar
  17. Y. Huang, J. Peia, J. Yanga, B. Wang, and X. Liu, “Neighborhood geometric center scaling embedding for SAR ATR,” IEEE Transactions on Aerospace and Electronic Systems, vol. 50, no. 1, pp. 180–192, 2014. View at: Publisher Site | Google Scholar
  18. X. Zhao, Y. Jiang, T. Stathaki, and H. Zhang, “Gait recognition method for arbitrary straight walking paths using appearance conversion machine,” Neurocomputing, vol. 173, pp. 530–540, 2016. View at: Publisher Site | Google Scholar
  19. X. Zhao, Y. Jiang, and W.-Q. Wang, “Efficient Clutter Suppression in SAR Images with Shedding Irrelevant Patterns,” IEEE Geoscience and Remote Sensing Letters, vol. 12, no. 9, pp. 1828–1832, 2015. View at: Publisher Site | Google Scholar
  20. X. Xing, K. Ji, H. Zou, and J. Sun, “Sparse representation based sar vehicle recognition along with aspect angle,” The Scientific World Journal, vol. 2014, Article ID 834140, pp. 174-175, 2014. View at: Publisher Site | Google Scholar
  21. G. Akbarizadeh, “A new statistical-based kurtosis wavelet energy feature for texture recognition of SAR images,” IEEE Transactions on Geoscience and Remote Sensing, vol. 50, no. 11, pp. 4358–4368, 2012. View at: Publisher Site | Google Scholar
  22. L. Bruzzone, M. Marconcini, U. Wegmüller, and A. Wiesmann, “An advanced system for the automatic classification of multitemporal SAR images,” IEEE Transactions on Geoscience and Remote Sensing, vol. 42, no. 6, pp. 1321–1334, 2004. View at: Publisher Site | Google Scholar
  23. Y. Wang, P. Han, X. Lu, R. Wu, and J. Huang, “The performance comparison of Adaboost and SVM applied to SAR ATR,” in Proceedings of the CIE International Conference on Radar (ICR '06), pp. 1–4, October 2006. View at: Publisher Site | Google Scholar
  24. U. Srinivas, V. Monga, and R. G. Raj, “SAR automatic target recognition using discriminative graphical models,” IEEE Transactions on Aerospace and Electronic Systems, vol. 50, no. 1, pp. 591–606, 2014. View at: Publisher Site | Google Scholar
  25. S. G. Mallat, “Theory for multiresolution signal decomposition: the wavelet representation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 11, no. 7, pp. 674–693, 1989. View at: Publisher Site | Google Scholar
  26. C. Qiu, H. Ren, H. Zou, and S. Zhou, “Performance comparison of target classification in SAR images based on PCA and 2D-PCA features,” in Proceedings of the 2009 Asia-Pacific Conference on Synthetic Aperture Radar, APSAR 2009, pp. 868–871, China, October 2009. View at: Publisher Site | Google Scholar
  27. R. E. Schapire and Y. Singer, “Improved boosting algorithms using confidence-rated predictions,” Machine Learning, vol. 37, no. 3, pp. 297–336, 1999. View at: Publisher Site | Google Scholar
  28. H. Zhang, H. Lin, and Y. Li, “Impacts of feature normalization on optical and SAR data fusion for land use/land cover classification,” IEEE Geoscience and Remote Sensing Letters, vol. 12, no. 5, pp. 1061–1065, 2015. View at: Publisher Site | Google Scholar
  29. Q. Zhao, D. Xu, and J. C. Principe, “Pose Estimation for SAR Automatic Target Recognition,” in Proceedings of Image Understanding Workshop, pp. 827–832, 1999. View at: Google Scholar
  30. J. C. Principe, D. Xu, and J. W. Fisher III, “Pose estimation in SAR using an information theoretic criterion,” in Proceedings of the Algorithms for Synthetic Aperture Radar Imagery V, pp. 218–229, usa, April 1998. View at: Publisher Site | Google Scholar
  31. L. M. Kaplan and R. Murenzi, “Pose estimation of SAR imagery using the two dimensional continuous wavelet transform,” Pattern Recognition Letters, vol. 24, no. 14, pp. 2269–2280, 2003. View at: Publisher Site | Google Scholar
  32. L. Voicu, R. Patton, and H. Myler, “Multi-criterion vehicle pose estimation for SAR-ATR,” in In Proceedings of SPIE - The International Society for Optical Engineering, vol. 372110. View at: Google Scholar
  33. S. Helgason, “The Radon transform,” Progress in Mathematics, vol. 13, no. 89, pp. 81–133, 2009. View at: Google Scholar
  34. G. Toussaint, “Solving geometric problems with the rotating calipers,” in Proceedings of the MELECON '83, Mediterranean Electrotechnical Conference. View at: Google Scholar
  35. J. Liu, S. Chen, X. Tan, and D. Zhang, “Comments on ‘Efficient and robust feature extraction by maximum margin criterion’,” IEEE Transactions on Neural Networks and Learning Systems, vol. 18, no. 6, pp. 1862–1864, 2007. View at: Publisher Site | Google Scholar
  36. S. Sanghavi, V. Tan, and A. Willsky, “Learning graphical models for hypothesis testing,” in Proceedings of the 2007 IEEE/SP 14th WorkShoP on Statistical Signal Processing, SSP 2007, pp. 69–73, USA, August 2007. View at: Publisher Site | Google Scholar
  37. V. Y. Tan, S. Sanghavi, I. Fisher, and A. S. Willsky, “Learning graphical models for hypothesis testing and classification,” IEEE Transactions on Signal Processing, vol. 58, no. 11, pp. 5481–5495, 2010. View at: Publisher Site | Google Scholar | MathSciNet
  38. T. M. Cover and J. A. Thomas, Elements of Information Theory, Wiley-Interscience Publication, New York, USA, 2nd edition, 2012. View at: MathSciNet
  39. S. Lauritzen, Graphical Models, Oxford University Press, Oxford,England, 1996.
  40. C. Chow and C. Liu, “Approximating discrete probability distributions with dependence trees,” IEEE Transactions on Information Theory, vol. 14, no. 3, pp. 462–467, 1968. View at: Publisher Site | Google Scholar
  41. J. Kruskal, “On the shortest spanning subtree of a graph and the traveling salesman problem,” Proceedings of the American Mathematical Society, vol. 7, pp. 48–50, 1956. View at: Publisher Site | Google Scholar | MathSciNet
  42. S. Kullback, Information Theory and Statistics, Wiley-Interscience Publication, New York, USA, 1959.
  43. R. E. Schapire, Y. Freund, P. Bartlett, and W. S. Lee, “Boosting the margin: a new explanation for the effectiveness of voting methods,” The Annals of Statistics, vol. 26, no. 5, pp. 1651–1686, 1998. View at: Publisher Site | Google Scholar | MathSciNet
  44. J. Friedman, T. Hastie, and R. Tibshirani, “Additive logistic regression: a statistical view of boosting (with discussion and a rejoinder by the authors),” The Annals of Statistics, vol. 28, no. 2, pp. 337–407, 2000. View at: Publisher Site | Google Scholar | MathSciNet
  45. P. Viola and M. Jones, “Robust real-time object detection,” International Journal of Computer Vision, vol. 4, pp. 51-52. View at: Google Scholar
  46. A. Vezhnevets and V. Vezhnevets, “Modest AdaBoost-teaching AdaBoost to generalize better,” in Graphicon, pp. 12987–12997, 2005. View at: Google Scholar
  47. S. Wu and H. Nagahashi, “A new method for solving overfitting problem of gentle AdaBoost,” in Proceedings of the 5th International Conference on Graphic and Image Processing, ICGIP 2013, China, October 2013. View at: Publisher Site | Google Scholar
  48. J. Milgram, M. Cheriet, and R. Sabourin, “Speeding Up the Decision Making of Support Vector Classifiers,” in Proceedings of the Ninth International Workshop on Frontiers in Handwriting Recognition, pp. 57–62, Tokyo, Japan. View at: Publisher Site | Google Scholar
  49. J. C. Mossing and T. D. Ross, “An evaluation of SAR ATR algorithm performance sensitivity to MSTAR extended operating conditions,” in Proceedings of the Algorithms for Synthetic Aperture Radar Imagery V, pp. 554–565, USA, April 1998. View at: Publisher Site | Google Scholar
  50. T. Ross, S. Worrell, V. Velten, J. Mossing, and M. Bryant, “Standard SAR ATR evaluation experiments using the MSTAR public release data set,” in Proceedings of the Algorithms for Synthetic Aperture Radar Imagery V, pp. 566–573, USA, April 1998. View at: Publisher Site | Google Scholar
  51. T. D. Ross and C. J. Mossing, “MSTAR evaluation methodology,” in AeroSense’99, pp. 705–713, International Society for Optics and Photonics, 1999. View at: Google Scholar
  52. Q. Zhao and J. C. Principe, “Support vector machines for SAR automatic target recognition,” IEEE Transactions on Aerospace and Electronic Systems, vol. 37, no. 2, pp. 643–654, 2001. View at: Publisher Site | Google Scholar
  53. R. Singh and B. V. K. Vijaya Kumar, “Performance of the extended maximum average correlation height (EMACH) filter and the polynomial distance classifier correlation filter (PDCCF) for multi-class SAR detection and classification,” Proceedings of SPIE - The International Society for Optical Engineering, vol. 4727, pp. 265–276, 2002. View at: Publisher Site | Google Scholar

Copyright © 2017 Xiaohui Zhao et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

834 Views | 1015 Downloads | 2 Citations
 PDF  Download Citation  Citation
 Download other formatsMore
 Order printed copiesOrder

We are committed to sharing findings related to COVID-19 as quickly and safely as possible. Any author submitting a COVID-19 paper should notify us at help@hindawi.com to ensure their research is fast-tracked and made available on a preprint server as soon as possible. We will be providing unlimited waivers of publication charges for accepted articles related to COVID-19.