Research Article  Open Access
Lin Song, Yongmei Cheng, Lu Yu, Liang Yu, "A Method of Waypoint Selection in Aerial Images for Vision Navigation", International Journal of Optics, vol. 2014, Article ID 161026, 8 pages, 2014. https://doi.org/10.1155/2014/161026
A Method of Waypoint Selection in Aerial Images for Vision Navigation
Abstract
We present a novel method to select waypoints from aerial images of candidate flying regions via matching suitability analysis, which is based on visual attention mechanism and feature attribute classification. At first, visual attention mechanism is used to get the saliency map of the initial image by lowrank recovery and sparse coding. The salient regions are selected to be as preparatory results with threshold constraint and nonmaxima suppression. Then we use support vector machine (SVM) to divide the preparatory results into two classes for suitable or unsuitable waypoints based on their feature attributes, which can be represented by two edgebased descriptors and two correlationbased descriptors. The experimental results show that the proposed method is valid and effective.
1. Introduction
In recent years, the research on vision navigation based on scene matching technique has attracted more and more attentions for its accuracy and independence [1]. In order to assure the matching precision, it is a primary step to select waypoints from aerial images of candidate flying regions for scene matching, which can be implemented via suitability analysis [2–4].
Some researchers have made efforts to solve the problem of suitability analysis. Yang et al. [2] used graybased descriptors and edgebased descriptors as the input of SVM to classify the matching suitability of the image. Jiang and Chen [5] provided a hierarchical way of selecting optimal scene matching areas. Liu et al. [6] presented the method of selecting matching areas using independent pixel numbers and variance. In [7] it was suggested that information entropy and summation of image gradient can be used for evaluating image suitability. However, features used in the above methods are inadequate to describe the suitability and there is redundancy information among the feature descriptors. Research has shown that the suitability is considered to consist of information, obviousness, stability, and uniqueness [1]. From the view of visual analysis, the information and obviousness can be analyzed via visual saliency. Visual saliency is a cognitive procedure which can rapidly select a small set of highly informative or visually salient objects from a scene for further processing [8]. The stability and uniqueness are the feature attributes of an image. So we can transform the suitability analysis to the combination of visual saliency analysis and feature attributes classification.
For visual saliency analysis, in order to avoid the uncertainty influence of statistic features, Li et al. [9] proposed the visual saliency analysis method based on length of incremental coding, which is based on sparse coding [10, 11] and centersurround model (CS); Han et al. [12] proposed saliency analysis method based on weighted length of incremental coding. They are all efficient approaches to select regions with local saliency, but for matching suitability the saliency should be with unique structure information, which means it is sparse in the global image. So lowrank recovery is introduced to analyze the global and local saliency with sparse coding for preparatory selection. For feature attributes classification, SVM is used to analyze stability and uniqueness for optimizing selection. So this paper presents a practical framework for waypoint selection, as illustrated in Figure 1.
The proposed framework consists of two major components: a preparatory selection model based on visual saliency analysis and an optimal selection model based on SVM. In the formal model, the initial image is decomposed as the sum of sparse matrix and lowrank matrix, and then saliency of sparse matrix and lowrank matrix is analyzed, respectively, to construct a new saliency map. The preparatory selection results are got with threshold constraint and nonmaxima suppression. In the second component, SVM is used for optimizing selection, and the input vector is composed of four measure parameters based on the edge and cross correlation surface.
The rest of the paper is organized as follows. Section 2 describes the preparatory selection model based on visual saliency analysis. Section 3 presents optimal selection of waypoints model based on feature attribute classification. Section 4 reports and analyzes evaluation results. Finally, conclusions are drawn in Section 5.
2. The Preparatory Selection Based on Visual Saliency Analysis
Salient objects can be viewed as a small number of foregrounds, which are different from the surrounding backgrounds. So lowrank matrix recovery is introduced to separate foreground from background [13]. The lowrank matrix and the sparse matrix are recovered by optimizing the constraint, where represents matrix rank, represents norm, and is the weight parameter which is used to weight the sparse relationship between the rank of and the sparsity of . Given a suitable , we want to be able to get a pair of . However, it is a nonconvex problem for the existence of and . Usually, (1) is relaxed by solving the convex optimization problem where is the sum of singular values of matrix and represents norm. The optimization problem (2) is solved by utilizing augmented Lagrange multiplier (ALM) to get and , respectively, as shown in Figures 2(b) and 2(c).
(a) Original image
(b) Lowrank image
(c) Sparse image
There are two kinds of saliency in aerial images. One is with its unique information of scene structure (shown as green dotted line in Figure 2(c)) and the other is some small manmade areas with high brightness where there is no structure information (shown as red dotted line in Figure 2(c)). We can see that the second saliency is the same as its surrounding objects in the image whereas the first saliency in both and is different from its surrounding objects. So we can separate the two kinds of saliency by local saliency analysis of and , respectively, based on centersurround model (CS).
Here sparse coding is introduced to describe the local saliency with CS model. Sparse coding codes the center patch over a dictionary constructed by the surrounding patches. If the center patch is similar to its surroundings, it has sparse coefficients. Denote by the th patch of , and is dictionary consisting of surrounding patches, which is represented by a set of vectors as a dictionary. Consider . So the problem of saliency based on sparse coding is shown as follows: where is the balance factor between sparsity and data integrity. The norm optimization problem can be solved efficiently by Lasso method [11]. The local saliency of image patch is obtained as (4). The process is shown as in Figure 3. Consider
All patches of the image can be calculated, so we can get the saliency map .
We calculated the local saliency in both and based on sparse coding. And then the new saliency could be represented by the following function:
A certain threshold is used to judge possible waypoints, and the rule is
For the region , the nonmaxima suppression is used to get peaks in , which are the centers of possible waypoints.
3. Optimal Selection Based on Feature Attribute Classification
It is a problem to judge whether there are stability and uniqueness of preparatory results. From the viewpoint of pattern recognition, it can be solved by twoclass classification. So, SVM is introduced.
3.1. Introduction of SVM
SVM lies in strong connection to the underlying statistical learning theory, where it implements the structural risk minimization for solving the problem of twoclass classification [2]. SVM has advantages in solving the problems like small samples, high dimensions, and large scale.
Given a training sample set of labeled examples, with each input and the output label , is the number of training samples. The best hyperplane ( is a constant) to separate two classes is achieved by satisfying the constraint: where is norm of a vector, so (7) is equivalent to minimize with the same constraint. The decision function is
When the original set will not be linearly separable, it is common to define a soft margin by including variables and parameter
The original sample can be mapped into a highdimensional space (named feature space) by nonlinear transform and is expressed as the dual form . So classification output can be predicted using the decision function, as where is the kernel function. There are three forms of kernel function: radial basis function and the linear and polynomial kernel. Preliminary researches suggest that the radial basis function outperforms the others. So the radial basis function kernel in the following equation will be used in the classification:
3.2. Feature Selection
Suitable descriptors as the input vector of SVM can optimize computational efficiency and gain the better classification results. Here measure parameters based on the edge and cross correlation surface are considered for stability and uniqueness analysis.
3.2.1. Stability
The stability is an important feature attribute of an aerial image, which is suitable for scene matching. So we need to select waypoints with stable features. Edge complexity and edge density are selected to be measure parameters of stability.
(1) Edge Complexity. Edge complexity is a homogeneity parameter of edge texture distribution. When it is smaller, the image is more smoother, which will lead to mismatching more easily. Edge complexity is calculated by where is a local neighborhood with the center and and are 1dimensional derivative and 2dimensional derivative, respectively.
(2) Edge Density. Edge density can show the concentration of features in the original image. It is computed by where is the number of points in the neighborhood with the center . is the number of pixels in the neighborhood.
3.2.2. Uniqueness
The global uniqueness of waypoints is analyzed to avoid the selection of repeated scenery areas. The uniqueness is determined by cross correlation plane statistic feature. Cross correlation plane is computed pixel by pixel in the whole image via matching waypoint image . We use two features of cross correlation, Submaxratio and Ngb8maxratio: where is the mean of and is the mean of an area with the same size as in .
(1) Submaxratio. Submaxratio denotes ratio of secondary maximum peak to maximum peak, which is computed using where is secondary high correlation peak and is maximal correlation peak. It means the waypoint has better uniqueness when the value of is closer to zero.
(2) Ngb8maxratio. Ngb8maxratio represents ratio of maximum of eight neighbor peaks to maximum peak, which is computed using where is maximum of eight neighbor peaks and is maximal correlation peak.
3.3. Training and Classifying
Select 100 sample images as a training set and label each image manually. 50 images are waypoints, and the other 50 images are nonwaypoints, which are shown as in Figure 4.
(a) Waypoint 1
(b) Waypoint 2
(c) Nonwaypoint 1
(d) Nonwaypoint 2
The image is decrypted by the measure parameter vector based on the edge and cross correlation surface. So the vector is used as the input for SVM. The feature vectors should be normalized before training. The best parameters and for the SVM classifying using (11) are obtained via training. In the testing, each of preparatory results is decrypted as and normalized, which is put into SVM for classifying suitable or unsuitable results.
4. Experiments
As known at present, automatic selection method of waypoints has not been reported up to now, so there is no public dataset for method validation. To evaluate the performance of our method, experiments on aerial images from Google Earth are conducted. The image pair of the same scene is taken at different time, as shown in Figure 5. One is used as reference image for waypoint selection, and the other is as sensed image used to verify the suitability of selected waypoints. For simplicity, the size of reference image and sensed image is set to be and the size of waypoint image is set to be in all experiments.
(a) Reference image
(b) Sensed image
There are two kinds of reference images. One is called Class 1 reference, part of which is with its unique information of scene structure and can be selected as waypoints. The other is called Class 2 reference which is without any unique information of scene structure at all, as shown in Figure 6. The quantities of the two kinds are 150, respectively.
(a) Class 1 reference 1
(b) Class 1 reference 2
(c) Class 2 reference 1
(d) Class 2 reference 2
4.1. Preparatory Selection
The method based on sparse coding and lowrank matrix recovery is used to implement preparatory selection of waypoints. The result is shown as in Table 1.

A few waypoints can be selected from one reference image, so the quantity of waypoints is larger than the quantity of references. Part results of Class 1 are shown as in Figure 7. In order to reduce the complexity of analysis, saliency coefficient is normalized between 0 and 1. Set Threshold_{saliency} = 0.75, and the regions are selected as waypoints, which of local saliency maximum in the centre are higher than Threshold_{saliency}. For comparison, two stateoftheart saliency analysis methods are introduced: Li et al.’s [9] and Han et al.’s [12]. The results of waypoints selection are shown in Figure 7.
(a)
(b)
From Figure 7 we can see that Li et al.’s [9] and Han et al.’s [12] saliency emphasized the areas with more brightness information, most of which are without unique information of scene structure. Our method can extract the areas with salient structure information and effectively inhibit the disturbance of brightness; for example, in the fourth line results, just the traffic intersections are extracted because of their structure information. Therefore, the number of the waypoint candidates is less than the results of the former methods.
When we analyzed the reference images from Class 2 in the preparatory selection, there were still saliency regions in the images of Class 2, as shown in Figure 8.
(a)
(b)
We note that the methods are ineffective in analyzing the saliency in the references from Class 2. It is because of the normalization of the saliency coefficients. The results in Class 2 are not suitable to be waypoints, so we need to classify the results in Class 1 and the results in Class 2 with SVM.
4.2. Optimal Selection
SVM is used to optimally select waypoints by classifying the feature attributes. To evaluate the results of classification, cross correlation matching method is used for verification (it is thought to be a correct matching when the matching error is smaller than 5 pixels). The result is shown as in Table 2.

There are two kinds of mistakes in the process of classification. One is called “undetected,” which means a waypoint is classified as nonwaypoint. The other is called “false detected,” which means a nonwaypoint is classified as waypoint. The former error is tolerable, but the latter is fatal for vision navigation. So the latter should be avoided or be reduced as much as possible. The classification rate is shown as in Table 3.

From Table 3, we know that though there are many waypoints in images of Class 1, false detection still exists because of the scenes changing with time. Undetected is produced by the difference between training samples and testing samples. There is no undetected mistake in Class 2. It is because that there is no waypoint in the reference image of Class 2.
For comparison, the algorithms [2, 5] are used for waypoints selection. Random sampling investigation was carried out on sets of Class 1 and Class 2 with the same quantity of waypoints as in Table 1. The times are 1000 and the result is shown as in Table 4.
We can see that, in the analysis of Class 1, our method is better than the other two methods, and, in the analysis of Class 2, our method is better than [2] and almost the same as [5]. It is because that the threshold in [5] is set manually.
5. Conclusions
A method of waypoints selection was proposed in this paper, which firstly selected salient areas as candidate waypoints and then classified the candidates based on their feature attributes. The method combined the visual saliency analysis and feature attributes classification and especially avoided the inference of some small manmade areas with high brightness where there is no structure information.
The sensed image and the reference image are both from Google Earth, which makes the suitability analysis only depend on the original reference image itself and does not consider the matching condition under the geometrical transformations such as image scaling and rotation. In the next stage, we plan to extend this work along the following directions. Firstly, the matching condition will be considered for suitability analysis, and more aerial images under the real flying condition will be used to test the validity of the approach. Secondly, we will incorporate more powerful features or improve the saliency analysis model and classifying model to improve the effectiveness of the method.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
Acknowledgments
This research was financially supported by Xi’an Science and Technology Project (CXY1350(2)) and Aeronautical Science Foundation of China (20100853010).
References
 L. Shen, Y. Bu, X. Xu, and L. Pan, “Research on matchingarea suitability for scene matching aided navigation,” Acta Aeronautica et Astronautica Sinica, vol. 31, no. 3, pp. 553–563, 2010. View at: Google Scholar
 Z. Yang, Y. Chen, X. Qian, M. Yuan, and E. Gao, “Predicting the suitability for scene matching using SVM,” in Proceedings of the International Conference on Audio, Language and Image Processing (ICALIP '08), pp. 743–747, July 2008. View at: Publisher Site  Google Scholar
 J. Du and T. Zhang, “Selection of matching region for scene matching,” Infrared and Laser Engineering, vol. 32, no. 4, pp. 368–371, 2003. View at: Google Scholar
 W.X. Fu, J.M. Wang, and S.L. Jin, “Practical method for selecting scene matching area,” Journal of Astronautics, vol. 24, no. 4, pp. 348–353, 2003. View at: Google Scholar
 B. Jiang and Y. Chen, “Rule of selecting scene matching area,” Journal of Tongji University, vol. 35, no. 6, pp. 830–833, 2007. View at: Google Scholar
 Y. Liu, F. Zhao, and S. Jin, “New method of selecting scene matching reference map,” Infrared and Laser Engineering, vol. 30, no. 3, pp. 168–171, 2001. View at: Google Scholar
 R. An, X.L. Jin, H.L. Wang, X.Z. Feng, and D.X. Xu, “Image ability to obtain correct matching based on feature matching,” Infrared and Laser Engineering, vol. 34, no. 4, pp. 469–473, 2005. View at: Google Scholar
 J. Han, P. Zhou, D. Zhang et al., “Efficient, simultaneous detection of multiclass geospatial targets based on visual saliency modeling and discriminative learning of sparse coding,” ISPRS Journal of Photogrammetry and Remote Sensing, vol. 89, pp. 37–48, 2014. View at: Publisher Site  Google Scholar
 Y. Li, Y. Zhou, L. Xu, X. Yang, and J. Yang, “Incremental sparse saliency detection,” in Proceedings of the IEEE International Conference on Image Processing (ICIP '09), pp. 3093–3096, November 2009. View at: Publisher Site  Google Scholar
 J. Wright, A. Y. Yang, A. Ganesh, S. S. Sastry, and Y. Ma, “Robust face recognition via sparse representation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 31, no. 2, pp. 210–227, 2009. View at: Publisher Site  Google Scholar
 R. Tibshirani, “Regression shrinkage and selection via the lasso,” Journal of the Royal Statistical Society B. Methodological, vol. 58, no. 1, pp. 267–288, 1996. View at: Google Scholar  MathSciNet
 B. Han, H. Zhu, and Y. Ding, “Bottomup saliency based on weighted sparse coding residual,” in 19th ACM International Conference on Multimedia ACM Multimedia 2011, MM'11, pp. 1117–1120, Scottsdale, Ariz, December 2011. View at: Publisher Site  Google Scholar
 Z. C. Lin, M. M. Chen, L. Q. Wu, and Y. Ma, “The augmented lagrange multiplier method for exact recovery of corrupted lowrank matrices,” Tech. Rep. UILUENG092215, University of Illinois at UrbanaChampaign, Champaign, Ill. USA, 2009. View at: Google Scholar
Copyright
Copyright © 2014 Lin Song et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.