Research Article  Open Access
A RealTime Optical Tracking and Measurement Processing System for Flying Targets
Abstract
Optical tracking and measurement for flying targets is unlike the close range photography under a controllable observation environment, which brings extreme conditions like diverse target changes as a result of high maneuver ability and long cruising range. This paper first designed and realized a distributed image interpretation and measurement processing system to achieve resource centralized management, multisite simultaneous interpretation and adaptive estimation algorithm selection; then proposed a realtime interpretation method which contains automatic foreground detection, online target tracking, multiple features location, and human guidance. An experiment is carried out at performance and efficiency evaluation of the method by semisynthetic video. The system can be used in the field of aerospace tests like target analysis including dynamic parameter, transient states, and optical physics characteristics, with security control.
1. Introduction
The performance testing of flying targets like aircraft and missile is an important process in their civil or military production. The main testing item is trajectory measurement, including dynamic parameters (such as pose, velocity, and acceleration) and optical physics characteristics analysis (such as infrared radiation, flame spectrum, and luminance brightness). Optical tracking and measurement method is widely adopted in testing process, because optical sensor is passive, of low power consumption, and of noncontact; in addition, optical images are objective and informative. Because of targets’ high maneuver ability and large cruising range, the observation environment is extreme, including target changes (such as pose, shape, scale, and motion blur) and circumstance changes (such as illumination and occlusion), but we still need high precision in performance analysis and high efficiency for security control. So an automatic and accurate interpretation method to conquer these extreme conditions is critical in realtime processing and this paper puts emphasis on the key components including foreground detection, online tracking, and feature location.
There are many worldwide famous video motion analysis systems, like TrackEye [1] from Sweden Image Systems cooperation, Visual Fusion [2] from America MIT, Movias Pro [3] from America NAC, and so forth. They all can provide transient state record, realtime image interpretation, and dynamic parameter estimation for flying targets using highspeed camera. But most of them depend on feature points or markers, whereas we care about the image interpretation for markerless targets with finite prior knowledge on the case that large sensor standoff distance and outdoor environment make a target region with a weak texture in the image.
This paper starts from the demand analysis of the optical tracking and measurement for flying targets. Section 2 designs the architecture of distributed realtime processing system and introduces the compositions in brief. In Section 3 a detailed image interpretation method and a short parameter estimation method are presented. Section 4 realizes the image interpretation subsystem and validates the performance by the experiment. Section 5 draws the conclusion.
2. Architecture
The realtime processing of optical tracking and measurement system has the characteristics of high efficiency, strong parallelism, and rigorous time series. As Figure 1 shows, the system adopts distributed architecture in terms of load balancing to improve efficiency and scalability, which is loosely decomposed into data receiving, image interpretation and parameter estimation three subsystems according to the course of first interpretation and then estimation. In view of the variety of different observation platforms and high computation resource demand of multisite synchronous interpretation, the parallel processing is needed which makes each interpretation subsystem accomplish one site task and the number of observation platforms decides the number of interpretation subsystems. The system communicates by network whose protocol is TCP/IP. Data server provides the data of the system via link A, which distinguishes data from different moments and sensors by time stamp and sensor serial number, and the results of estimation are exported to the external system for analysis and display.
It can provide two kinds of work modes including realtime and postprocessing. Data transmitting subsystem is the manager of the whole system, which receives interpretation task data package from data server by link A, splits the package according to the sensor serial number to send each package to the subsystem by link B or C, and records the flying scene grouped by task in the database for postrecurrence and interpretation; image interpretation subsystem picks up the data package from data transmitting subsystem, extracts the image in the package to realize one site interpretation which can be guided by interpreter when the result is abnormal, and integrates the interpretation result with other information in the package to send it to the estimation subsystem through link D or E. Parameter estimation subsystem decides whether the data is from one site or multisite by time stamp alignment to select one site or multisite measurement manner automatically and send results to the external system for further analysis.
3. RealTime Processing
The difficulty of realtime processing is accuracy and efficiency of the image interpretation, because parameter estimation methods during realtime or postprocessing are nearly the same. So in this chapter, we propose a detailed realtime image interpretation method and make a brief introduction about parameter estimation.
3.1. Image Interpretation
Image interpretation is an image understanding process, which needs to detect the foreground target, track it online, and locate the feature for parameter estimation during realtime processing. A human guidance policy is essential on account of more false positives and false negatives during longterm flying.
3.1.1. Foreground Detection
Automated systems need to use some form of foreground detection mechanisms to identify the target region to be tracked. Foreground detection can be divided into two types including single frame detection and sequence frames detection according to the number of used frames. Single fame detection is a visual object recognition problem [5] which needs the target prior appearance information from offline training and online update; sequence frames detection is a change detection problem which uses background subtraction like parameter estimation method GMM [6] or nonparameter estimation method VIBE [7] for fixed field of view and adopts a interframe registration such as optical flow [8] to describe motion information followed by a trajectory analysis based on tensor voting [9] or epipolar geometry [10] for varied field of view. Here we propose a simple detection algorithm which combines twoframe motion segmentation with onesingle shape recognition.
The motion detector relies on twoframe optical flow filed. Many methods use a pixellevel analysis which is of high computational cost. In view of the scene whose background is nearly a plain region like sky or gobi in flying target tests, we downsample the image with grids which is marked with one pixel in each grid. The pixel may be a strong KLT feature point or a central point. If it is a KLT point, a pyramid LucasKanade is used to track it; otherwise a pyramid NCC can be used to match the grid with a template size and a search range . The feature selection is shown in Figure 2(a) with optical flow estimation in Figure 2(b) where white points are KLT feature points, green ones are central points, and red lines are optical flow. The orientation and magnitude of can be computed in the following: which can be used to count the histogram of optical flow orientation noted as HOOF assigning a weight according to the following: where is the set of sampling pixels, sgn is the sign function, and bin is the capacity of . The normalized is in (3). In light of , a backprojection is shown in Figure 2(c) without stationary sampling points. Figure 2(d) shows the motion segmentation result with occupy map based on HOOF where the same color means an identical motion region. Because the target may be a small part of the image which provides little motion information to the total optical flow field, we concern the connected region more than occupy value and tend to select larger occupy value when there are multiple connected regions
(a) Feature selection ( )
(b) Optical flow ( )
(c) Backprojection with HOOF (bin = 18)
(d) Motion segmentation and occupy map
The appearance detector depends on the shape information, because the image intensity is varying along the observation distance, and the shape is more robust feature which is affected by visual angle and acquired easily. In addition, the observation distance is larger than the target dimension, so a projective transformation can be approximated by an affine warp.
As in Figure 3(a) we render partial 2D aircraft shape models at some step according to the 3D model from Trimble 3D warehouse [11] and the visual angle. The affineinvariant boundary moment [12] in (4) is adopted to represent the shape which can bear small quantity of variable visual angle and noise, where is the normalized boundary central moment. Compute the boundary moment of the th target shape model offline noted as the set , where is the sample identifier, is the target type, is the visual angle, the number of contour sampling points is , the number of predefined interest points is , and is the boundary moment descriptor. Calculate the boundary moment descriptor of the th target contour online like Figure 3(b), and the similarity metric between and by a revised chisquare distance in (5) for the value of or may be nonpositive, where is a small amount and less means more similar. The detection is a nonminimum suppression process and a similarity metric between and is shown in Figure 3(c). The association of motion detector and appearance detector is easy, which outputs the biggest intersection as the target region with a bounding box definition.
(a) Partial shape models
(b) Contour detection
(c) Shape similarity
3.1.2. Target Tracking
Tracking must have the function of detection and online learning, because flying targets may leave/reenter the fieldofview and gradually change the pose. TLD [13] is a novel framework based on semisupervised learning which decomposes the tracking task into tracking, detection, integration, and learning four components. We improve it from adaption and efficiency for realtime application and propose an AATLD [14] as an acronym for adaptive and accelerated trackinglearningdetection.
Figure 4 depicts the workflow of AATLD whose solid rectangle is a component, dashed rounded rectangle is a unit, and shaded dashed rounded rectangle is the modified unit. According to the bounding box of the target from foreground detection, initiator builds only the current scale space of the target with a sampling step and trains an initial fern detector and a NN detector according to a handful of samples with a threshold adjustment by crossvalidation.
Unlike TLD’s sequential execution, AATLD parallelizes tracker and detector to enhance the efficiency for independence by OPENMP. The tracker is the same with TLD, which uses median flow based on discrete sampling points in the target region and can be substituted with other easy tracking methods. The detector is realized in the particle filter framework to generate new scale scanning grids online according to the predicted scale and uses the same cascade process but fixes the number of the positive and negative samples which ensures a constant retrieving time by ordering the sample contribution ratio as (6), where is the positive or negative sample, is the potential target grid, and is the similarity metric in TLD
TLD associates the results of tracker and detector by weightaverage which assigns the weight of tracking result 10, and each detection result 1. Maybe it is not convincing because the results are from different scales but are assessed by normalized correlation coefficient (NCC). AATLD adopts a weight computing method considering both NCC and scale size according to [15] as (7), where is NCC, is the number of pixels, and is a small quantity
The learning stage of TLD only puts emphasis on updating the feature library by PN learning but not adjusting the threshold which is also important in decision process. AATLD adds a distance metric parameter adjustment component to online update the threshold by crossvalidation like initiator. Partial tracking results of Shenzhou IX datasets which are shown in Figures 5 and 6 show the performance contrast details. We can conclude that AATLD is faster than TLD but less accurate than TLD at center location which is all right for the emphasis is an initial target region for feature location.
(a)
(b)
3.1.3. Feature Location
The feature can be contour, line, or point which provides image coordinates to parameter estimation. Target tracking provides a bounding box of the target, and locating the feature in the bounding box will bring more accurate results than the total image.
The contour can be detected by a realtime approximate levelset method in [16]. The axis and the edge line are the major line feature in the image. After a stateoftheart lineartime line segment detector LSD [17], clustering and additional criterion can be used to locate these edge line features. The axis can be extracted using moment of inertia of the target region or the halving line of two edge line segments.
The feature point can be centroid or interest points. The priority of the centroid computing is locating target pixel sets which can be acquired by contour scanning or saliency detection [18]. There are two kinds of interest points. One may be corners or bloblike points and can be established stereo correspondence by image matching like in [19]; the other can be the projection of the physical points as in [20, 21] with a prior knowledge of the target like 3D or 2D model. Here we propose simple and fast multiple interest physical points locating method based on ICP [22] with 2D shape models in Algorithm 1, which can offer image coordinates to pose estimation based on points.

Some examples of feature location are shown in Figure 7. Different features adapt to different targets and observation conditions.
3.1.4. Human Guidance
For correctness and efficiency of the system, when interpretation error occurs, the interpreter instantly provides guidance information which can be a simple seed point or a coarse rectangle region by humancomputer interaction. The guidance does not interrupt the continuous running; when the system receives the guidance information, it will seek the bounding box of the target using region growing algorithm based on a guidance point or search it using grab cut [23] based on a guidance rectangle and learn the region feature to update the sample library at the next frame. The update strategy contains adding the feature which puts the positive sample into the library and deleting the feature that removes the similar sample with the guidance region from the library by nearest neighbor (NN) search as shown in Figure 8.
(a) Adding target
(b) Deleting target
3.2. Parameter Estimation
Optical sensor can provide more appearance information like color, texture, and gradient than radar which relies purely on position and motion information [24]. Optical measurement can supply not only dynamic parameter estimation but also optical characteristic analysis.
3.2.1. Dynamic Parameter Estimation
Different observation manners and image features need different parameter estimation methods. The conventional observation apparatus for flying target tests is optoelectronic theodolite for longdistance measurement or highspeed camera for closerange measurement.
The measurement based on optoelectronic theodolite [25] can realize onesite localization, because the apparatus outputs the angular altitude, azimuth angle, and slope distance of the target by missing distance interpretation of one point standing for the target. And multisite triangulation will bring higher precision than onesite localization. If there is no slope distance information like highspeed camera observation, one site interpretation with material point hypothesis cannot locate the target, and the triangulation measurement [26] is necessary.
One point interpretation cannot estimate the attitude. But multiple points from single sensor with known correspondence between object points and image points is a classical PNP problem which estimates the pose like [27, 28]. The axis is the common line segment feature for rigid rotation, which can be used to locate with triangulation and estimate the pose in [29, 30]. With 3D model and the contour in the image of the target, modelbased pose tracking method can be used to estimate the pose, like PWP3D [20]. The trajectory, velocity, and acceleration can be computed with time information.
3.2.2. Optical Characteristic Analysis
The analysis of optical physics characteristics about infrared radiation, flame spectrum, and luminance brightness needs the silhouette of the target in the image. The histogram of the intensity of the pixels inside represents the characteristics, and the histogram is defined as follows: where is the number of pixels inside , is the intensity bins, and is the number of pixels which belongs to the th intensity bin. The statistical optical characteristics can derive from the mean, variance, coefficient of skewness, coefficient of kurtosis, energy, entropy, and other criteria of . Figure 9 illustrates an example of optical characteristic analysis. Figure 9(a) shows an infrared image of Shenzhou X emission marked with the contour, whose histogram is in Figure 9(b) with intensity range from 0 to 255 and 8 intensity bins. Figure 9(c) presents the statistical analysis.
(a) Shenzhou X infrared image
(b) Histogram
(c) Statistics
4. Experiment
We test our realtime processing system at precision and efficiency based on the image interpretation subsystem, which is developed on VS2008 with VC++ and implemented on an Intel Core i53470 3.20 GHz CPU with 3.46 GB RAM/ Windows X8632 bit machine. Figure 10 shows the system UI including the menu bar and the windows. The menu bar consists of the human guidance buttons and the configuration button. The windows are the live image, the model image, the curve diagram, and the last five target regions.
4.1. Performance
Supposing the scene where two following aircrafts (noted as FA1 and FA2) are tracking and recording the measured aircraft (MA) during 80 s by a 25 Hz camera with focal length from 150 mm to 300 mm and pixel size 6 m shown in Figure 11, we generate 2000 digital gray scale image sequences with resolution of . Dynamic parameter estimation needs six reference coordinates, which are the world coordinate system , the coordinate system of MA , the camera coordinate system of FA1 with the image coordinate system , and the camera coordinate system of FA2 with the image coordinate system . We first locate the four interest points of reference points of MA in the image coordinate then estimate the position and the attitude of MA in with known camera extrinsic parameters and coordinates of reference points.
4.1.1. Image Feature Location Precision
The four predefined interest points of the aircraft model are shown in Figure 12(b) marked with to . Figures 12(a) and 12(b) are the foreground detection result signed with bounding box and online interpretation result marked by crosses from two views.
(a) Result in left view
(b) One model
(c) Result in right view
The location precision of multiinterest points is shown in Figure 13. The true value coordinate of the target is (, ), and the interpretation coordinate is (, ), where is the sequence number of frames, is the label of the interpretation points, is the amount of interpretation frames, and is the number of interpretation points. Here equals 4, and location standard deviation is (0.3, 0.5) pixel for left view and (0.6, 0.4) pixel for right view on the basis of the following:
(a)
(b)
(c)
(d)
4.1.2. Dynamic Parameter Estimation Precision
The measurement precision about position and attitude is shown in Figure 14. The position is estimated by twoview triangulation with standard deviation (18.0, 13.6, and 24.1) m, and the attitude is estimated by absolute orientation with standard deviation (1.88, 1.00, and 3.34)° in one experiment, when there are 0.1° system error with 0.05° standard deviation of rotation angle, 10 m system error with 2 m standard deviation of translation about camera extrinsic parameters, and 0.5 m object error. The excellent filter like Kalman can be used to estimate the velocity and the angular velocity with position, attitude, and time information.
(a)
(b)
4.2. Efficiency
Figure 15 shows the execution time of the single frame about 12 ms/frame, and more initialization time is required than online processing. The location of multiinterest points needs an iterative optimization; for realtime, we not only use the frequency limitation, but also build shape samples library with some fixed step to ensure a superior initial value. In our experiment, because of the pattern of side direction observation, the number of iteration is less than 2.
(a)
(b)
5. Conclusion
This paper proposes a realtime processing method of optical tracking and measurement system for flying targets with a detailed image interpretation method and a brief parameter estimation review and realizes a distributed realtime image interpretation and parameter estimation system. The simulated experiments validate the accuracy and the efficiency of the system. The Achilles’ heel of the method and the system is not adapting to all the outdoor rigorous observation conditions, like uneven illumination and longtime occlusion, which seriously affect the feature location. Next we will put emphasis on robust feature location algorithms to improve the image interpretation precision and use more sophisticated real experiments for the system optimization. The system can be used to flying targets tests to meet the users’ need of realtime interpretation.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
Acknowledgment
This work was supported by the National Natural Science Foundation of China (Grant nos. 11072263 and 11272347) and the Program for New Century Excellent Talents in University (NCET120140).
References
 Image Systems, “TrackEye,” 2013, http://www.imagesystems.se/. View at: Google Scholar
 MIT, “Visual Fusion,” 2013, http://pages.swcp.com/~spsvs/VisualFusion/. View at: Google Scholar
 NAC, “Movias Pro,” 2013, http://www.nacinc.com/products/MotionAnalysisSoftware/. View at: Google Scholar
 M. Muja and D. G. Lowe, “Fast approximate nearest neighbors with automatic algorithm configuration,” in Proceedings of the International Conference on Computer Vision Theory and Applications (VISAPP '09), pp. 331–340, February 2009. View at: Google Scholar
 K. Grauman and B. Leibe, Visual Object Recognition, Morgan & Claypool Publishers, 2011.
 Z. Zivkovic, “Improved adaptive Gaussian mixture model for background subtraction,” in Proceedings of the 17th International Conference on Pattern Recognition (ICPR '04), vol. 2, pp. 28–31, August 2004. View at: Google Scholar
 O. Barnich and M. van Droogenbroeck, “ViBe: a universal background subtraction algorithm for video sequences,” IEEE Transactions on Image Processing, vol. 20, no. 6, pp. 1709–1724, 2011. View at: Publisher Site  Google Scholar
 D. Q. Sun, S. Roth, and M. J. Black, “Secrets of optical flow estimation and their principles,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR '10), pp. 2432–2439, San Francisco, Calif, USA, June 2010. View at: Publisher Site  Google Scholar
 Q. Yu and G. Medioni, “Motion pattern interpretation and detection for tracking moving vehicles in airborne video,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR '09), pp. 2671–2678, Miami, Fla, USA, June 2009. View at: Publisher Site  Google Scholar
 J. M. Kang, I. Cohen, G. Medioni, and C. Yuan, “Detection of independently moving objects in nonplanar scenes via multiframe monocular epipolar constraint,” in Proceedings of the 12th European Conference on Computer Vision (ECCV '12), pp. 860–873, 2012. View at: Publisher Site  Google Scholar
 Google, “Trimble 3D warehouse,” 2013, http://sketchup.google.com/3dwarehouse/. View at: Google Scholar
 J. Flusser and T. Suk, “Pattern recognition by affine moment invariants,” Pattern Recognition, vol. 26, no. 1, pp. 167–174, 1993. View at: Publisher Site  Google Scholar
 Z. Kalal, K. Mikolajczyk, and J. Matas, “Trackinglearningdetection,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 34, no. 7, pp. 1409–1422, 2011. View at: Google Scholar
 P. Y. Guo, X. Li, S. W. Ding, Z. H. Tian, and X. H. Zhang, “Adaptive and accelerated trackinglearningdetection,” in International Symposium on Photoelectronic Detection and Imaging 2013: Imaging Sensors and Applications, vol. 8908 of Proceedings of SPIE, Beijing, China, June 2013. View at: Publisher Site  Google Scholar
 S. Bernstein and R. Bernstein, Elements of Statistics II: Inferential Statistics, McGrawHill, New York, NY, USA, 1999.
 Y. G. Shi and W. C. Karl, “A realtime algorithm for the approximation of levelsetbased curve evolution,” IEEE Transactions on Image Processing, vol. 17, no. 5, pp. 645–656, 2008. View at: Publisher Site  Google Scholar
 R. G. Von Gioi, J. Jakubowicz, J. M. Morel, and G. Randall, “LSD: a fast line segment detector with a false detection control,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 32, no. 4, pp. 722–732, 2010. View at: Publisher Site  Google Scholar
 X. D. Hou and L. Q. Zhang, “Saliency detection: a spectral residual approach,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR '07), pp. 1–8, Minneapolis, Minn, USA, June 2007. View at: Publisher Site  Google Scholar
 H. Bay, A. Ess, T. Tuytelaars, and L. van Gool, “Speededup robust features (SURF),” Computer Vision and Image Understanding, vol. 110, no. 3, pp. 346–359, 2008. View at: Publisher Site  Google Scholar
 V. A. Prisacariu and I. D. Reid, “PWP3D: realtime segmentation and tracking of 3D objects,” International Journal of Computer Vision, vol. 98, no. 3, pp. 335–354, 2012. View at: Publisher Site  Google Scholar
 S. Belongie, J. Malik, and J. Puzicha, “Shape matching and object recognition using shape contexts,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 24, no. 4, pp. 509–522, 2002. View at: Publisher Site  Google Scholar
 P. J. Besl and N. D. McKay, “A method for registration of 3D shapes,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 14, no. 2, pp. 239–256, 1992. View at: Publisher Site  Google Scholar
 C. Rother, V. Kolmogorov, and A. Blake, ““GrabCut”—interactive foreground extraction using iterated graph cuts,” ACM Transactions on Graphics, vol. 23, no. 3, pp. 309–314, 2004. View at: Publisher Site  Google Scholar
 K. Cannos, “A review of visual tracking,” Tech. Rep., York University, 2008. View at: Google Scholar
 L. S. Liu, PostFlight Data Processing of Trajectory Measurement, National Defense Industry Press, 2000.
 R. Hartley and A. Zisserman, Multiple View Geometry in Computer Vision, Cambridge University Press, New York, NY, USA, 2003.
 C. P. Lu, G. D. Hager, and E. Mjolsness, “Fast and globally convergent pose estimation from video images,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 22, no. 6, pp. 610–622, 2000. View at: Publisher Site  Google Scholar
 V. Lepetit, F. MorenoNoguer, and P. Fua, “EPnP: an accurate O(n) solution to the PnP problem,” International Journal of Computer Vision, vol. 81, no. 2, pp. 155–166, 2009. View at: Publisher Site  Google Scholar
 Q. F. Yu, X. Y. Sun, and G. J. Chen, “A new method of measure the pitching and yaw of the axes symmetry object through the optical image,” Journal of National University of Defense Technology, vol. 22, no. 2, pp. 15–19, 2000. View at: Google Scholar
 K. P. Wang, X. H. Zhang, L. C. Li, and X. C. Liu, “A new method for measurement of rolling angle by optical image,” Journal of Projectiles, Rockets, Missiles and Guidance, vol. 29, no. 1, 2009. View at: Google Scholar
Copyright
Copyright © 2014 Pengyu Guo et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.