Research Article | Open Access
Yanshuang Hao, Yixin Yin, Jinhui Lan, "Vehicle Tracking Algorithm Based on Observation Feedback and Block Symmetry Particle Filter", Journal of Electrical and Computer Engineering, vol. 2014, Article ID 520342, 8 pages, 2014. https://doi.org/10.1155/2014/520342
Vehicle Tracking Algorithm Based on Observation Feedback and Block Symmetry Particle Filter
This paper proposes a novel particle filter algorithm for vehicle tracking, which feeds observation information back to state model and integrates block symmetry into observation model. In view of the proposal distribution in traditional particle filter without considering the observation data, a new state transition model which takes the observation into account is presented, so that the allocation of particles is more familiar with the posterior distribution. To track the vehicles in background with similar colors or under partial occlusion, block symmetry is proposed and introduced into the observation model. Experimental results show that the proposed algorithm can improve the accuracy and robustness of vehicle tracking compared with traditional particle filter and Kernel Particle Filter.
In recent years, the traffic pressure in some cities has become higher and higher, and the role of the intelligent transportation system (ITS) is more and more important. Video sensors are particularly important in traffic applications due to their fast response, easy installation, operation, and maintenance, and their ability to provide wide area monitoring allowing analysis of traffic flow, speed measurement, vehicle classification, and incident detection . Moving vehicle tracking is the key technology in the video-based traffic monitoring system .
Particle filter is a kind of popular and robust approach for vehicle tracking. It is also called Sequential Monte Carlo method  or Condensation algorithm . Different from the Kalman filter, which can only solve the linear, Gaussian system state estimation problems, the particle filter does not suppose system model to be linear and can deal with nonlinear non-Gaussian distribution state model. It is a method based on Monte Carlo theory and Recursive Bayesian Estimation . The particle filtering predicts a new set of particles using a proposal distribution, updates the weight for every particle using an observation function, and estimates the posterior probability density of the states by computing the weighted sum of the set of particles. The propagation of particles and the updating of weights are very important steps during the particle filtering.
In order to get a preferable proposal distribution, Chang and Ansari  proposed the Kernel Particle Filter (KPF) for visual tracking, in which particles are allocated based on the gradient of the kernel density to form a continuous estimate of the posterior density function. The gradient estimation and particle allocation are implemented by the mean shift algorithm. Rui and Chen  introduced Unscented Particle Filter (UPF) into audio and visual tracking, which used Unscented Kalman Filter to generate sophisticated proposal distribute that seamlessly integrates the current observation. KPF and UPF can improve the tracking performance but increase the computation complexity. Zhu et al.  combined particle filter and grey prediction model and proposed a novel visual tracking algorithm, called GMPF. GMPF utilized the history of state estimation sequence to set up grey prediction model for generating proposal distribution. Compared with UPF, the GMPF consumed less time per frame but presented worse results . Wei et al.  proposed a layered particle filter (LPF) architecture for vehicle tracking. The strategy of LPF is dividing the particle set into two subsets randomly first, then predicting a new subset via state transition equation, selecting the particle with largest weight in another subset and predicting another new subset at the new position obtained via Camshift algorithm, and finally merging the two subsets.
In process of weights updating, the observation model combining multiple features was often used. Wei et al.  built a fusion observation model with color and local integral orientation descriptor. Tian et al.  used the weighted color histogram and rotated complex wavelet filter to describe color and texture features, respectively, and fused the two features under particle filter framework to improve the robustness and accuracy of face tracking. Rezaee et al.  fused color, edge, texture, and motion cues in particle filter algorithm for vehicle tracking. Maggio et al.  proposed a multifeature tracking algorithm that adaptively weights the reliability of each feature in particle filtering framework. The target representation is based on color and orientation histograms. Gao et al.  constructed proposal distribution by use of state partition technique and parallel extended kalman filter and fused adaptively color model and shape model in particle filter framework to enhance the performance of object tracking. Chan et al.  proposed a vision-based automatic system to detect multiple vehicles under various lighting conditions with a single particle filter. The system fuses four cues including underneath shadow, vertical edge, symmetry, and taillight. With BSAS clustering technique and initial data-driven sampling, multiple vehicles can be tracked with only one set of particles. Niknejad et al.  proposed a method for multiple vehicles tracking combining a deformable object model with particle filter. The likelihood of particles is updated by integrating the detection scores map and intensity correlations for vehicle’s root and parts between two sequential frames.
In this paper, the novelty of the proposed algorithm lies in a new state model fed back observation information and an observation model based on block symmetry. The rest of the paper is organized as follows. Section 2 introduces the normal particle filter framework. Section 3 proposes a novel particle state transition model. In Section 4, the block symmetry observation model based on color histogram and block symmetry are given. Section 5 presents the algorithmic description of the proposed particle filter. The experimental results are discussed in Section 6. Finally, Section 7 concludes the paper.
2. Review of Particle Filter Algorithm
A nonlinear, non-Gauss system can be represented by the state equation and the observation equation as follows : where , denote the system state vector and the state noise vector at time , respectively, and , denote the system observation vector and the observation noise vector. Particle filtering is a method to approximate the state posterior distribution by finite weighted particles.
At the beginning, the initial set of particles is generated from the priori density , and every particle is assigned an equal weight.
At time , the new set of particles is derived from the proposal distribution , which is also called the importance function. Supposing that the states of particles comply with a first order Markov process and the observations are independent of each other, the weight of the particle at time can be updated by  where is the likelihood function defined by the observation equation (2) and is the state transition density distribution described by the state equation (1). Normalize the weights:
According to the theory of Monte Carlo Integration, the unknown particle state at time can be estimated as follows:
The resampling step is implemented to eliminate the particles with weak weights and reduce particle degeneracy problem .
3. State Model
The normal particle filters select the state transition density distribution as the proposal distribution , that is, predict the new particles according to the state equation (1). Equation (3) can be simplified as
This method is simple and effective but requires an accurate model-of-state transition. Due to not considering the observation information, the proposal distribution cannot represent the true distribution effectively when the state equation is not exact enough or the observation noises become larger . To solve the problem, this paper proposes a new model-of-state transition which takes the observation information into account during the process of particle propagation.
Let us represent the target vehicle area with a rectangle centered in , with the width of and the height of . Then, we define the state vector of the target as .
The state equation is designed as follows:
In (7), means the estimate of target center displacement from the frame to the frame ; is the coefficient matrix of , and is a Gaussian noise vector with zero mean. The displacement of target is updated adaptively by (8), in which is the estimate of target center at frame and is the similarity between the target color histogram and the color histogram of the estimated state .
Karavasilis et al.  updated the translation of target using the Earth Mover’s Distance (EMD) between the source and target GMM. In (8), the displacement of target is evaluated by employing the similarity . A similarity close to 1 indicates that the estimated state has a small error with the true state at the frame , while a similarity close to 0 indicates a large error. Therefore, the displacement of target center at the th frame will be mainly determined by the difference between and if is close to 1. On the other hand, the displacement will be mainly decided by if is close to 0. Obviously, the estimated state includes the observation information, so the displacement is a function of the observation. Consequently, our proposed model-of-state transition also considers the observation information.
The size of vehicle in a video sequence increases when the vehicle moves towards the camera, and it decreases while the vehicle gets far way, apparently. The real distance between a vehicle and the camera is reflected in the vertical coordinate of the vehicle in the video image. The size of rectangle area can be corrected by the displacement of target in vertical direction and then the coefficient matrix can be defined as where , are constants, whose values vary with the angle between the camera and road.
4. Observation Model
The color histogram is invariant to the scale and orientation changes of the target and is robust to partial occlusions but cannot provide geometry information of the target , which makes it difficult to track the target in the background with similar colors. Considering that most of vehicles keep a good symmetry, this paper integrates the color histogram and the symmetry measurements to design observation model for vehicle tracking.
4.1. Color Histogram
The RGB color space can be quantized into bins, and the color histograms of the reference target region is represented by . At time , the color histograms of the candidate target region defined by particle can be denoted by . The Bhattacharyya coefficient  is used to measure the similarity between target color histogram and candidate histogram as follows:
The distance between the two histograms can be calculated on the basis of the Bhattacharyya coefficient :
The observation model based on color feature is defined as where denotes the normal standard deviation set by experience.
4.2. Block Symmetry
Assume that the state of the particle at time is and denotes the gray value of the pixel on row and column in the image of frame . We compute the symmetry within the rectangle region defined by particle state using the symmetry model proposed in .
The mean gray value of the pixels on row in the rectangle region can be expressed by :
Then we calculate the following two functions for , :
The symmetry property of the row in the rectangle region can be computed as follows :
The range of is the interval . If the gray scale image on the row has a perfect symmetry, should be equal to 1. On the other hand, shows an ideal asymmetry.
Figure 1 gives the example of symmetry computing, and Figure 1(a) shows two rectangle regions to be extracting the symmetry, the red one is symmetry and the blue one is asymmetry. Figure 1(b) gives the symmetry of every row in the two rectangle regions.
(a) Two rectangle regions
(b) Symmetry of every row
Different from , we propose the block symmetry-based observation model to accurately track the vehicles under partial occlusion. When the vehicle tracked is under partial occlusion, it can be divided into two parts, the occluded part and the unoccluded part. The former no longer has symmetry, while the latter still keeps the symmetry. Therefore, we split the rectangle region defined by particle into the upper half and lower half and calculate the symmetries of the upper half and lower half, respectively, as follows:
If both and are less than a given threshold , the rectangle area is considered asymmetry. If either or is greater than or equal to the threshold , the rectangle area is regarded as a region with mirror symmetry about a vertical axis. Thus, we define the global symmetry of the rectangle region as where the threshold is a value approximate to but less than 0 (e.g., −0.05), which is selected by experiment. The range of is the interval and 0.
The observation model based on symmetry feature is defined as
4.3. Combination of Color and Block Symmetry
The observation model which fused color histogram and symmetry feature is defined as follows:
If the candidate target region defined by particle has no symmetry, and the observation density . If the symmetry of the candidate target region is acceptable, the observation density is modified by multiplying the symmetry measurement .
5. Proposed Algorithm
The complete vehicle tracking algorithm proposed is presented below.
Given the target color histogram , consider particle set at time and its weights , estimation of state at time , and the similarity and the estimate of target displacement at time .
Step 1. Generate a predicted set at time according to (7).
Step 3. Estimate the state of target at time by .
Step 4. Calculate the similarity and displacement at time by (8).
Step 5. Compute the number of effective sample by and resample the set of particles using the importance sampling method if is less than a threshold.
Step 6. If the frame is not the last one, return to Step 1; otherwise end the algorithm.
6. Experimental Results
In order to test the tracking performance of the algorithm proposed above, video sequences obtained from real roads are used. The camera is fixed on pedestrian overpasses. The experiments are performed by MATLAB on PC with a 2.1 GHz Intel(R) Core (TM) 2 Duo CPU and a 2 GB memory. In all of the following experiments, we locate the tracked vehicle in first frame manually. The standard deviations of the Gaussian noise vector of (7) are set to , and the standard deviation of (12) is set to .
6.1. Tracking with Different State Models
Under particle filter framework, the state models in common use are the random walking model [18, 19] and the second-order linear regression model [8, 20]. The former can be represented as , where is a multivariate zero-mean Gaussian random variable. The latter can be expressed as .
Figure 2 shows the tracking results of three particle filter algorithms with random walking model, 2nd order linear model, and proposed state model, respectively. The observation models of three methods are based on color histogram only. The image sizes of the video sequence are 640 × 480. In this video, the road condition is simple, but the target vehicle moves fast. The number of particles is 60 in each method, and the tracking results are marked with red rectangle bounding boxes.
(a) Random walking model (Frames 2, 11, 21, and 26 are shown)
(b) Second-order linear regression model (Frames 2, 11, 21, and 26 are shown)
(c) Proposed state model (Frames 2, 11, 21, and 26 are shown)
Figure 2(a) shows that the method using a random walking model takes on an obvious deviation of the estimated vehicle position at frame 11 and loses the target vehicle at frame 26. Figure 2(b) gives the results using second-order linear regression model, in which errors of target size appear from frame 21 and deflection of target position appears at frame 26. The tracking results of proposed state model are presented in Figure 2(c). Because the displacement of target between adjacent frames and the observation information are considered during the particle propagation, the tracking results can keep up with the true target in time. The sizes of the tracking windows are changed adaptively due to a suitable coefficient matrix , as shown in Figure 2(c).
6.2. Tracking with Different Observation Models
To verify the effect of the observation model proposed, we compare it with the observation model based on color histogram only by using a video sequence, in which two vehicles with similar colors are close to each other. Figure 3(a) shows the tracking results with the observation model based on color histogram, and Figure 3(b) presents the results with the proposed observation model. The state models of two approaches are both the model proposed in Section 3.
(a) Observation model based on color histogram only (Frames 2, 10, 20, and 60 are shown)
(b) Proposed observation model (Frames 2, 10, 20, and 60 are shown)
From Figure 3(a), we can see that the color-histogram based approach cannot locate the target accurately, and the tracking results wander back and forth between the target vehicle and adjacent vehicle because of similar color interference. In contrast, the proposed approach can track the target vehicle stably owing to taking advantage of symmetry, as shown in Figure 3(b).
6.3. Tracking under Partial Occlusion
In this experience, we select a video sequence, in which a car is partially occluded by a white minibus when the minibus moves from the middle lane to the inside lane. Figure 4 presents the tracking results of two approaches with same state models and different observation models. Figure 4(a) employs the symmetry particle filter, which fuses the symmetry feature, not block symmetry, and color histogram. Figure 4(b) applies the proposed method. The number of particles is 60 in both methods.
(a) Symmetry particle filter (Frames 10, 30, 50, and 80 are shown)
(b) Proposed approach (Frames 10, 30, 50, and 80 are shown)
From Figure 4, we can see that the symmetry particle filter gives the results with distinct deflection compared to the proposed approach at frame 10 and frame 30. To evaluate the tracking accuracy, the deviation from the vehicle real position is calculated by  where (, ) and (, ) mean the estimated center position of tracked vehicle and the true center of the vehicle at frame , respectively.
The deviations in frames 1 to 90 by symmetry particle filter and the proposed approach are computed and presented in Figure 5. The 90 frames can be divided into two periods, frame 1–40 (asymmetry occlusion), frame 41–90 (symmetry occlusion). Figure 5 indicates that the deviations by symmetry particle filter are larger than those by the proposed approach during the period of asymmetry occlusion and become comparable during the period of symmetry occlusion. The results of Figures 4 and 5 show that the using of block symmetry can improve the tracking accuracy.
6.4. Compare with Kernel Particle Filter
In this example, the particle filter algorithm proposed is compared with Kernel Particle Filter (KPF)  and the traditional particle filter. In the approach of Kernel Particle Filter, we choose the random walking model as the state model, the color histogram as the observation model and quantize the RGB color space into bins. In the procedure of gradient estimation, each particle is moved to its mean using the mean shift method.
The image sizes of the video sequence are ; the number of particles is 60 in three methods, traditional particle filter, KPF, and proposed algorithm. The results shown in Figure 6 indicate that both the proposed algorithm and KPF can track vehicles in the video, while the traditional PF fails to track the target at frame 90. In addition, the sizes of tracking windows obtained by the proposed algorithm are more close to the true sizes of target vehicle than those obtained by KPF.
(a) Traditional Particle Filter (Frames 2, 40, 60, and 90 are shown)
(b) Kernel Particle Filter (Frames 2, 40, 60, and 90 are shown)
(c) Proposed algorithm (Frames 2, 40, 60, and 90 are shown)
We also measure the average time consumptions per frame of the two algorithms with different numbers of particles. The results are presented in Figure 8. Obviously, the KPF spends more time than our approach because of the procedure of gradient estimation for every particle.
Figure 8 demonstrates that the more particles are allocated the more time isconsumed. On the other hand, the less particles are allocated, the worse tracking accuracy will be. Thus we compute and compare the deviation in the different number of particles by the proposed approach to determine a suitable number of particles. The results are shown in Figure 9. In this example, the proposed particle filter with 30 particles only can accurately track vehicles in real time.
This paper proposes a novel state transition model based on observation feedback and an observation model which fused block symmetry and color features in the particle filter framework. During the process of particles propagation, the latest observation information is employed to update the displacement of particles, and size of target region defined by particle is corrected by the particle displacement on vertical direction. In order to improve the tracking performance when the tracked vehicle is under occlusion, the block symmetry is put forward. The observation model takes color histogram and block symmetry measurements as observation features. Experimental results illustrate the effectiveness of the observation feedback and block symmetry particle filter.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
The work is supported by Coconstruction Project of Beijing, 2013-Major Construction Item, and National Natural Science Foundation of China (Grant no. 61174181).
- V. Kastrinaki, M. Zervakis, and K. Kalaitzakis, “A survey of video processing techniques for traffic applications,” Image and Vision Computing, vol. 21, no. 4, pp. 359–381, 2003.
- T. Gao, Z.-G. Liu, and J. Zhang, “Motion vehicle recognition and tracking in the complex environment,” in Proceedings of the International Seminar on Future Information Technology and Management Engineering (FITME '08), pp. 278–282, November 2008.
- A. Doucet, S. Godsill, and C. Andrieu, “On sequential Monte Carlo sampling methods for Bayesian filtering,” Statistics and Computing, vol. 10, no. 3, pp. 197–208, 2000.
- E. B. Koller-Meier and F. Ade, “Tracking multiple objects using the Condensation algorithm,” Robotics and Autonomous Systems, vol. 34, no. 2-3, pp. 93–105, 2001.
- C. Chang and R. Ansari, “Kernel particle filter for visual tracking,” IEEE Signal Processing Letters, vol. 12, no. 3, pp. 242–245, 2005.
- Y. Rui and Y.-Q. Chen, “Better proposal distributions: object tracking using unscented particle filter,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. II786–II793, December 2001.
- M.-Q. Zhu, Z.-L. Wang, and Z.-H. Chen, “Visual tracking algorithm based on grey prediction model and particle filter,” Control and Decision, vol. 27, no. 1, pp. 53–57, 2012.
- Q. Wei, Z. Xiong, C. Li, Y.-X. Ouyang, and H. Sheng, “A robust approach for multiple vehicles tracking using layered particle filter,” International Journal of Electronics and Communications, vol. 65, no. 7, pp. 609–618, 2011.
- H. Tian, T.-Z. Shen, S.-Y. Zhao, and B. Hao, “Face tracking algorithm combing color and texture features based on particle filter,” Transaction of Beijing Institute of Technology, vol. 30, no. 4, pp. 469–473, 2010.
- H. Rezaee, A. Aghagolzadeh, and H. Seyedarabi, “Vehicle tracking by fusing multiple cues in structured environments using particle filter,” in Proceedings of the Asia Pacific Conference on Circuit and System (APCCAS '10), pp. 899–902, December 2010.
- E. Maggio, F. Smerladi, and A. Cavallaro, “Adaptive multifeature tracking in a particle filtering framework,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 17, no. 10, pp. 1348–1359, 2007.
- S.-W. Gao, L. Guo, N. Yang, L. Chen, and Y.-Q. Du, “New particle filter object tracking algorithm,” Journal of Shanghai Jiaotong University, vol. 43, no. 3, pp. 485–489, 2009.
- Y.-M. Chan, S.-S. Huang, L.-C. Fu, P.-Y. Hsiao, and M.-F. Lo, “Vehicle detection and tracking under various lighting conditions using a particle filter,” IET Intelligent Transport Systems, vol. 6, no. 1, pp. 1–8, 2012.
- H. T. Niknejad, A. Takeuchi, S. Mita, and D. McAllester, “On-road multivehicle tracking using deformable object model and particle filter with improved likelihood estimation,” IEEE Transaction on Intelligent Transportation Systems, vol. 13, no. 2, pp. 748–758, 2012.
- S. Maskell and N. Gordon, “A tutorial on particle filters for on-line nonlinear non-Gaussian Bayesian tracking,” in Proceedings of the IEE Workshop Target Tracking: Algorithms and Applications, vol. 2, pp. 1–15, 2001.
- V. Karavasilis, C. Nikou, and A. Likas, “Visual tracking using the Earth Mover's Distance between Gaussian mixtures and Kalman filtering,” Image and Vision Computing, vol. 29, no. 5, pp. 295–305, 2011.
- D. Comaniciu, V. Ramesh, and P. Meer, “Kernel-based object tracking,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 25, no. 5, pp. 564–577, 2003.
- H. Liu, F. Sun, and K. He, “Symmetry-aided particle filter for vehicle tracking,” in Proceedings of the IEEE International Conference on Robotics and Automation (ICRA '07), pp. 4633–4638, Rome, Italy, April 2007.
- W. Ye, H. Liu, F. Sun, and M. Gao, “Vehicle tracking based on co-learning particle filter,” in Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS '09), pp. 2979–2984, St. Louis, Mo, USA, October 2009.
- G.-C. Liu and Y.-J. Wang, “An algorithm of multi-target tracking based on improved particle filter,” Control and Decision, vol. 24, no. 2, pp. 317–320, 2009.
Copyright © 2014 Yanshuang Hao et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.