Research Article  Open Access
YuLong Qiao, KaiLong Yuan, ChunYan Song, XueZhi Xiang, "Detection of Moving Objects with Fuzzy Color Coherence Vector", Mathematical Problems in Engineering, vol. 2014, Article ID 138065, 8 pages, 2014. https://doi.org/10.1155/2014/138065
Detection of Moving Objects with Fuzzy Color Coherence Vector
Abstract
Background subtraction is a popular method for detecting foreground that is widely adopted as the fundamental processing for advanced applications such as tracking and surveillance. Color coherence vector (CCV) includes both the color distribution information (histogram) and the local spatial relationship information of colors. So it overcomes the weakness of the conventional color histogram for the representation of an object. In this paper, we introduce a fuzzy color coherence vector (FCCV) based background subtraction method. After applying the fuzzy cmeans clustering to color coherence subvectors and color incoherence subvectors, we develop a regionbased fuzzy statistical feature for each pixel based on the fuzzy membership matrices. The features are extracted from consecutive frames to build the background model and detect the moving objects. The experimental results demonstrate the effectiveness of the proposed approach.
1. Introduction
With the increasing significance and popularity of digital multimedia applications such as video surveillance, object tracking, action understanding, anomaly detection, and gait recognition, the detection of moving objects in video frames occupies a critical role in the subsequent tasks. A successful detection system with great accuracy promotes the overall performance of the further processing. Therefore, various methods have been proposed for detecting moving objects in the past years. However, robust detection of moving objects in presence of complex backgrounds is yet a challenging problem.
The basic detection method is based on frame difference. For consecutive frames of the video, it compares the intensities of the pixels at the same location and extracts the objects with an appropriate threshold. But this approach is sensitive to various disturbs, such as illumination changes, swaying vegetation, rippling water, and changes in the background geometry. When the number of frames in the image sequence is large and there is little change between consecutive frames, the background modeling and subtraction method is another solution for the detection problem [1].
The heart of a background subtraction method is background modeling that uses the new video frame to calculate and update a background model. This resulting background model provides a description of the entire background scene. The moving objects can be detected by identifying the pixels that are not consistent with the background model. Many different approaches have been proposed for modeling the background. The simple but often effective parametric approach is to model each pixel of a video frame with the single Gaussian distribution. The model parameters (mean and covariance) can be recursively updated with a simple adaptive filter [2]. This model works well for the static background. In the natural environments, even though the camera is fixed, the background appears to be dynamic because it always includes repetitive motions like swaying vegetation, rippling water, flickering monitors, and so forth. To model the dynamic background, the use of a single Gaussian becomes inappropriate and the mixture of Gaussians model (MoG) is preferable [3, 4]. When changes in the background are fast, the Gaussian assumption based modeling method is not suited. Therefore, a nonparametric approach was proposed for background modeling [5]. This method makes use of a general nonparametric kernel density estimation technique for building a statistical representation of the scene background. A major drawback of this approach is that it ignores the timeseries nature of the problem. Moreover, KDE requires training data from a sequence of examples that have a relatively “clean” background [6].
The pixelbased method is partially inspired by an observation that the background model will be well established if the background is static and there are sufficient training frames. Therefore, for the dynamic texture background and realtime applications, these pixelbased methods do not perform well. To overcome their limitation, researchers pay attention to the regionbased methods that usually divide an image into (overlapping or nonoverlapping) blocks and calculate block features, from which we can build the background model and detect the foreground. Mason and Duric [7] extracted a feature vector based on the edge histogram to describe a block and then to detect the moving objects. Heikkilä and Pietikäinen [2] proposed a blockbased background subtraction method based on the local binary pattern (LBP) texture measure. Each image block was modeled as a group of weighted adaptive LBP histograms. Zhang et al. [8] proposed the spatiotemporal local binary pattern (STLBP) to model dynamic textures. In another work [9], they furthermore used the covariance descriptor, defined based on various spatial and texture features, to efficiently suppress dynamic textures in the background. Recently, several authors have explored the adoption of fuzzy approaches to deal with different aspects of detecting moving objects. Sigari et al. [10] proposed a fuzzy version of running average method (FG) for background subtraction. Maddalena and Petrosino [11] introduced a fuzzy learning factor into the background model update procedure (FASOM). W. Kim and C. Kim [12] introduced a fuzzy color histogram (FCH) based block features for background subtraction. The background model is reliably constructed by computing the similarity between local FCH features with an online update procedure.
A color histogram provides only a distribution of the color. The FCH could attenuate color variations generated by background motions. However, color histogram is a very coarse image characterization so that images with similar histograms can have dramatically different appearances. Color coherence vector (CCV) does include both the color distribution information (histogram) and the local spatial relationship information. Therefore, this paper proposes a background subtraction method based on fuzzy color coherence vector that is derived from CCV by applying the fuzzy cmeans clustering. The local features extracted from the fuzzy color coherence vector are used to build and update the background model and detect the moving objects.
This paper is organized as follows. In Section 2, we provide a brief overview of color coherence vector and introduce the fuzzy color coherence vector. Section 3 presents the background modeling and moving objects detection method. To demonstrate the effectiveness of the proposed background subtraction method, the experimental results on various image sequences will be shown and analyzed in Section 4. Section 5 draws a conclusion for the paper.
2. Fuzzy Color Coherence Vector
2.1. Color Coherence Vector
Color is an essential and important property of an image; thus, it has been explored in various image processing problems including background subtraction. A color histogram can be regarded as the probability density function that describes the probability for pixels in the image to belong to a color bin. It is computationally efficient and generally insensitive to small changes in camera position and geometric transformation such as rotation. However, a color histogram provides only a distribution of the color. It is a very coarse image characterization so that images with similar histograms can have dramatically different appearances.
In order to overcome the limitations of the conventional color histogram, Pass and Zabih [13] propose a histogram refinement method and give a new color characterization, color coherence vector (CCV), that includes both the color distribution information (histogram) and the local spatial relationship information. As shown in Figure 1, there are two images with the same color histograms. However, they may represent different objects. Because the descriptor CCV covers spatial relationship information of the color of pixels, the two images can be distinguished with their CCVs.
For a usual RGB image, there are three color channels, and each color level is 256, which is of heavy computational burden to obtain CCV. In addition, the human visual perception system does not match with RGB color space very well. Therefore, we first convert RGB into HSV color space. In general, taking into account computational complexity and perceptual representation performance, the HSV image will be quantized into a gray image with bins (levels). This paper adopts the quantization schemes: 4 hues, step 90 (315, 45, 135, 225, 315), 4 saturations, step 0.25 (0, 0.25, 0.5, 0.75, 1.0), and 8 values, step 0.125 (0, 0.125, 0.25, …, 1.0), which will generate a gray image L_{HSV} with 128 bins [14]. In order to improve the detection performance, we also quantize value () into 128 bins resulting in another gray image L_{V} with 128 bins. The images L_{HSV} and L_{V} will be combined to detect the moving objects.
In the CCV framework, a pixel within a given color bin is classified as either coherent or incoherent. A coherent pixel is a part of a sizable contiguous region, while an incoherent pixel is not. Each CCV represents this classification for a color bin in the image. To determine the coherent pixel, we compute connected components using 8connected neighbors with a given color bin. A pixel is coherent if the size of its connected component exceeds a threshold; otherwise, the pixel is incoherent.
Suppose that and are the numbers of coherent and incoherent pixels in the th color bin of the histogram, respectively. Then the color histogram of the image with color bins is described as and the CCV of the image is defined as [13, 15]. The judging threshold for the coherence pixel is defined as follows: where denotes the number of pixels of the th individual coherent regions for the th color bin, is the total number of the regions, and is a weighting coefficient to control the accuracy of detection and is obtained from practical training. As shown in Figure 1, there are six pixels with green color in each image resulting in the same histograms. However, if we set the threshold to be 2, then the corresponding elements in the CCVs of the two images are and , respectively.
When an object is moving into a region, not only the local color histograms but also the local spatial relationships of the color of this region are changing. The CCV includes both the distribution and spatial relationship information of the color. Hence, the CCVs of L_{HSV} and L_{V} should give better clues for moving objects detection.
2.2. Fuzzy Color Coherence Vector
Based on the observation that background motions (dynamic texture) do not make severe alterations of the scene structure, W. Kim and C. Kim [12] adopt the fuzzy color histogram (FCH) to attenuate color variations generated by background motions and improve moving objects detection performance. The CCV is the refinement of the color histogram by considering the local spatial relationship. Therefore, this paper combines fuzzy cmeans clustering with CCV and introduces the fuzzy color coherence vector (FCCV) for background subtraction.
After obtaining the conventional CCVs of L_{HSV} and L_{V}, and , we apply the fuzzy cmean algorithm to the combined coherence subvector and obtain a membership matrix and clusters. The matrix describes the membership degree of every element belonging to each cluster. For conducting the FCM algorithm, given the initial membership values that are randomly generated and subject to the condition , , the initialization values of clustering centers are obtained by using (2), and then they are updated iteratively as [12], where is a constant and set to 2, which controls the spread degree among the fuzzy clusters. The clustering process stops when the maximum number of iterations 100 is reached, or when the objective function improvement between two consecutive iterations is less than the specified threshold. The same clustering method is applied to the combined incoherence subvector , and obtains a membership matrix .
Once we obtain the two fuzzy membership matrices, each pixel of the HSV image can be endued with the membership values to coherence clustering centers and incoherence clustering centers, respectively, by using the corresponding gray values of L_{HSV} and L_{V} images. Both the membership matrices and for the first video frame are calculated and stored in advance. They could be treated as two lookup tables. For each pixel of remaining video frames, we can obtain its membership values to coherence and incoherence clustering centers by using gray values of L_{HSV} and L_{V} images and tablelookup method without computing membership matrices.
In order to construct the background models, we define the regionbased feature FCCV that can be easily built by referring to the stored membership matrices and . For a video frame, we convert it into HSV color space and then quantize the HSV image into two grey images as the first frame. For a pixel at the position , it is assumed that the grey values of the L_{HSV} and L_{V} images are and , respectively. Then the membership value of this pixel belonging to the th coherence clustering centers is the th element of the matrix , while the membership value to the th incoherence clustering centers is the th element of the matrix . The regionbased features vector (FCCV) of the th pixel is defined as follows: where indicates the position set of the region centered at the location . is the membership value of the pixel located at the position to the th coherence (incoherence) clustering center. By comparing the feature vectors of the same pixel location in the consecutive frames, we can detect the moving objects and update the background model.
3. Background Subtraction Methods
The background model BF is initialized with the feature vector FCCV of each pixel of the first frame of the image sequence. For the current frame, we compute the FCCV for every pixel and compare it with the FCCV of the corresponding position of the background model by using some similarity measure. Then the pixel is determined as background () or foreground () objects as follows: where is a similarity measure defined as where is an empirical parameter to integrate the incoherent and coherent features to compare the similarity of the two descriptions. That is to say, when the higher similarity exists between a pixel of the current frame and the background model, this pixel is implied to be a background. Otherwise, the lower similarity suggests that the pixel should belong to the moving objects.
In the dynamic background, some change will take place in the background as time goes on. In order to take these changes into account and build a sound background model, the model should be online updated with the detection results of the current frame. The scheme for the th pixel is adopted as the following method [16]: where is the updating rate belonging to the coverage . The updated background model is used to detect the objects of the th frame.
The background subtraction algorithm is shown in Figure 2 and summarized as follows.
(A) Initialization Phase (for the first video frame, )(a)Convert RGB into HSV color space and quantize it into two gray images L_{HSV} and L_{V} with color bins.(b)Calculate the color coherence vectors of L_{HSV} and L_{V} and then apply FCM clustering method to the combined coherence subvector and incoherence subvector to obtain membership matrices and , respectively.(c)Derive the FCCV for each pixel using (3).(d)Initialize the background model () with the FCCV calculated in step (c).
(B) Detection Phase (for the th frame, )(a)Convert RGB into HSV color space and quantize it into two grey images as the initialization phase.(b)Calculate all feature vectors (FCCVs) for the th frame using (3).(c)Detect the moving object by masking each pixel with “1” and “0” using (4) and the background model .(d)Update the background model using (6).
4. Results and Discussions
In this section, we conduct the experiments on several benchmark image sequences. In order to fully demonstrate the detection performance, the results of the proposed algorithm are compared to typical background subtraction methods including frame difference (FD) [17], MoG [3, 17], FCH [12], FG [10, 17], and FASOM [11, 17]. The numbers of the quantized color bins and clustering centers are chosen as and , respectively. The updating weighting rate is set to be 0.01. The size of the local window for extracting FCCV is 5 × 5 pixels. In order to be fair, no postprocessing is used to improve the background subtraction performance.
The benchmark image sequences include FT, WS, CT and CA [18], WT [19], CJ [20], CE, and FT2 [21]. The frame numbers of original image sequence and ground truths are listed in Table 1. The selected original video frames, ground truths, and the results (foreground masks) of these methods are shown in Figures 3 and 4. It can be seen that the proposed method provides a reliable background model so that the dynamic background, such as waving leaves and turbulent water, is almost separated from the moving objects. The detection results suggest that the introduced method could effectively find the moving objects and it is better than the existing methods.

(a)
(b)
(c)
(d)
(e)
(f)
(g)
(h)
(a)
(b)
(c)
(d)
(e)
(f)
(g)
(h)
The foreground detection could be considered as binary classification of each pixel. With the ground truth, we can use the measures, recall and precision, to explain the correctness of this classification. Thus, they could be used to evaluate the performance of background subtraction methods: For the quantitative evaluation of our algorithm, we utilize a compositive measure, measure [22], to evaluate the effectiveness of our algorithm quantitatively. Higher values indicate better performance of the algorithm.
The measures of background subtraction algorithms for different image sequences are listed in Table 2. It is shown from the quantitative analysis that the proposed FCCV provides higher values on the average, which indicates better performance of the introduced method.

In our experiments, the foreground pixel is determined by using (4) with the empirical parameter . In order to further evaluate the proposed algorithm, we also analyze the detection performance with different values of the parameter . The results for the image sequences WT, WS, and CA are shown in Figure 5. It can be seen that the detection performance has acceptable change when the parameter lies in the interval , from which we can choose a proper parameter value for detecting the moving objects.
5. Conclusions
The detection of moving objects in video frames occupies a critical role in lots of applications. The traditional color histogram provides important color distribution information for moving object detection. However, there is a lack of the local spatial relationship information of colors, which could be captured by CCV. Therefore, this paper proposes a background subtraction algorithm based on FCCV using fuzzy mean clustering and CCV, in which the introduced feature vector FCCV for each pixel is used to build background model and detect the moving objects. In addition, the fuzzy property of the FCCV makes the algorithm robust to varying changes of environment, so the method could detect the moving objects in the dynamic background. The experimental results on some benchmark datasets demonstrate that the approach achieves better detection performance than several existing methods.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
Acknowledgments
This work is partially supported by the National Natural Science Foundation of China 61371175, Postdoctoral ScienceResearch Developmental Foundation of Heilongjiang Province LBHQ09128, and Fundamental Research Funds for the Central Universities HEUCFQ1411. The authors also gratefully acknowledge the helpful comments and suggestions of the reviewers, which have improved the presentation.
References
 A. Bugeau and P. Pérez, “Detection and segmentation of moving objects in complex scenes,” Computer Vision and Image Understanding, vol. 113, no. 4, pp. 459–476, 2009. View at: Publisher Site  Google Scholar
 M. Heikkilä and M. Pietikäinen, “A texturebased method for modeling the background and detecting moving objects,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 28, no. 4, pp. 657–662, 2006. View at: Publisher Site  Google Scholar
 C. Stauffer and W. E. L. Grimson, “Adaptive background mixture models for realtime tracking,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR '99), pp. 246–252, Ft. Collins, Colo, USA, June 1999. View at: Google Scholar
 N. Friedman and S. Russell, “Image segmentation in video sequences: a probabilistic approach,” in Proceedings of the 13th Conference on Uncertainty in Artificial Intelligence, pp. 175–181, Rhode Island, RI, USA, 1997. View at: Google Scholar
 A. Elgammal, D. Harwood, and L. Davis, “Nonparametric model for background subtraction,” in Proceedings of the European Conference on Computer Vision, pp. 751–767, Dublin, Ireland, June 2000. View at: Google Scholar
 L. Cheng, M. Gong, D. Schuurmans, and T. Caelli, “Realtime discriminative background subtraction,” IEEE Transactions on Image Processing, vol. 20, no. 5, pp. 1401–1414, 2011. View at: Publisher Site  Google Scholar
 M. Mason and Z. Duric, “Using histograms to detect and track objects in color video,” in Proceedings of the 30th Applied Imagery Pattern Recognition Workshop, pp. 154–159, Washington, DC, USA, October 2001. View at: Google Scholar
 S. Zhang, H. Yao, and S. Liu, “Dynamic background modeling and subtraction using spatiotemporal local binary patterns,” in Proceedings of the IEEE International Conference on Image Processing (ICIP '08), pp. 1556–1559, San Diego, Calif, USA, October 2008. View at: Publisher Site  Google Scholar
 S. Zhang, H. Yao, S. Liu, X. Chen, and W. Gao, “A covariancebased method for dynamic background subtraction,” in Proceedings of the 19th International Conference on Pattern Recognition (ICPR '08), 4, p. 1, Tampa, Fla, USA, December 2008. View at: Google Scholar
 M. H. Sigari, N. Mozayani, and H. R. Pourreza, “Fuzzy running average and fuzzy background subtraction: concepts and application,” International Journal of Computer Science and Network Security, vol. 8, no. 2, pp. 138–143, 2008. View at: Google Scholar
 L. Maddalena and A. Petrosino, “A fuzzy spatial coherencebased approach to background/foreground separation for moving object detection,” Neural Computing and Applications, vol. 19, no. 2, pp. 179–186, 2010. View at: Publisher Site  Google Scholar
 W. Kim and C. Kim, “Background subtraction for dynamic texture scenes using fuzzy color histograms,” IEEE Signal Processing Letters, vol. 19, no. 3, pp. 127–130, 2012. View at: Publisher Site  Google Scholar
 G. Pass and R. Zabih, “Histogram refinement for contentbased image retrieval,” in Proceedings of the 3rd IEEE Workshop on Applications of Computer Vision (WACV '96), vol. 10, pp. 96–102, December 1996. View at: Google Scholar
 S. Niranjanan and S. P. R. Gopalan, “Performance efficiency of quantization using HSV colour space and intersection distance in CBIR,” International Journal of Computer Applications, vol. 42, no. 21, pp. 48–55, 2012. View at: Google Scholar
 C. Huang and G. Wang, “Method of image retrieval based on color coherence vector,” Computer Engineering, vol. 32, no. 2, pp. 194–199, 2006. View at: Google Scholar
 P. Chiranjeevi and S. Sengupta, “New fuzzy texture features for robust detection of moving objects,” IEEE Signal Processing Letters, vol. 19, no. 10, pp. 603–606, 2012. View at: Publisher Site  Google Scholar
 A. Sobral, “BGSLibrary: a background subtraction library,” in Proceedings of the 9th Workshop de Visão Computacional (WVC '13), Rio de Janeiro, Brazil, 2013. View at: Google Scholar
 L. Y. Li, W. M. Huang, I. Y.H. Gu, and Q. Tian, “Statistical modeling of complex backgrounds for foreground object detection,” IEEE Transactions on Image Processing, vol. 13, no. 11, pp. 1459–1472, 2004. View at: Publisher Site  Google Scholar
 K. Toyama, J. Krumm, B. Brumitt, and B. Meyers, “Wallflower: principles and practice of background maintenance,” in Proceedings of the 7th IEEE International Conference on Computer Vision (ICCV '99), pp. 255–261, Kerkyra, Greece, September 1999. View at: Google Scholar
 Y. Sheikh and M. Shah, “Bayesian modeling of dynamic scenes for object detection,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 27, no. 11, pp. 1778–1792, 2005. View at: Publisher Site  Google Scholar
 N. Goyette, P. M. Jodoin, F. Porikli, J. Konrad, and P. Ishwar, “Changedetection.net: a new change detection benchmark dataset,” in Proceedings of the IEEE Workshop on Change Detection (CDW '12), pp. 1–8, 2012. View at: Google Scholar
 S. Brutzer, B. Höferlin, and G. Heidemann, “Evaluation of background subtraction techniques for video surveillance,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR '11), pp. 1937–1944, Colorado Springs, Colo, USA, June 2011. View at: Publisher Site  Google Scholar
Copyright
Copyright © 2014 YuLong Qiao et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.