Journal of Electrical and Computer Engineering

Journal of Electrical and Computer Engineering / 2013 / Article

Research Article | Open Access

Volume 2013 |Article ID 598708 | https://doi.org/10.1155/2013/598708

Farah Yasmin Abdul Rahman, Aini Hussain, Wan Mimi Diyana Wan Zaki, Halimah Badioze Zaman, Nooritawati Md Tahir, "Enhancement of Background Subtraction Techniques Using a Second Derivative in Gradient Direction Filter", Journal of Electrical and Computer Engineering, vol. 2013, Article ID 598708, 12 pages, 2013. https://doi.org/10.1155/2013/598708

Enhancement of Background Subtraction Techniques Using a Second Derivative in Gradient Direction Filter

Academic Editor: Vijayan K. Asari
Received27 Mar 2013
Revised05 Sep 2013
Accepted24 Sep 2013
Published11 Nov 2013

Abstract

A new approach was proposed to improve traditional background subtraction (BGS) techniques by integrating a gradient-based edge detector called a second derivative in gradient direction (SDGD) filter with the BGS output. The four fundamental BGS techniques, namely, frame difference (FD), approximate median (AM), running average (RA), and running Gaussian average (RGA), showed imperfect foreground pixels generated specifically at the boundary. The pixel intensity was lesser than the preset threshold value, and the blob size was smaller. The SDGD filter was introduced to enhance edge detection upon the completion of each basic BGS technique as well as to complement the missing pixels. The results proved that fusing the SDGD filter with each elementary BGS increased segmentation performance and suited postrecording video applications. Evidently, the analysis using F-score and average accuracy percentage proved this, and, as such, it can be concluded that this new hybrid BGS technique improved upon existing techniques.

1. Introduction

Object extraction is a technique used in suppressing the background of a video scene to detect subjects that appear in the frame. The technique involves comparing or subtracting the current frames from the background frame and treating the remaining pixels as foreground [1]. Prior research on background subtraction (BGS) used several parametric BGS techniques, such as running average [24], running Gaussian average [57], approximate median filter [7, 8], and Gaussian Mixture Model [911]. These parametric techniques determine the foreground and update the subsequent background based on the distribution of intensity value [12]. Aside from these techniques, other studies have introduced nonparametric models that detect foreground and background based on the intensity of statistical properties [13]. Other non-parametric models include a kernel density estimator [14] and mean shift estimation [15].

This work focuses on basic BGS techniques, such as frame differencing, approximate median, running average, and running Gaussian average. The motivation of this work lies in the fact that most edge pixels are undetected after performing object extraction techniques based on FD, AM, RA, and RGA. In this study, however, we have overcome this limitation by detecting all edge pixels; hence a perfect blob can be retrieved through morphological procedures.

This is done by applying an SDGD filter on the results of background suppression and combining the foreground pixels generated by BGS techniques with the detected edge as our extracted object. The edge pixels are expected to fill in the boundary gap, which creates better connections among pixels in the boundary. This leads to better foreground frame detection.

This paper is organized as follows. Section 2 presents an overview of several basic BGS techniques and the SDGD filter. Section 3 describes the methodology. Section 4 discusses the results. Finally, Section 5 concludes our paper.

2. Literature Review

This section provides a review of the literature on the four BGS techniques evaluated in this study, namely, frame differencing, approximate median, running average, and running Gaussian average. SDGD filter studies are also presented.

2.1. Frame Differencing

Frame differencing (FD) is the most fundamental technique in BGS. FD involves finding the absolute difference between the current frame and a previous or background frame [1]. The absolute difference is then compared with an appropriate threshold value, , to detect the object as shown in (1), where is the current frame intensity value, is the background intensity value, and is the foreground intensity value. This technique uses the same background frame for all video sequences:

2.2. Approximate Median Filter

The approximate median (AM) algorithm is adaptive, dynamic, nonprobabilistic, and intuitive [8]. AM is obtained by calculating the difference between two video frames and using this difference in determining the perfect method for updating the background. AM is considered one of the most acceptable methods because it provides the most accurate pixel identification.

Several studies have evaluated the efficacy of the AM algorithm. He et al. [7] tested the effectiveness of the AM algorithm as part of their optimized algorithm for vehicle detection in an embedded system. Their approach yields highly accurate information with less computational time when detecting and tracking vehicles in a traffic scene. Equation (2) presents AM updates in the reference frame of every video sequence. The succeeding background frame, , is dependent on the intensity value of both the present frame, , and background frame, :

2.3. Running Average

Running average (RA) is another technique for updating a background image. A pixel is classified as background when the pixel value belongs to the corresponding distribution of the background model, and if otherwise, the mean of the distribution is updated [4]. The updated image is then used in the changing scene. The computational effect in an RA background is lesser because only the weighted sums of the two images are computed, and thus, low computational and space complexities are produced [3]. Moreover, several researchers have utilized this method to detect moving objects in video captured by a static camera.

Several studies were conducted to enhance the efficiency of BGS based on the RA method [24]. The outcome by Park et al. in [4] showed that application of a hierarchical data structure significantly increased the processing speed with accurate motion detection. This outcome can be attributed to the updating of the background frame by the RA method. Equation (3) shows a specified learning rate based on the previous background frame, where is the learning rate and is the threshold value:

2.4. Running Gaussian Average

This method combines both the Gaussian function and RA. Overall, the running Gaussian average (RGA) method has a significant advantage over other approaches because it requires less processing time and utilizes less memory compared with nonrecursive methods such as mixture of Gaussian (MoG) and kernel density estimation (KDE) [9]. Equation (4) shows how the reference frame that is represented by the mean is updated in each video sequence using this method. Unlike AM and RA, this method uses as a threshold value:

2.5. Second Derivative in a Gradient Direction (SDGD) Filter

In image processing studies, researchers use first- and second-order derivatives to detect the edge of an object based on its gradient. By using the first derivative, the edge location is defined at the maximum position of the steep and the descent [16]. Traditional edge detection methods, such as those by Prewitt, Sobel, and Roberts, convolved the image with a specific kernel [16, 17]. However, these techniques were reported to be sensitive to noise and inaccurate [17]. In 1986, the Canny edge detector was introduced, which represented an improvement over the traditional methods [17, 18]. The detector applied Gaussian smoothing to reduce noise, unwanted details, textures, nonmaximum suppression, and hysteresis thresholding to find the edges [19].

The second-order derivative approach defines the edge pixels based on changes in brightness or zero crossing in the image area [19, 20]. SDGD is a nonlinear operator that can be expressed in the first and second derivatives. Additionally, similar to Canny, SDGD is combined with a Gaussian low pass filter for smoothing purposes [21]. Moreover, a Laplace operator is used to simplify the SDGD operation [22].

The Laplacian is defined as where and are the second derivative filters.

The basic versions in the second derivative filters are given by

Associating (5) with the Gaussian filter yields where is the Gaussian low pass filter.

Five partial derivatives are used in the SDGD filter, as follows: Therefore, A detailed explanation of SDGD can be found in [20, 21, 23].

In [19, 24], SDGD was presented as a filter used in finding edges and measuring objects. Several studies utilize the SDGD filter. For example, Aarnick et al. [25] used this filter in analyzing the ultrasound images of male kidneys and prostate in their study on preprocessing algorithms for edge detection at multiple resolution scales. Their study [25] reported that detecting the contour of objects in grey medical images could be improved by applying an adaptive filter size in SDGD. Nader El-Glaly [24] used SDGD as part of her work in a digital inpainting algorithm. Hagara and Moravcik [23] introduced the PLUS operator, a combination of the SDGD filter and Laplace operator for edge detection. Similar findings were obtained by using PLUS and SDGD filters based on a kernel size of nine or lower, and PLUS yielded better results when the kernel size was greater than nine and was suitable in locating the edges of small objects. Similarly, Verbeek and Van Vliet [22] compared Laplace, SGDG, and PLUS derived from 2D and 3D images. Their research confirmed the findings in [23].

The idea of combining two methods in one algorithm was inspired by a study made by Zheng and Fan [3], where RA and temporal difference were combined to detect the moving object. Another example of hybrid research in BGS was conducted by Lu and Wang [26]. They crossbred optical flow and double background filtering to detect the moving object. Zaki et al. [27] combined frame differencing with a scale invariant feature detector to detect the moving object in various environments.

Based on the study of Persoon et al. [19], the SDGD filter gave better surface localization, especially in highly-curved areas, compared with the Canny edge detection technique. Thus, we adopted this filter in our present work. In addition, Persoon et al. showed that SDGD guaranteed minimal detail smoothing that led to better visualization of polyps in computed tomography (CT) scan data. This finding is aligned with our results reported in [28]. Further, the study by Nader El-Glaly [24] used the SGDG filter in developing an enhanced partial-differential equation-based digital inpainting algorithm to find the missing data in digital images.

To the best of our knowledge, this study is new because no prior work exists that integrates the SDGD filter with the BGS technique. Although Al-Garni and Abdennour used edge detection and the FD technique to find moving vehicles, no information was provided on the edge detection technique they utilized [29]. We used an SDGD filter to enhance the performance of the existing background subtraction technique by combining foreground pixels generated by BGS techniques with the detected edge as our extracted object. The edge pixels are expected to fill in the boundary gap, which will create better connections among pixels in the boundary. This research is an extension of our previous work published in [28] which uses more data from a variety of data sources and more detailed analysis is being done.

3. Methodology

This section discusses the database used and the proposed method.

3.1. Dataset

This study utilized datasets that were acquired from selected prerecorded video collections of several online databases.(a)Smart Engineering System Research Group (SESRG) UKM Collections. This video collection consists of various human actions and activities recorded by students involved in SESRG studies on smart surveillance systems. Besides humans, this database also has a collection of moving cars that are used as nonhuman samples for classification that will be explained further in Section 3.4.(b)CMU Graphic Lab Motion Capture (MoCap) Database [30]. This database, which is owned by Carnegie Mellon University, contains 2,506 trials in 6 categories and 23 subcategories. The videos were recorded in an indoor environment. (c)CMU Motion of Body (MoBo) Database [31]. This database, which is also owned by Carnegie Mellon University, consists of videos showing six different angles of the subject who is walking on a treadmill. (d)Multicamera Human Action Video Data (MuHAVi) [32]. This database is owned by the Digital Imaging Research Centre at Kingston University. It presents 17 action classes with 8 camera views. (e)Human Motion Database (HMDB51) [33]. This database consists of collections of edited videos from digitized movies and YouTube. The collection contains 51 action categories with 7,000 manually annotated clips.

The initial background for videos taken from databases (a), (b), (d), and (e) was modeled using the median value from a selected frame interval. We used five frames from the video sequences with 10 intervals between the frames, that is , , , , and . Next, the median value of these selected frames was used as the reference image. Details of this technique are presented in [34]. Meanwhile, database (c) provided the background reference frame. All datasets except the MuHAVi videos were manually segmented to obtain the ground truth.

3.2. The Proposed Algorithm

The methodology of our proposed techniques is as described in the flowchart shown in Figure 1.

First, the dataset was tested using the following basic parametric BGS techniques: FD, RA, AM, and RGA. Next, the SDGD filter was fused with each technique by combining the output of a background technique with the SDGD filter output. The SDGD filter was selected as a segmentation tool because it produced better results compared to other edge detection techniques (Sobel, Canny, and Roberts) [19, 28].

3.3. Evaluation Method

To evaluate the performance of each BGS technique, we calculated the average of recall, precision, score, and accuracy.(a) Recall (Rcl) refers to the detection rate, which is calculated by comparing the total number of detected true positive pixels with the total number of true positive pixels in the ground truth frame [35]. This is also known as sensitivity. The following equation shows how recall is calculated: where FP is the false positive and FN is false negative.(b) Precision (Prcsn) is the ratio between the detected true positive pixels and the total number of positive pixels detected by the method [35, 36]. This is also known as specificity: where FP is false positive, FN is false negative, TP is true positive, and TN is true negative.(c)-measure or balance -score is the weighted harmonic mean of recall and precision. It is used as a single measurement for comparing different methods [35, 36]: (d) Accuracy is the percentage of correct data retrieval. It is calculated by dividing the number of pixels with true positive plus pixels with true negative pixels over the total number of pixels in the frame. The following equation displays the calculation of accuracy [36]:

This study utilized videos with a multiple number of frames. Hence, we presented a comparison of the average F score and average accuracy percentage as the overall performance benchmark in each BGS technique with and without SDGD filter. The following equation is used to calculate the percentage of improvement:

3.4. Classification

Using the artificial neural network (ANN) as classifier, the segmented image from the proposed technique was subjected to classification testing. Training input of the ANN was extracted from 1,500 randomly chosen segmented frames/images. 750 human blob images represented human samples, and another 750 car blob images represented non-human samples.

A scale conjugate gradient and the back propagation rules were chosen to train the classifier. The ANN was designed with one hidden layer containing ten hidden neurons and an output layer containing two hidden neurons. Both layers used sigmoid as the activation function and the mean squared error value as the performance function. Images were classified as either human or non-human. Next, another set of 1,000 frames/images were chosen for testing. We applied leave-one-out cross validation in our study. Ten experiments and the average classification rate were used to evaluate whether the recognition performance was human or not.

The evaluation was based on the segmented frames/images generated from the enhanced FD technique, with the SDGD filter added to the algorithm. In this experiment, we only used the FD instead of other BGS techniques because it has the fastest processing time. We also performed statistical tests such as recall, precision, and F-score on the obtained classification result by taking positive identification on the classification of human and negative for the classification of non-human.

4. Results and Discussion

This section discusses the robustness of the proposed technique based on videos taken from five different databases. To confirm the robustness of the proposed algorithm, we tested using video that presented a different environment or camera angle. Specifically, the video sequences were recorded in multiple environments (indoors and outdoors) and multiple views.

Figures 2(a)2(f) show some of the background frames used in this research. These background frames were generated by using the median value of selected frames in the video sequences, except for the videos obtained from the MoBo database. Rather than stating the filename, we assigned letters, from A to E, to identify the videos representing the data: A refers to SESRG, B to MoCap, C to MoBo, D to MuHAVi, and E to HMDB51. The words in brackets indicate whether the video environment was either indoor or outdoor.

Next, we present the subjective results of our object extraction. Since this study involves video data with a multiple number of frames, we only depict the results obtained for frame number 10 in data A1 and frame number 29 in data B1 to represent the outdoor and indoor scene samples. Figures 3(a) and 4(a) illustrate the original images of a frame in videos A1 and B1, respectively. Figures 3(b) and 4(b) depict the ground truth images.

Figures 5 and 6 present the extracted subjects in both indoor and outdoor environments using FD, AM, RA, and RGA, with and without the SDGD filter.

The first column of Figures 5 and 6 show that all the basic BGS techniques were capable of detecting the object in the scene of interest based on the tested videos. However, many pixels were missing and resulted in a smaller blob compared with the ground truth image. Our proposed technique solves the problem of missing pixels and reduced blob size because it combines the SDGD filter with FD, AM, RA, and RGA, as shown in column 2 of Figures 5 and 6. The pixel size of the extracted object is slightly enlarged and becomes more compound. Foreground detection showed significant improvement after the proposed technique was applied to all datasets.

Tables 1 and 2 show the number of TP, TN, FP, and FN for the selected indoor and outdoor samples with and without the addition of the SDGD filter in the BGS techniques. Then the values of recall, precision, and F-score were calculated based on mathematical equations (10)–(12).


MethodTPFPFNTNRecallPrecision -Score

FD without SDGD204612791091800640.650.620.63
With SDGD2521804616805390.800.760.78

AM without SDGD198617851151795580.630.530.57
With SDGD2374951763803920.760.710.73

RGwithout SDGD195917811178795620.620.520.57
With SDGD23031022834803210.730.690.71

RGAwithout SDGD2441912696804310.780.730.75
With SDGD2654499483808440.850.840.84


MethodTPFPFNTNRecallPrecision -Score

FDwithout SDGD23361372809799630.740.630.68
With SDGD27889201165796070.710.750.73

AMwithout SDGD27711437808799640.740.610.67
With SDGD27269821077796950.720.740.73

RGwithout SDGD22891419858799140.730.620.67
With SDGD27239851106796660.710.730.72

RGAwithout SDGD256410921175796490.690.700.69
With SDGD29796771535792890.670.810.73

Based on the findings shown in Tables 1 and 2, we can see that the number of TPs has increased significantly, which proves that our technique is able to detect more of the compound blob than the original methods. This finding is also in line with the increment of precision values. Tables 1 and 2 also show that the F-score increased for both samples when we add the SDGD filter.

The graphs in Figures 7 and 8 show the F-score trends in both A1 and B1 for the four BGS techniques, namely, FD, AM, RA, and RGA. The solid lines represent the F-score results using our proposed technique, that is, with the SDGD filter, whereas the dashed lines represent the F-score results without using the SDGD filter. Based on Figures 7 and 8, higher F-score values were noted for the A1 and B1 videos when using the proposed hybrid technique compared with those when using the basic BGS techniques. Thus, our proposed technique improves upon traditional BGS techniques.

To confirm the effectiveness of the proposed technique, we tested the algorithm in six different videos with six different backgrounds, as obtained from the five different databases. Table 3 shows the performance of FD, AM, RA, and RGA with and without the SDGD filter, in terms of the F-score and average accuracy percentage of all six video samples. Columns 5 and 8 show the percentage of improvement for both F-score and average accuracy percentage. Based on Table 3, the use of an SDGD filter improved the average F-score values for all data compared with the values produced by the methods without SDGD. Column 5 of Table 3 shows that the F-score values improved by 1% to 9%. Therefore, the proposed technique, compared with existing techniques, enhances object extraction.


NameMethodEvaluation performance based on average value
-scoreAccuracy (%)
Without SDGDWith SDGD%im*Without SDGDWith SDGD%im*

A1FD0.750.78398.4598.470.02
AM0.690.75698.1698.290.13
RA0.740.77398.4298.440.02
RGA0.790.81298.4298.470.03

A2FD0.780.79196.7496.65−0.09
AM0.800.81197.3197.14−0.17
RA0.760.78296.9196.83−0.08
RGA0.760.77195.7595.65−0.10

B1FD0.680.74697.9998.120.13
AM0.710.76598.2298.370.15
RA0.710.76598.2598.380.13
RGA0.640.68497.4797.510.04

C1FD0.840.85195.6495.690.05
AM0.790.80194.8795.030.16
RA0.570.65891.6992.170.48
RGA0.80.81194.9595.070.12

D1FD0.680.71397.8498.070.23
AM0.710.75498.1398.370.22
RA0.710.74398.1498.360.22
RGA0.700.74498.0498.250.21

E1FD0.800.81196.6696.690.03
AM0.800.81196.7096.710.01
RA0.800.81196.6996.700.01
RGA0.680.77996.4596.500.05

im: percent improvement.

Additionally, Table 3 depicts an increment in average accuracy percentage for each tested technique except for videos A2 and E1. A2 had poor video quality because the footage was taken in a corridor without proper lighting. Because of the poor video quality and bad lighting condition, the SDGD filter was unable to segment the foreground subjects properly and produced an unwanted shadow in the foreground. Nevertheless, our proposed technique is capable of detecting foreground pixels with over 90% accuracy on all videos tested.

Meanwhile, Table 4 exhibits the results of classifying human and non-human recognition based on the segmented frame generated by the proposed technique. In Table 4, ANN successfully recognized human and non-human images from the generated frames, using the improved FD technique. Incorrect classifications were minimal in all ten experiments.


Exp #HumanNon-human%

1Correct49849299
Missed281.0

2Correct49949199.0
Missed191.0

3Correct49849599.3
Missed250.7

4Correct49550099.5
Missed500.5

5Correct48149297.3
Missed1982.7

6Correct49749098.7
Missed3101.3

7Correct48449698.0
Missed1642.0

8Correct49349899.1
Missed720.9

9Correct49749699.3
Missed340.7

10Correct49748698.3
Missed3141.7

Figure 9 presents a matrix describing the overall results of classification testing. The average recognition rate for human and non-human categories was 98.78% and 98.72%, respectively. The rate is higher for both categories as our algorithm provided good human and non-human blob images, which allowed ANN to distinguish between the two categories. Further, the overall performance in both classes was 98.75%. The findings confirmed that the proposed algorithm, which combines FD with the SDGD filter, could generate better silhouette images, thereby facilitating recognition of human and non-human images in the segmented images/frames.

From the matrix, the information of TP, TN, FP, and FN can be extracted; hence Table 5 shows results of the statistical analysis done on the classification findings by ANN.


TPTNFPFNRecallPrecision -Score

4939493661640.9870.9880.987

5. Conclusion

We presented a new hybrid approach that incorporates the SDGD filter with four basic BGS techniques, namely, FD, AM, RA, and RGA. This hybrid technique enhanced segmentation performance, as indicated by the F-score values and average accuracy percentages. The technique was tested on six different videos from five different databases, and each video was taken either indoors or outdoors and showed different scenes. An ANN classifier was used to classify human and nonhuman images appearing in the segmented images generated by our algorithm. As the algorithm was capable of providing good blob images, ANN recognition of human and non-human images in the silhouette images was facilitated.

Although computational time increased, this aspect is acceptable considering the enhancement and the characteristics of second-order derivatives. Therefore, this study is valid and suitable for implementation in non-real-time applications. The proposed hybrid technique can improve upon traditional BGS techniques as indicated by the improved F-score, accuracy, and ANN recognition values after testing various data sources and data environments. The technique can also be considered in detecting moving objects in non-real-time applications, such as investigations of human actions or traffic conditions.

Acknowledgments

This work was supported by Universiti Kebangsaan Malaysia (UKM) research Grant DPP-2013-003 and Ministry of Higher Education (MoHE) research Grant LRGS/TD/2011/ICT/04/02.

References

  1. A. M. McIvor, “Background subtraction techniques,” in Proceedings of the Image and Vision Computing Conference, Auckland, New Zealand, 2000. View at: Google Scholar
  2. M. H. Sigari, N. Mozayani, and H. M. Pourreza, “Fuzzy running average and fuzzy background subtraction: concepts and application,” International Journal of Computer Science and Network Security, vol. 8, pp. 138–143, 2008. View at: Google Scholar
  3. Y. Zheng and L. Fan, “Moving object detection based on running average background and temporal difference,” in Proceedings of the IEEE International Conference on Intelligent Systems and Knowledge Engineering (ISKE '10), pp. 270–272, November 2010. View at: Publisher Site | Google Scholar
  4. J. Park, A. Tabb, and A. C. Kak, “Hierarchical data structure for real-time background subtraction,” in Proceedings of the IEEE International Conference on Image Processing (ICIP '06), pp. 1849–1852, October 2006. View at: Publisher Site | Google Scholar
  5. C. R. Wren, A. Azarbayeni, T. Darrel, and A. P. Petland, “Pfinder: real-time tracking of the human body,” IEEE Transaction on Pattern Analysis and Machine Intelligence, vol. 19, no. 7, pp. 780–785, 1997. View at: Publisher Site | Google Scholar
  6. Z. Tang, Z. Miao, and Y. Wan, “Background subtraction using running Gaussian average and frame difference,” in Proceedings of the International Conference Entertainment Computing (ICEC '07), vol. 4740 of Lecture Notes in Computer Science, pp. 411–414, 2007. View at: Publisher Site | Google Scholar
  7. Z. He, Y. Liu, H. Yu, and X. Ye, “Optimized algorithm for traffic information collection in an embedded system,” in Proceedings of the IEEE Congress on Image and Signal Processing, May 2008. View at: Publisher Site | Google Scholar
  8. A. Singh, S. Sawan, M. Hanmandlu, V. K. Madasu, and B. C. Lovell, “An abandoned object detection system based on dual background segmentation,” in Proceedings of the 6th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS '09), pp. 352–357, IEEE Press, September 2009. View at: Publisher Site | Google Scholar
  9. S. Su and Y. Chen, “Moving object segmentation using improved running Gaussian average background model,” in Proceedings of the International conference on Digital Image Computing: Techniques and Applications, 2008. View at: Publisher Site | Google Scholar
  10. H. Kim, R. Sakamoto, I. Kitahara, T. Toriyama, and K. Kogure, “Background subtraction using generalised Gaussian family model,” Electronics Letters, vol. 44, no. 3, pp. 189–190, 2008. View at: Publisher Site | Google Scholar
  11. Z. H. Huang and K. W. Chau, “A new image thresholding method based on Gaussian mixture model,” Journal Applied Mathematics and Computation, vol. 205, pp. 899–907, 2008. View at: Publisher Site | Google Scholar
  12. Y. Liu, H. Yao, W. Gao, X. Chen, and D. Zhao, “Nonparametric background generation,” in Proceedings of the 18th International Conference on Pattern Recognition (ICPR '06), pp. 916–919, August 2006. View at: Publisher Site | Google Scholar
  13. P. Vu, V. Phong, T. H. Vu, and H. B. Le, “GPU implementation of Extended Gaussian mixture model for background subtraction,” in Proceedings of the 8th IEEE-RIVF International Conference on Computing and Communication Technologies: Research, Innovation and Vision for the Future (RIVF '10), November 2010. View at: Publisher Site | Google Scholar
  14. A. Mittal and N. Paragios, “Motion-based background subtraction using adaptive kernel density estimation,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR '04), pp. II302–II309, July 2004. View at: Google Scholar
  15. F. Porikli and O. Tuzel, “Human body tracking by adaptive background models and mean shift analysis,” in Proceedings of the IEEE International Workshop on Performance Evaluation of Tracking and Surveillance (PETS '03), July2003. View at: Google Scholar
  16. I. J. Young, J. J. Gerbrands, and L. J. van Vliet, “Fundamentals of Image Processing,” Version 2. 3, 1995–2007. View at: Google Scholar
  17. V. G. Narendra and K. S. Hareesh, “Study and comparison of various image edge detection techniques,” International Journal of Image Processing, vol. 4, no. 2, article 83, 2009. View at: Google Scholar
  18. J. F. Canny, “A computational approach to edge detection,” IEEE Transaction Pattern Analysis and Machine Intelligent, vol. 8, no. 6, pp. 679–697, 1986. View at: Publisher Site | Google Scholar
  19. M. P. Persoon, I. W. O. Serlie, F. H. Post, R. Truyen, and F. M. Vos, “Visualization of noisy and biased volume data using first and second order derivative techniques,” in Proceedings of the 14th IEEE Visualization Conference, pp. 379–385, October 2003. View at: Publisher Site | Google Scholar
  20. I. T. Young, “Generalized convolutional filtering,” in Proceedings of the 19th CERN School of Computing, pp. 51–65, 1996. View at: Google Scholar
  21. “Image processing fundamentals: derivative-based operation,” 2011, http://www.mif.vu.lt/atpazinimas/dip/FIP/fip-Derivati.html. View at: Google Scholar
  22. P. W. Verbeek and L. J. van Vliet, “Location error of curved edges in low-pass filtered 2-D and 3-D images,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 16, no. 7, pp. 726–733, 1994. View at: Publisher Site | Google Scholar
  23. M. Hagara and J. Moravcik, “PLUS operator for edge detection in digital images,” in Proceedings of the International Conference of Radioelektonika, pp. 467–470, 2002. View at: Google Scholar
  24. Y. Nader El-Glaly, Development of PDE-Based Digital Inpainting Algorithm Applied to Missing Data in Digital Images [M.S. thesis], Ain Shams University, 2007.
  25. R. G. Aarnick, J. de la Rosette, W. Feitz, F. Debruyne, and H. Wijkstra, “A preprocessing algorithm for edge detection with multiple scales of resolution,” European Journal of Ultrasound, vol. 5, pp. 113–126, 1997. View at: Publisher Site | Google Scholar
  26. N. Lu and J. Wang, “Motion detection based on accumulative optical flow and double background filtering,” in Proceedings of the World Congress on Engineering,, vol. 1, 2007. View at: Google Scholar
  27. Zaki, W. M. D. W, A. Hussain, and M. Hedayati, “Moving object detection using Keypoints reference model,” EURASIP Journal on Image and Video Processing, vol. 2011, 13 pages, 2011. View at: Publisher Site | Google Scholar
  28. F. Y. A. Rahman, A. Hussain, N. M. Tahir, S. A. Samad, and M. H. M. Saad, “Hybrid background subtraction techniques with second derivative on gradient direction filter,” in Proceedings of the International Workshop on Advanced Image Technology (IWAIT '12), 2012. View at: Google Scholar
  29. S. M. Al-Garni and A. A. Abdennour, “Moving vehicle detection using automatic background extraction,” in proceedings of the World Academy of Science, Engineering and Technology, vol. 24, pp. 82–86, Sydney, Australia, December 2006. View at: Google Scholar
  30. “CMU graphic lab motion capture database,” 2010, http://mocap.cs.cmu.edu/. View at: Google Scholar
  31. R. Gross and J. Shi, “The CMU motion of body (MoBo) database,” Technical Report CMU-RI-TR-01-18, Robotics Institute, Carnegie Mellon University, 2001. View at: Google Scholar
  32. S. Singh, S. A. Velastin, and H. Ragheb, “MuHAVi: a multicamera human action video dataset for the evaluation of action recognition methods,” in Proceedings of the 7th IEEE International Conference on Advanced Video and Signal Based (AVSS '10), pp. 48–55, Boston, Mass, USA, September 2010. View at: Publisher Site | Google Scholar
  33. H. Kuehne, H. Jhuang, E. Garrote, T. Poggio, and T. Serre, “HMDB: a large video database for human motion recognition,” in Proceedings of the IEEE International Conference on Computer Vision (ICCV '11), pp. 2556–2563, November 2011. View at: Publisher Site | Google Scholar
  34. F. Y. A. Rahman, A. Hussain, N. M. Tahir, and W. M. D. Zaki, “Modeling of initial reference frame for background subtraction,” in Proceedings of the 6th International Colloquium on Signal Processing and Its Applications (CSPA '10), pp. 125–128, 2010. View at: Google Scholar
  35. E. Fauske, L. M. Eliassen, and R. H. Bakken, “A comparison of learning based background subtraction,” in Proceedings of the Norwegian Artificial Intelligens Symposium (NAIS), pp. 181–192, 2009. View at: Google Scholar
  36. H. Kim, B. Ku, D. K. Han, S. Kang, and H. Ko, “Adaptive selection model in block-based background subtraction,” Electronics Letter, vol. 48, no. 8, 2012. View at: Google Scholar

Copyright © 2013 Farah Yasmin Abdul Rahman et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


More related articles

4294 Views | 1246 Downloads | 6 Citations
 PDF Download Citation Citation
 Download other formatsMore
 Order printed copiesOrder

Related articles

We are committed to sharing findings related to COVID-19 as quickly as possible. We will be providing unlimited waivers of publication charges for accepted research articles as well as case reports and case series related to COVID-19. Review articles are excluded from this waiver policy. Sign up here as a reviewer to help fast-track new submissions.