Engineering Applications of Intelligent Monitoring and Control 2014
View this Special IssueResearch Article  Open Access
Jun Qiang, Jun Chen, Jun Wang, "An Interfield and Intrafield Weighted Interpolative Deinterlacing Algorithm Based on LowAngle Detection and Multiangle Extraction", Mathematical Problems in Engineering, vol. 2014, Article ID 972540, 8 pages, 2014. https://doi.org/10.1155/2014/972540
An Interfield and Intrafield Weighted Interpolative Deinterlacing Algorithm Based on LowAngle Detection and Multiangle Extraction
Abstract
In the process of converting interlaced scanning to progressive scanning, interline flickers, sawtooth, and creeping could be found in motion images. This paper proposes an interfield and intrafield weighted interpolative algorithm based on lowangle detection and multiangle extraction. Using interframe difference of vertical edge area in current field and interfield difference of other areas as the input of motion detection, incorrect judgment could be avoided by fourfields motion detection so that static and motive image can be distinguished. Based on the motion image from the motion detection, multipixels are used in lowangle horizontal detection to extract higher correction angles so as to make interfield and intrafield weighted interpolation. The experiment results show that the proposed algorithm could improve peak signaltonoise ratio (PSNR), restrain some interlace phenomenon, and reach better visual effects. Meanwhile, there is a good balance between efficiency and computation cost.
1. Introduction
Traditional video sources mostly adopt interlaced scanning mode. It can reduce the complexity of the device. But, in the frontend of intelligent video surveillance systems with the development of high definition digital video, when inputting interlaced scanning video into display devices with progressive scanning mode, flaws, for example, interline flickers, sawtooth, and creeping, appear, which result in bad video quality and visual fatigue [1, 2]. The function of deinterlacing is to convert interlaced images to progressive ones and obtain fine video effects.
In the past decades, some deinterlacing algorithms, for example, linear filter, adaptive nonlinear filtering, and motion compensation (MC), have been proposed. Linear filter which has low cost of hardware and low algorithm complexity is easy to realize, but an ideal result is difficult to achieve. Compared with linear filtering, adaptive nonlinear filter has fine results of deinterlacing but needs more costs of hardware. The algorithm of motion compensation uses the related information of spatial and temporal domain to get smooth and natural effects in motion. With its processing, it can overcome stagnation and shake which the previous two algorithms could bring on. Motion compensation has the best performance among these three algorithms, while it has the biggest amount of calculation and the most sensitivity in error. It may cause some serious problems in visual effects such as blocking effect and localized distortion [3, 4].
In order to address the problem of MC, nonmotion compensation algorithm (nonMC) is proposed [5]. Among nonMC techniques, intrafield interpolation (in spatial domain) and interfield interpolation (in temporal domain) are distinguished. Interfield algorithm interpolates the missing lines by employing pixels from different fields, using field insertion and line average methods. Interfield algorithm works properly in the static parts of the image. Spatial interpolation algorithms calculate the lines by interpolating the adjacent lines from the same field, such as line averaging or directional [1, 6, 7]. Intrafield algorithms work much better in the presence of motion but cannot work well in static parts. The main idea of motion adaptive deinterlacing algorithm is proposed in [8].
The biggest problem of motion adaptive deinterlacing algorithms is the detection of the motion. So motion detection should be made before deinterlacing. The algorithms in [9, 10] use a fuzzy inference rule base to choose deinterlacing method according to the motion detection. Recently, several methods have been proposed to deinterlace by focusing on edges [1, 6].
Our research mainly focuses on two points. Firstly, an accurate motion detection method will be used. Secondly, based on the result of motion detection, we find a suitable deinterlacing algorithm to solve the interlace effect.
As for the motion detection, we adopt fourfield motion detection which obtains the motion information by using adjacent four fields’ luminance difference. It is easy to obtain the motion information which is used commonly. But, in previous research, this motion detection method may cause deviation. In [11], Chang et al. proposed the same parity fourfield motion detection applied in MC deinterlacing which has achieved accurate effect. And Yao et al. [12] mentioned a modified fourfield method for motion adaptive interpolation. They both mentioned the importance of fourfield motion detection. In this paper, we adopt further detection. Interframe difference is used as the input of motion detection in vertical edge area of current field. And other areas use interfield difference as the input motion detection. It can avoid incorrect judgment.
In order to improve the accuracy of horizontal directive detection, the interpolative method, and the effects of deinterlacing, in this paper, we proposed an interfield and intrafield weighted interpolative algorithm based on lowangle detection and multiangle extraction. First, adjacent four fields for motion detection are used to distinguish static and motive parts. Then, based on the result of the motion detection, more correlative pixels of interpolated pixels in adjacent lines are increased to improve the accuracy of horizontal directive detection. Finally, through extracting pixel angles of higher correlation degree and setting weighted coefficient, the intrafield and interfield weight interpolation are taken. The proposed method could solve the lowangle edge detection which could not be solved by traditional algorithms. Moreover, combined with intrafield and interfield interpolation, using weighted average to calculate the value of the interpolated pixel, the effect of interpolation gets a fine result. The experimental results show that this algorithm efficiently improves PSNR and has an efficient computation time ratio.
The remainder of the paper is organized as follows: Section 2 presents fourfield motion detection. Section 3 describes the proposed algorithm of deinterlacing by a lowangle extraction based on intrafield and interfield weighted interpolation. Section 4 gives the comparison of experiments and reports the results, followed by the conclusion in Section 5.
2. Motion Detection
2.1. Problem Statement
In the research of deinterlacing, we find that motion detection plays an important role. Motion detection is usually used to distinguish motive image and static image, and then different deinterlacing algorithm will be adopted accordingly; thus, better effects could be obtained. If the same deinterlacing algorithm is used in all the processes, the effect of the algorithm could not be demonstrated. Consequently, phenomenon such as no smooth edge, virtual image, breakpoint, and hazy detail will appear. In [13–15], Lee et al. mentioned the importance of motion detection. However, the incorrect motion detection methods cause two kinds of result, motion misjudgment and motion missing. Motion misjudgment mistakes the motive image as a static one. Motion missing is a situation that the motive images cannot be detected. These two problems cause serious effects on deinterlacing. So, motion detection needs to minimize the phenomenon of incorrect judgments [16]. With the accurate motion detection, followup deinterlacing algorithm could adaptively choose suitable method to deal with motive or static images.
2.2. Algorithm Description
Fourfield motion detection, which is shown in Figure 1, is used in the proposed algorithm. It denotes the pixels needed in four fields.
Motion information could be obtained by using adjacent four fields’ luminance difference. Prior two fields, prior one field, current field, and rear one field are denoted by , , , and , separately. The interpolated pixel in motion detection is denoted by .
Two kinds of absolute image errors are defined:
Detection input is
Comparing threshold which was set in advance with from (2), the result of detection is outputted. Motive images are defined as “1” and static images as “0.”
In previous research, this motion detection method may cause deviation. So, further detection is needed. A suitable method is adopted as follows: interframe interpolation is used as the input of motion detection in vertical edge area of current field, while interfield interpolation is used as the input motion detection in other areas. It can avoid incorrect judgment.
Fields differences are defined as follows: where represents interframe difference and and represent interfield difference. Motion estimate of is
Extended in spatial domain by , calculated by nonlinear twovalued function, the result of motion detection will be obtained [16, 17].
3. Deinterlacing Algorithm by a LowAngle Extraction
3.1. Problem Statement
According to the previous analysis, in the adaptive motion deinterlacing algorithm, it is important to distinguish motion and static status. But, correct detection to the edge information in motive image and proper application of the edge information to the spatial interpolate filter are also very important to improve the quality of image.
Edgebased line average (ELA) algorithm is widely used in deinterlacing. But, its defect is that it is hard to detect the horizontal edge. ELA only detects the edges in 45°, 90°, and 135° and is unable to detect lowangle edge [18]. Kuo et al. proposed the ELA based on 3 + 3 taps which improved image horizontal edge detection by comparing the differences of adjacent pixels [19]. Lee et al. expand the 3 + 3 taps ELA to 5 + 5 taps [20] and de Haan and Lodder to 7 + 7 taps [21]. All these improved algorithms help to improve the lowedge detect function, but they are unable to reach the horizontal edge. The 7 + 7 taps can only detect the smallest angle in 18°. Chang et al. proposed EIELA using adaptive method, which has large amount of calculation and complexity [22]. It is hard to realize in hardware and not a suitable method in video surveillance system. Yong et al. proposed a spatialtemporal weight and edge adaptive deinterlacing algorithm (STWEA) [23]. It is a good method to use lowangel edge detection and spatialtemporal weight calculation. It detects lowangel edge by using an adaptive searching radius in which the 6° edge can be detected. This algorithm achieves high image quality and low hardware complexity, but it still has problems in dealing with details, especially in high speed motive image.
3.2. Algorithm Design and Description
In order to improve the effect of deinterlacing and get high quality of details, we propose an algorithm combined with lowangle detection and intrafield and interfield bidirectional interpolation based on angle extraction. Considering the advantages and disadvantages of the existing edge detection algorithm, an improved algorithm is designed from two aspects: expand detection region and search more pixels in one line in order to find the edge direction accurately; use angle analysis result and extract the highest correlated angle to interpolate, both in intrafield and interfield.
Let be the interpolated pixel and related pixels in adjacent lines are of line and of line , shown in Figure 2.
Pixels in diagonal direction on two rows represent the probable edge direction. By comparing the absolute difference of luminance values among the pixels in diagonal direction, the correlation of pixels in diagonal direction and interpolated pixel could be determined. The smaller the value of the absolute difference is, the bigger correlation it has. The direction of minimum difference is the edge direction. We select 17 pixels in one row; it means that 18 angels with horizontal direction will be detected, such as 90° (270°), 45° (135°), 26.57° (123.43°), 18.43° (161.57°), 14.03° (163.97°), 11.3° (168.7°), 9.46° (170.54°), 8.13° (171.87°), and 7.12° (172.88°).
Proposed algorithm is described as follows.
Step 1. The angles mentioned above are used to find the pixels which correspond to interpolated pixel. In (5), (6), and (7), all 18 angles are shown:
Step 2. In order to state clearly, the diagonal pixels are divided into three parts: middle part, 90° angles, a_{8} and b_{8}; right part, pixels on the right of a_{8} and right of b_{8}; left part, pixels on the left of a_{8} and right of b_{8}. Just as defined in (5), (6), and (7), the angles are divided into three parts. Equation (5) denotes the middle part, (6) denotes the right part, and (7) denotes the left part.
Then, the threshold functions are defined as follows, which are the absolute difference of diagonal pixels:
denotes the luminance component of pixel and denotes the th field.
Step 3. Intrafield: we choose the optimal direction angle by comparing these differences to select the minimum value, and then the diagonal direction which is corresponding to minimum value is generally the edge direction:
If all the absolute differences of two pixels in diagonal direction are over the , we consider that no edge exists. Line averaging is managed.
Step 4. Interfield: considering the accuracy of edge detection, other two adjacent fields are used for calculating the temporal correlation. The direction of the (get from Step 3) is the most correlation direction. Two corresponding pixels in fields and are also the most correlative pixels. They are considered as the two positions before and after the motion. Adopting bidirectional interpolation, this algorithm uses the average of these two pixels to interpolate.
At the basis of Step 3, adjacent fields’ information is used. All directions’ absolute differences of the same line in adjacent fields and the updown lines in adjacent fields are defined as follows (Figure 3).
Same line:
Adjacent line:
The minimum of all directions from native and adjacent fields , as shown in (14), is considered the most correlated direction in all directions, which will be interpolated.
Step 5. According to Steps 3 and 4, intrafield and interfield interpolation are both managed. Spatial difference “” (get from (11)) is the absolute difference of correlated pixels in forward and back fields. Temporal difference “” (get from (14)) is the absolute difference of diagonal pixels in updown lines which is in the edge direction of intrafield. We define the spatial weighted coefficient and temporal weighted coefficient as follows:
Then, using which is the value of the intrafield interpolation and which is the value of the interfield interpolation to calculate the value of weighted interpolated pixel, formula is as follows:
Step 6. The value of interpolated pixel is used for filter. If the adjacent direction consistent with interpolated pixel, no operation will be taken. Otherwise, median filtering is adopted to revise the adjacent pixel which has been interpolated.
4. Experimental Results
4.1. Experiment Setup and Result
The whole deinterlacing system is simulated by using ISE 14.1 design tools, 2v1000fg2564 simulation device by Xinlinx, and ModelSim SE 7.0 simulation tools.
After simulation, PSNR and MSE are contrasted between original image and the processed image by proposed algorithm. The definitions are
denotes resolution of original image. is the luminance value of original image in point . is the luminance value of deinterlaced image in point .
The experiment is designed as follows. Two oddeven overlapped fields test sequences of 25 frames length are used for algorithm testing. The 50th frame progressive image of sequence plays fluently and has fine test result by deinterlacing the proposed algorithm. Experimental results are shown in Figure 4 and Table 1.

(a)
(b)
(c)
(d)
(e)
(f)
Original images in Figure 4 are the 35th frame in test sequence. The proposed algorithm has also been tested by using several standard video sequences and the results have been compared with other deinterlacing algorithm, such as ELA, ELA using 3 + 3 and 5 + 5 windows, STWEA, and ECA. By contrast to Table 1, PSNR of test sequences are both improved by the proposed algorithm. We got the fluent visual effect and removed sawtooth phenomenon.
4.2. Analysis
From Figure 4, we could find that traditional ELA has the drawback of low vertical definition. And using intrafield interpolation directly leads to fuzzyedge phenomenon. ELA using 3 + 3 and 5 + 5 windows are also making errors when processing nonclear edges or ambiguous situations, as we analyzed in Section 2. The proposed algorithm combined with interfield interpolation and intrafield lowangle edge detection solves these problems well.
Computational time ratio, which means quality/cost ratio, is calculated to evaluate the proposed algorithm if it is competitive versus the other deinterlacing algorithms. Table 1 (PNSR) shows the PNSR of every test sequence, which indicates the “quality.” Table 2 (CPU time) shows the consumed CPU times of every algorithm which means “cost.” In Tables 1 and 2, the last columns both mean the average value of every algorithm, which are used in calculating the computational time ratio in Table 3.


In Table 3, the proposed algorithm gets the best quality/cost ratio. But traditional algorithm ELA’s computational time ratio is better than ELA 3 + 3, ELA 5 + 5, and STWEA. Analyses show that ELA 3 + 3, ELA 5 + 5, and STWEA are more complex than ELA and have better PSNR. ELA has less cost of CPU times but has bad deinterlacing quality. It is not recommended in intelligent video surveillance systems. As all things are considered, our approach offers a better quality/cost in comparison with other traditional algorithms.
5. Conclusion
Through the research and analysis of existed deinterlacing algorithms, this paper summarizes their merits and drawbacks and proposes a novel deinterlacing algorithm. Effective fourfield motion detection is adopted to avoid the incorrect judgment, which could improve the accuracy of motion detection. Lowangle detection and multiangle extraction can obtain higher corrected pixels angle. Interpolation with weighted interfield and intrafield increases accuracy of interpolation. Due to the proposed algorithm with higher ability of multiangle extraction and interpolation, we get good effect on deinterlacing. By theoretical and experimental analysis, this algorithm overcomes the drawbacks of traditional algorithms, gets subjective evaluation and PSNR and computational ratio contrasted in objective evaluation, and obtains better visual effects.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
Acknowledgment
This research was supported by the Natural Science Foundation of the Anhui Higher Education Institutions of China (Grant no. KJ2013B031).
References
 G. Jeon, M. Anisetti, V. Bellandi, E. Damiani, and J. Jeong, “Designing of a type2 fuzzy logic filter for improving edgepreserving restoration of interlacedtoprogressive conversion,” Information Sciences, vol. 179, no. 13, pp. 2194–2207, 2009. View at: Publisher Site  Google Scholar
 P. Brox, I. Baturone, S. SánchezSolano, J. GutiérrezRíos, and F. FernándezHernández, “A fuzzy edgedependent motion adaptive algorithm for deinterlacing,” Fuzzy Sets and Systems, vol. 158, no. 3, pp. 337–347, 2007. View at: Publisher Site  Google Scholar  MathSciNet
 G. de Haan and E. B. Bellers, “Deinterlacing an overview,” Proceedings of the IEEE, vol. 8, pp. 819–825, 1998. View at: Google Scholar
 H. Hwang, M. H. Lee, and D. I. Song, “Interlaced to progressive scan conversion with double smoothing,” IEEE Transactions on Consumer Electronics, vol. 39, no. 3, pp. 241–246, 1993. View at: Publisher Site  Google Scholar
 G. de Haan, “Deinterlacing,” in Digital Video Post Processing, vol. 9, pp. 185–201, University Press Eindhoven, 2006. View at: Google Scholar
 G. Jeon, M. Anisetti, J. Lee, V. Bellandi, E. Damiani, and J. Jeong, “Concept of linguistic variablebased fuzzy ensemble approach: application to interlaced HDTV sequences,” IEEE Transactions on Fuzzy Systems, vol. 17, no. 6, pp. 1245–1258, 2009. View at: Publisher Site  Google Scholar
 S. Park, G. Jeon, and J. Jeong, “Deinterlacing algorithm using edge direction from analysis of the DCT coefficient distribution,” IEEE Transactions on Consumer Electronics, vol. 55, no. 3, pp. 1674–1684, 2009. View at: Publisher Site  Google Scholar
 A. M. Bock, “Motionadaptive standards conversion between formats of similar field rates,” Signal Processing: Image Communication, vol. 6, no. 3, pp. 275–280, 1994. View at: Publisher Site  Google Scholar
 D. van de, B. Ville, W. Philips, and I. Lamahieu, “Fuzzybased motion detection and its application to deinterlacing,” in Fuzzy Techniques in Image Processing, vol. 52 of Studies in Fuzziness and Soft Computing, pp. 337–369, Springer, 2000. View at: Publisher Site  Google Scholar
 D. van de Ville, B. Rogge, W. Philips, and I. Lamahieu, “Evaluation of several operators for fuzzybased motion adaptive deinterlacing,” in Proceedings of the PRORISC IEEE Benelux Workshop on Circuits. Systems and Signal Processing, vol. 11, pp. 535–544, 1999. View at: Google Scholar
 Y. Chang, P. Wu, S. Lin, and L. Chen, “Four field local motion compensated deinterlacing,” in Proceeding of the IEEE International Conference on Acoustics, Speech, and Signal Processing, vol. 5, pp. V253–V256, May 2004. View at: Publisher Site  Google Scholar
 Y. Yao, H. Chen, and Y. Cheng, “A low power IC for efficient deinterlacing based on refined motion adaptive method,” in Proceedings of the IEEE International Symposium on Industrial Electronics (ISIE '12), pp. 1094–1099, Hangzhou, China, May 2012. View at: Publisher Site  Google Scholar
 S. Lee and D. Lee, “A motionadaptive deinterlacing method using an efficient spatial and temporal interpolation,” IEEE Transactions on Consumer Electronics, vol. 49, no. 4, pp. 1266–1271, 2003. View at: Publisher Site  Google Scholar
 S.F. Lin, Y.L. Chang, and L.G. Chen, “Motion adaptive deinterlacing by horizontal motion detection and enhanced ELA processing,” in Proceedings of the International Symposium on Circuits and Systems (ISCAS '03), vol. 2, pp. II696–II699, May 2003. View at: Publisher Site  Google Scholar
 B. Yoo, B. Kim, and K. Lee, “An efficient motion adaptive deinterlacing algorithm using spatial and temporal filter,” in PProceedings of the IEEE Region 10 Conference: Trends and Development in Converging Technology Towards (TENCON '11), pp. 288–292, Indonesia, November 2011. View at: Publisher Site  Google Scholar
 S. F. Lin, Y. L. Chang, and L. G. Chen, “Motion adaptive interpolation with horizontal motion detection for deinterlacing,” IEEE Transactions on Consumer Electronics, vol. 49, no. 4, pp. 1256–1265, 2003. View at: Publisher Site  Google Scholar
 J. Qiang, M. Zhou, and J. Wang, “Algorithm research on eliminating sawtooth phenomenon in motion image,” Video Engineering, vol. 36, no. 15, pp. 32–35, 2012. View at: Google Scholar
 S. G. Lee and D. H. Lee, “A motionadaptive deinterlacing method using an efficient spatial and temporal interpolation,” IEEE Transactions on Consumer Electronics, vol. 49, no. 4, pp. 1266–1271, 2003. View at: Publisher Site  Google Scholar
 C. J. Kuo, C. Liao, and C. C. Lin, “Adaptive interpolation technique for scanning rate conversion,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 6, no. 3, pp. 317–321, 1996. View at: Publisher Site  Google Scholar
 H. Y. Lee, J. W. Park, T. M. Bae, S. U. Choi, and Y. H. Ha, “Adaptive scan rate upconversion system based on human visual characteristics,” IEEE Transactions on Consumer Electronics, vol. 46, no. 4, pp. 999–1006, 2000. View at: Publisher Site  Google Scholar
 G. de Haan and R. Lodder, “Deinterlacing of video data using motion vectors and edge information,” in Proceedings of the International Conference on Consumer Electronics (2002 Digest of Technical Papers), pp. 70–71, Los Angeles, Calif, USA, 2002. View at: Google Scholar
 Y. Chang, S. Lin, and L. Chen, “Extended intelligent EDGEbased line average with its implementation and test method,” in Proceedings of the IEEE International Symposium on Circuits and Systems (ISCAS '04), vol. 2, pp. 341–344, Vancouver, Canada, May 2004. View at: Publisher Site  Google Scholar
 D. Yong, L. Shengli, and S. Longxin, “Spatiotemporal weight and edge adaptive deinterlace,” Chinese Journal of Computers, vol. 30, no. 4, pp. 655–660, 2007. View at: Google Scholar
Copyright
Copyright © 2014 Jun Qiang et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.