Research Article  Open Access
Qiuwen Zhang, Nana Li, Yong Gan, "Low Complexity Mode Decision for 3DHEVC", The Scientific World Journal, vol. 2014, Article ID 392505, 12 pages, 2014. https://doi.org/10.1155/2014/392505
Low Complexity Mode Decision for 3DHEVC
Abstract
High efficiency video coding (HEVC) based 3D video coding (3DHEVC) developed by joint collaborative team on 3D video coding (JCT3V) for multiview video and depth map is an extension of HEVC standard. In the test model of 3DHEVC, variable coding unit (CU) size decision and disparity estimation (DE) are introduced to achieve the highest coding efficiency with the cost of very high computational complexity. In this paper, a fast mode decision algorithm based on variable size CU and DE is proposed to reduce 3DHEVC computational complexity. The basic idea of the method is to utilize the correlations between depth map and motion activity in prediction mode where variable size CU and DE are needed, and only in these regions variable size CU and DE are enabled. Experimental results show that the proposed algorithm can save about 43% average computational complexity of 3DHEVC while maintaining almost the same ratedistortion (RD) performance.
1. Introduction
With the development of the technology of 3D television (3DTV) and free viewpoint television (FTV), 3D video coding attracts more and more attention. The typical 3D video is represented using the multiview video plus depth (MVD) format [1], in which few captured texture videos as well as associated depth maps are used. The depth maps provide perpixel with depth corresponding to the texture video that can be used to render arbitrary virtual views by using depth image based rendering (DIBR) [2, 3]. In recent years, high efficiency video coding (HEVC) based 3D video coding (3DHEVC) technology [4, 5] is now being standardized by joint collaborative team on 3D video coding (JCT3V) as an extension to HEVC [6, 7]. From the JCT3V meetings, the developed coding schemes for 3DHEVC mainly use HEVC together with exploiting temporal and interview correlation. Thus, many coding tools applied in 3DHEVC are based on the hybrid coding scheme and highly related to HEVC. Different from singleview video coding, 3DHEVC uses disparity estimation (DE) to reduce the interview redundancy. The test model of 3DHEVC uses the variable size coding unit (CU) and DE to exploit both temporal and view correlation within temporally successive pictures and neighboring views. This technique achieves the highest possible coding efficiency, but it results in extremely large encoding time which obstructs it from practical use. Therefore, it is necessary to develop a method that can reduce complexity of 3DHEVC with minimal loss of image quality. Fast CU size and DE decision algorithms for encoding multiview video plus depth are extremely necessary.
Recently, a number of efforts have been proposed to reduce the computational complexity for the HEVC encoder. An effective CU size decision method is proposed in [8] to reduce encoding complexity of HEVC. Besides, a fast mode decision method is proposed based on the direction information of the spatially adjacent CUs in [9]. Fast intramode decision method [10] uses edge information of the current prediction unit (PU) to choose a reduced set of candidate prediction directions. A complexity control method [11] is proposed based on fast mode decision algorithm that dynamically adjusts the depth of the CU defined by quadtreebased structures. A fast CU splitting and pruning method [12] is presented based on early CU split and pruning decision for HEVC intracoding. All these methods are efficient in reducing computational complexity with small degradation in coding performance. However, these methods are not directly applicable to the 3DHEVC, where high computational complexity is intrinsically related to the use of new prediction coding structures for depthenhanced multiview formats.
To the best of our knowledge, studies devoted to investigation of complexity reduction of the 3DHEVC are rarely found in the literatures. To this end, this paper proposes a fast CU size and DE mode decision algorithm to reduce 3DHEVC computational complexity. Our algorithm exploits the correlation between depth map and motion activity to reduce the 3DHEVC computational complexity. Since the depth maps can efficiently describe the actual 3D information, the potential of utilizing depth map for fast CU size and DE algorithm is promising. The proposed algorithm consists of two approaches: fast CU size decision and selective disparity estimation. Based on these two criteria, the number of candidate modes in a view is reduced, the disparity search is selectively enabled, and the search range of CU size is adaptively determined. Experimental results demonstrate that the proposed algorithm can significantly reduce the computational complexity of 3DHEVC while maintaining almost the same ratedistortion (RD) performance as the original encoder.
The rest of the paper is organized as follows. In Section 2, we analyze the correlation of depth maps among motion activities and propose a fast CU size and DE algorithm based on depth map. Experimental results and conclusions are given in Sections 3 and 4, respectively.
2. Low Complexity Mode Decision Algorithm
2.1. Fast CU Size Decision Based on Depth Map
3DHEVC inherits an advanced quadtreebased coding approach from HEVC, wherein a picture is divided into coding tree units (CTUs) [13]. Those are equivalent to macroblocks (MBs) in previous video coding standards such as H.264/AVC. The CTU can then be split into four CUs, and the CU is the basic unit of region splitting used for inter/intraprediction, which allows recursive subdividing into four equally sized blocks. This process gives a contentadaptive coding tree structure comprised of CU blocks that may be as large as a treeblock or as small as 8 × 8 pixels. Figure 1 shows the architecture of tree structured CUs. A specified maximum depth level is set to limit the CU split recursion. At each depth level (CU size), 3DHEVC performs motion estimation (ME) with different sizes including , , , and . Similar to the joint model of HEVC, the mode decision process in 3DHEVC is performed using all the possible CU sizes (depth levels) and prediction modes to find the one with the least RD cost using Lagrange multiplier [8]. This achieves the highest coding efficiency but requires a very high computational complexity. In this paper, we propose a fast CU size decision algorithm for 3DHEVC. Since the optimal depth level is highly contentdependent, it is not efficient to use all levels. We can determine CU depth range (including the minimum depth level and the maximum depth level) and skip some specific depth levels rarely used in the mode decision process.
3DHEVC usually allows a maximum CU size that equals 64, and the depth level range is from 0 to 3. The CU depth level has a fixed range for a whole video sequence in 3DHEVC reference software [15]. In fact, small depth level values tend to be chosen for CUs in small global motion or homogeneous texture region, and large depth level values are chosen for CUs with large global motion or rich texture region. We can see from experiments of 3DHEVC coding that the depth value of “0” occurs very frequently for large homogeneous texture region. On the other hand, the depth value of “0” is rarely chosen for treeblocks with active motion or rich texture region. These results show that CU depth level range should be adaptively determined based on the motion and texture property of treeblocks. In 3D video coding, a depth map represents a relative distance from a camera to an object in the 3D space, it can be regarded as a grayscale image using dark and bright values to represent far and close object, and the object depth not only represents the physical object position in 3D space but also indicates the motion activity of the object itself on the image plane. Under the condition that cameras are set up in a close parallelized structure, the depth maps are correlated to the texture video motion fields. Consequently, the optimal depth value level of current treeblock may have a strong correlation with its associated depth maps. Based on this concept, we can make use of depth map and motion activity correlations to analyze region properties and skip ME on unnecessary CU sizes.
In 3D space, the motion of the close object in depth map is usually higher than that of the far object in depth map [16]; the major interesting ratio of the video object should be put in the middle region. To achieve great saving in coding time while minimizing the loss in coding efficiency, the depth level of a treeblock having limited contribution to coding efficiency should be skipped. So we use depth information to filter out the unsuitable mode candidates to speed up the encoding process. Since the depth map indicates the physical position of the object in the 3D space, the potential of utilizing depth map for fast CU size decision is promising. In a nature video test sequence, the degree of motion activity for the object with near region from the 3D space will usually be higher than that of the object with far region. Based on this observation, the depth map can be used to classify the motion activities of video objects by this property. Thus, we first classify the objects motion activities into three regions: near region, middle region, and far region according to the depth map value. Near region represents a medium local motion or a smooth texture on the 3D space plane, middle region represents a large global motion or rich texture on the 3D space plane, and far region represents a small global motion or a homogeneous texture on the 3D space plane. Considering that the optimal CU depth level is highly dependent on object motion activities and texture characteristic mentioned above (small depth level values are suitable for CUs in small global motion or homogeneous texture region, and large depth level values are reasonable for CUs with large global motion or rich texture region), we can establish a relationship between the depth level and the depth map. By utilizing the depth map, we can classify scenes according to the corresponding depth map value and assign the suitable CU depth level candidates. Based on the aforementioned analysis, the 3D video space is classified into three regions based on the depth information: near region, middle region, and far region; the treeblocks classification based on depth map can be represented by the following equations: where is the depth map value of current treeblocks and and are chosen based on the sequence characteristics and set to 0 and 255, respectively. The selection of the thresholds and should greatly reduce the 3DHEVC computational complexity while keeping a high accuracy in CU size decision. Based on extensive experiments, the thresholds and are, respectively, set to 200 and 30, which achieve a good and consistent performance on a variety of test sequences with different texture characteristics and motion activities.
Extensive simulations have been conducted on 8 video sequences with different resolutions to analyze the depth level distribution for these three types of treeblocks. Among these test sequences, Kendo, Balloons, and Newspaper are in 1024 × 768 resolution, while Undo_Dancer, GT_Fly, Poznan_Street, Poznan_Hall2, and Shark are in 1920 × 1088 resolution. The test conditions are as follows: there is IBP view structure, there is test of 200 frames for each sequence, quantization parameter (QP) is chosen with 26, 31, 36, and 41, group of pictures (GOP) size is 8, treeblock size is 64, and contextadaptive binary arithmetic coding (CABAC) is used for entropy coding. By exploiting the exhaustive intramode decision in HTM under the aforementioned test conditions, we investigate the depth level distribution for these three types of treeblocks.
Table 1 shows the depth level distribution for each type of treeblocks, where “level 0,” “level 1,” “level 2,” and “level 3” are the depth levels of treeblocks. It can be seen that, for treeblocks with near region mode, about 70% of total treeblocks choose the optimal depth level with “0” and about 21% treeblocks choose the optimal depth value with “1.” In other words, if the maximum depth level is set to be “1,” it will most likely cover about 91% of treeblocks. For treeblocks with middle region mode, about 96% of treeblocks choose depth levels with “1,” “2,” and “3.” If the minimum depth level is set to be “1” and the maximum depth level is set to be “2,” it will most likely cover about 96% of treeblocks. On the other side, the probability of choosing the depth level with “0” is very low, less than 4%, and thus intraprediction on depth level of “0” (CU size 64 × 64) can be skipped. For treeblocks with far region mode, the probability of choosing the depth levels of “0” is more than 90%, and thus intraprediction on depth levels of “1,” “2,” and “3” (CU sizes 32 × 32, 16 × 16, and 8 × 8) can be skipped. Based on the above analysis, the candidate depth levels that will be tested using RD optimization (RDO) for each treeblock are summarized in Table 2. With the proposed fast CU size decision, most of treeblocks can skip one to three tested depth levels. A flowchart of the proposed fast CU size decision is given in Figure 2.


2.2. Selective Disparity Estimation Based on Depth Map
One of the most important aspects for efficient MVD coding is the redundancy reduction among different views at the same time instance, for which the content is usually rather similar and only varies by a slightly different viewing position. As a coding tool for dependent views, the concept of disparity estimation has been involved as an alternative to motion estimation in 3DHEVC encoders. Here, ME refers to interpicture prediction that uses already coded pictures of the same view at different time instances, while DE refers to interpicture prediction that uses already coded pictures of other views at the same time instance [17]. DE is also used in the multiview video coding (MVC) extension of H.264/MPEG4 AVC and similarly the coding treeblock syntax and decoding process of HEVC remain unchanged when adding DE to 3DHEVC codec. Only the highlevel syntax has been modified so that already coded video pictures of the same access unit can be inserted into the reference picture lists [1]. Thus, in the joint mode of 3DHEVC, both ME and DE are included in the encoding process. This achieves the highest coding efficiency but requires a very high computational complexity. Disparity estimation is to search the best matched block in frames from neighbor views. As mentioned above, disparity prediction is used to exploit interview dependence. Although temporal prediction is generally the most efficient prediction mode in 3DHEVC, it is sometime necessary to use both DE and ME rather than only use ME to achieve better predictions. In general, temporal motion cannot be characterized adequately, especially for regions with nonrigid motion and regions with motion boundaries. For the former, ME based on simple translation movement usually fails and thus produces a poor prediction. For the latter, regions with motion boundaries are usually predicted using small mode sizes with larger magnitude of motion vectors and higher residual energy. As mentioned above, the depth map indicates the motion activity of the object itself. Normally, areas with homogeneous motion probably belong to the far depth region, and areas with complex motion probably belong to the middle depth region. Since, for a normal parallelized camera setting, the major object motion should be put in the middle region, thus, the regions with far depth region are more likely to choose temporal prediction, and regions with middle depth region are more likely to choose interview prediction.
Table 3 shows probabilities of choosing interview prediction and temporal prediction for each type of treeblocks classified based on depth map. For treeblocks with near region mode, the average probabilities of choosing temporal prediction and interview prediction are 87% and 13%, respectively. For treeblocks with middle region, they are 69% and 31%. For treeblocks with far region mode, they are 97% and 3%. We can see from Table 3 that treeblocks with far region mode are much more likely to choose temporal prediction. Thus, for far region mode treeblocks, the procedure of the interview prediction can be skipped with only a very low miss detection ratio, by using the optimal prediction mode chosen by the full interview and temporal prediction modes. But, for middle region mode treeblocks (treeblocks with near region mode), the average probabilities of choosing interview prediction are 31% (13%). Although the test sequences such as “Poznan_Hall2” and “Newspaper” contain large area of the homogeneous textures and lowactivity motion, which are more likely to be encoded with temporal prediction, the probability of interview prediction for treeblock with middle region and near region is still the highest. Thus, if we disable interview prediction in middle region and near region, the coding efficiency loss is not negligible. Based on the aforementioned analysis, we propose a selective disparity estimation algorithm in which disparity search is selectively enabled. For treeblocks with far region mode, disparity search is skipped, while, for treeblocks with middle region mode, disparity search is enabled. For treeblocks with near region mode, the RD cost of the motion vector predictor (MVP) is compared with that of the disparity vector predictor (DVP). If the RD cost of MVP is larger than that of DVP, disparity search is enabled; otherwise it is disabled. A flowchart of the scheme is given in Figure 3.
 
“T” and “V” represent temporal prediction and view prediction, respectively. 
3. Experimental Results
In order to evaluate the performance of the proposed fast algorithms, the fast CU size decision and selective disparity estimation algorithms are implemented on the recent 3DHEVC reference software (3DVHTM 4.1) [18]. All the simulations are defined under the common test conditions (CTC) [19] defined by JCT3V. We have tested the proposed algorithms on eight sequences defined in the CTC with two resolutions (1024 × 768 and 1920 × 1088). The encoder configuration is as follows: there are 3 view cases, the GOP length is 8 with an intraperiod of 24, HEVC codecs are configured with 8bit internal processing, the coding treeblock has a fixed size of 64 × 64 pixels and a maximum CU depth level of 4, resulting in a minimum CU size of 8 × 8 pixels, and CABAC is used as the entropy coder. The proposed algorithm is evaluated with QP combinations for texture video and depth map (25, 34), (30, 39), (35, 42), and (40, 45). The experiments test 200 frames for each sequence. Each sequence is composed of three texture videos and three depth map views: the centerthe leftthe right views (in coding order). After encoding, the intermediate rendered views were synthesized between all views. The intermediate rendered views are generated at the receiver using view synthesis reference software (VSRS) algorithm provided by MPEG [20].
3.1. Individual Performance Results of the Proposed Algorithm
Table 4 gives the individual evaluation results of the proposed algorithm compared with the original 3DVHTM algorithm, that is, fast CU size decision (FCUS) and selective disparity estimation (SDE), respectively. The Bjontegaard delta PSNR (BDPSNR) [21] represents the average PSNR gain, bitrate (BDBR) represents the improvement of total bitrates for 3D video coding, and “Dtime (%)” represents the entire coding time change in percentage. The “texture” represents average PSNR for coded texture video views. The “rendered” represents average PSNR for rendered views. Rendered PSNR on rendered view distortion can be measured by comparing the coded rendered view with the image rendered with uncompressed texture videos and depth map [22]. The bitrate under consideration is the sum of the bitrates of the three coded texture videos and depth map views.

The proposed two approaches can greatly reduce the coding time with similar coding efficiency for all sequences. FCUS can save about 35% coding time over all sequences. The coding efficiency loss is very negligible with 0.02 dB–0.08 dB PSNR drop for average texture videos and 0.01 dB–0.04 dB PSNR drop for average rendered views. This result indicates that FCUS can efficiently skip unnecessary depth levels in CU size decision. As far as the SDE algorithm is concerned, 13% coding time has been reduced. And, for average texture videos coding, the average PSNR drop is 0.03 dB, or the bitrate increases about 0.84% on average. For average rendered views coding, the average PSNR drop is 0.015 dB, or the increase of bitrate is about 0.24% on average, which is negligible. The foregoing result analysis indicates that SDE can efficiently reduce the coding time while maintaining almost the same coding performance as the 3DHEVC encoder.
3.2. Combined Results
In the following, we will analyze the experimental result of the proposed overall algorithm, which incorporates FCUS and SDE. The comparison results of the overall algorithm are shown in Table 5. The proposed overall algorithm can greatly reduce coding time for all sequences. It reduces 43% coding time under average texture videos and average rendered views conditions and achieves the better gain in coding speed compared to FCUS and SDE. Also a consistent gain in coding speed for all test sequences with the lowest gain of 36% for “Poznan_Street” and the highest gain for 53% for “Poznan_Hall2” is shown. The computation reduction is particularly high because the exhaustive CU size decision procedures of a significant number of CUs are reasonably skipped, and disparity estimation procedures of a significant number of CUs are not processed by the 3DHEVC encoder. On the other hand, the coding efficiency loss is negligible; specifically, for average texture videos coding, the average PSNR drop is 0.055 dB, and the increase of bitrate is about 1.65% on average. For average rendered views coding, the average PSNR drop is 0.039 dB, and the increase of bitrate is about 1.16% on average.

Figure 4 gives RD curves of the proposed algorithms compared to 3DHEVC, which are the total bitrates for the multiview texture video and depth map coding and average PSNR over all virtual views. As shown in Figure 4, the proposed algorithms (FCUS and SDE and the overall algorithm) perform almost the same coding efficiency from low to high bitrate compared to 3DHEVC. Therefore, the proposed algorithm can efficiently reduce coding time while keeping nearly the same RD performance as 3DHEVC.
(a) Kendo
(b) Balloons
(c) Newspaper
(d) Shark
(e) Undo_Dancer
(f) GT_Fly
(g) Poznan_Street
(h) Poznan_Hall2
3.3. Results of the Proposed Overall Algorithm Comparison with the StateoftheArt Fast Algorithm
The comparison results of the overall algorithm and a stateoftheart fast algorithm (contentadaptive complexity reduction scheme, CACRS [14]) are given in Table 6. Experimental results shown in Table 6 indicate that the proposed overall algorithm consistently outperforms CACRS. The proposed overall algorithm can save 12% encoding time on average compared to CACRS, with the lowest gain of 3% for “Undo_Dancer” and the highest gain of 22% for “Poznan_Hall2.” Additionally, the proposed overall algorithm achieves a better coding performance. For average texture videos coding, with 0.14 dB PSNR increases or 1.96% bitrate decreases compared to CACRS. For average rendered views coding, with 0.12 dB PSNR increases or 1.61% bitrate decreases compared to CACRS. Therefore, the proposed overall algorithm is more efficient than CACRS with better time saving and fewer bits. The above experimental results indicate that the proposed overall algorithm is efficient for all test sequences and consistently outperforms the recent fast algorithm for 3DHEVC.

4. Conclusion
In this paper, we propose a low complexity mode decision algorithm to reduce the computational complexity of the 3DHEVC encoder, which includes two fast approaches: fast CU size decision approach and selective disparity estimation approach. The recent 3DHEVC reference software 3DVHTM is applied to evaluate the proposed algorithm. The comparative experimental results show that the proposed algorithm can significantly reduce the computational complexity of 3DHEVC while maintaining almost the same RD performances.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
Acknowledgments
The authors would like to thank the editors and anonymous reviewers for their valuable comments. This work was supported in part by the National Natural Science Foundation of China under Grant nos. 61302118, 61374014, 61340059, and 61272038, the Scientific and Technological Project of Henan Province under Grant no. 14A520034, and in part by the Doctorate Research Funding of Zhengzhou University of Light Industry, under Grant no. 2013BSJJ047.
References
 K. Müller, H. Schwarz, D. Marpe et al., “3D highefficiency video coding for multiview video and depth data,” IEEE Transactions on Image Processing, vol. 22, no. 9, pp. 3366–3378, 2013. View at: Publisher Site  Google Scholar  MathSciNet
 P. Kauff, N. Atzpadin, C. Fehn et al., “Depth map creation and imagebased rendering for advanced 3DTV services providing interoperability and scalability,” Signal Processing: Image Communication, vol. 22, no. 2, pp. 217–234, 2007. View at: Publisher Site  Google Scholar
 Q. Zhang, P. An, Y. Zhang, L. Shen, and Z. Zhang, “Efficient depth map compression for view rendering in 3D video,” Imaging Science Journal, vol. 61, no. 4, pp. 385–395, 2013. View at: Publisher Site  Google Scholar
 G. Tech, K. Wegner, Y. Chen, and S. Yea, “3DHEVC test model 2,” Joint Collaborative Team on 3D Video Coding Extensions (JCT3V) document JCT3VB1005, 2nd Meeting: Shanghai, China, 2012. View at: Google Scholar
 Q. Zhang, L. Tian, L. Huang, X. Wang, and H. Zhu, “Rendering distortion estimation model for 3D high efficiency depth coding,” Mathematical Problems in Engineering, vol. 2014, Article ID 940737, 7 pages, 2014. View at: Publisher Site  Google Scholar
 B. Bross, W.J. Han, J.R. Ohm, G. Sullivan, Y.K. Wang, and T. Wiegand, “High Efficiency Video Coding (HEVC) text specification draft 10 (for FDIS & Last Call,” ITUT SG16 WP3 & ISO/IEC JTC1/SC29/WG11 JCTVCL1003, January 2013. View at: Google Scholar
 Q. Zhang, N. Li, and Y. Gan, “Effective early termination algorithm for depth map intra coding in 3DHEVC,” Electronics Letters, vol. 50, no. 14, pp. 994–996, 2014. View at: Publisher Site  Google Scholar
 L. Shen, Z. Liu, X. Zhang, W. Zhao, and Z. Zhang, “An effective CU size decision method for HEVC encoders,” IEEE Transactions on Multimedia, vol. 15, no. 2, pp. 465–470, 2013. View at: Publisher Site  Google Scholar
 L. Zhao, L. Zhang, S. Ma, and D. Zhao, “Fast mode decision algorithm for intra prediction in HEVC,” in Proceedings of the IEEE Visual Communications and Image Processing (VCIP '11), pp. 1–4, November 2011. View at: Publisher Site  Google Scholar
 T. L. da Silva, L. V. Agostini, and L. A. da Silva Cruz, “Fast HEVC intra prediction mode decision based on EDGE direction information,” in Proceedings of the 20th European Signal Processing Conference (EUSIPCO '12), pp. 1214–1218, August 2012. View at: Google Scholar
 G. Correa, P. Assuncao, L. Agostini, and L. A. da Silva Cruz, “Complexity control of high efficiency video encoders for powerconstrained devices,” IEEE Transactions on Consumer Electronics, vol. 57, no. 4, pp. 1866–1874, 2011. View at: Publisher Site  Google Scholar
 S. Cho and M. Kim, “Fast CU splitting and pruning for suboptimal CU partitioning in HEVC intra coding,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 23, no. 9, pp. 1555–1564, 2013. View at: Google Scholar
 E. Mora, J. Jung, M. Cagnazzo, and B. PesquetPopescu, “Initialization, limitation and predictive coding of the depth and texture quadtree in 3DHEVC Video Coding,” IEEE Transactions on Circuits and Systems for Video Technology, no. 99, p. 1, 2013. View at: Publisher Site  Google Scholar
 H. R. Tohidypour, M. T. Pourazad, P. Nasiopoulos, and V. Leung, “A content adaptive complexity reduction scheme for HEVCbased 3D video coding,” in Proceedings of the 18th International Conference on Digital Signal Processing (DSP '13), pp. 1–5, July 2013. View at: Google Scholar
 L. Shen, Z. Zhang, and P. An, “Fast CU size decision and mode decision algorithm for HEVC intra coding,” IEEE Transactions on Consumer Electronics, vol. 59, no. 1, pp. 207–213, 2013. View at: Publisher Site  Google Scholar
 J. Lu, H. Cai, J. Lou, and J. Li, “An epipolar geometrybased fast disparity estimation algorithm for multiview image and video coding,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 17, no. 6, pp. 737–750, 2007. View at: Publisher Site  Google Scholar
 Q. Zhang, P. An, Y. Zhang, L. Shen, and Z. Zhang, “Low complexity multiview video plus depth coding,” IEEE Transactions on Consumer Electronics, vol. 57, no. 4, pp. 1857–1865, 2011. View at: Publisher Site  Google Scholar
 G. Tech, K. Wegner, Y. Chen, and S. Yea, “3DHEVC draft text 1,” in Proceedings of the 5th Meeting of Joint Collaborative Team on 3D Video Coding Extensions (JCT3V), Document JCT3VE1001, Vienna, Austria, August 2013. View at: Google Scholar
 D. Rusanovskyy, K. Mueller, and A. Vetro, “Common test conditions of 3DV core experiments,” in Proceedings of the 5th Meeting of Joint Collaborative Team on 3D Video Coding Extensions (JCT3V '13), Vienna, Austria, August 2013. View at: Google Scholar
 M. Tanimoto, T. Fujii, and K. Suzuki, “View Synthesis Algorithmin View Synthesis Reference Software 2.0 (VSRS 2.0),” ISO/IEC JTC1/SC29/WG11 M16090, Lausanne, Switzerland, 2008. View at: Google Scholar
 G. Bjontegaard, “Calculation of average PSNR difference between RDcurves,” in Proceedings of the 13th VCEGM33 Meeting, Austin, Tex, USA, April 2001. View at: Google Scholar
 H. W. Chang, H. Yang, Y. Gan, and M. H. Wang, “Sparse feature fidelity for perceptual image quality assessment,” IEEE Transactions on Image Processing, vol. 22, no. 10, pp. 4007–4018, 2013. View at: Publisher Site  Google Scholar  MathSciNet
Copyright
Copyright © 2014 Qiuwen Zhang et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.