Research Article  Open Access
HsiChin Hsin, TzeYun Sung, YawShih Shieh, Carlo Cattani, "Adaptive Binary Arithmetic CoderBased Image Feature and Segmentation in the Compressed Domain", Mathematical Problems in Engineering, vol. 2012, Article ID 490840, 14 pages, 2012. https://doi.org/10.1155/2012/490840
Adaptive Binary Arithmetic CoderBased Image Feature and Segmentation in the Compressed Domain
Abstract
Image compression is necessary in various applications, especially for efficient transmission over a bandlimited channel. It is thus desirable to be able to segment an image in the compressed domain directly such that the burden of decompressing computation can be avoided. Motivated by the adaptive binary arithmetic coder (MQ coder) of JPEG2000, we propose an efficient scheme to segment the feature vectors that are extracted from the code stream of an image. We modify the Compressionbased Texture Merging (CTM) algorithm to alleviate the influence of overmerging problem by making use of the rate distortion information. Experimental results show that the MQ coderbased image segmentation is preferable in terms of the boundary displacement error (BDE) measure. It has the advantage of saving computational cost as the segmentation results even at low rates of bits per pixel (bpp) are satisfactory.
1. Introduction
Image segmentation is important in many applications, ranging from industrial monitoring to medical diagnosis. Among numerous techniques, the featurebased approach has received a lot of attention due largely to its computational efficiency [1]. However, the segmentation result is dependent on the selection of feature vectors [2–7]. Early research work on feature extraction is mainly at a single scale. It is noted that an image is decomposed into bandpass subimages by simple visual cortical cells in the human visual system (HVS) [8], which can be modeled by Gabor filters with spatial frequencies and orientations properly tuned [9]. Wavelet transform (WT) provides an efficient multiresolution representation, in which the higher detail information of an image is projected onto the shorter basis function with higher spatial resolution, and the lower detail information is projected onto the larger basis function with higher spectral resolution. This property matches the characteristics of HVS [10]. Various WTbased schemes were proposed to extract image features at multiple scales [11–14]. In addition, the advantage of WT is to take account of the phenomena of multiscales [15–17], which is fundamental in nonlinear time series [18–20] and fractal time series [21].
With the rapid growth of multimedia technologies [22–25] and the Internet applications, image compression is still in great demand [26]. As one can see, it is desirable to extract image features in the compressed domain directly, such that the burden of decompressing an image can be avoided [14, 27, 28]. The Joint Photographic Expert Group (JPEG) standard shows satisfactory results at moderate compression rates. The JPEG2000 standard, which adopts WT as the underlying transform, is preferable for additional advantages, for example, embedded coding and progressive transmission [29, 30]. In embedded coding, the original image is coded into a single code stream, from which the decoded image at any bit rate can be obtained. For progressive transmission, which is especially beneficial to the image browsing and Internet streaming applications, JPEG2000 uses the postcompression rate distortion (PCRD) algorithm to arrange the code stream of an image in decreasing order of information importance [31]. It is based on the rate distortion theory; more specifically, the rate distortion slope (RDS) should be nonincreasing as the number of coding bits increases. For the image segmentation applications, two interesting questions are thus raised. (1) Is it possible that image segmentation can be carried out in the compressed domain such that the burden of decoding computations can be avoided? (2) Is there a common piece of information, based on which image features can be constructed at both encoder and decoder? If so, there is no need to transmit these features from encoder to decoder.
This paper presents an efficient scheme to segment an image in the compressed domain. It is a twostep algorithm. In the first step, the MQ coderbased image features are coarsely clustered into small regions known as superpixels by using the simple means algorithm. The inherently oversegmented superpiexels are merged recursively by using the Compressionbased Texture Merging (CTM) algorithm [32]. In order to avoid over merging, we propose a simple RDSbased method to terminate CTM accordingly. The remainder of this paper proceeds as follows. In Section 2, the JPEG2000 standard is briefly reviewed. In Section 3, the modified CTM algorithm with the MQ coderbased image features is proposed to segment JPEG2000 images. Experimental results are presented in Section 4. Conclusion is given in Section 5.
2. Introduction to JPEG2000
The core of JPEG2000 is the embedded block coding with optimized truncation (EBCOT) algorithm [29], which adopts wavelet transform (WT) as the underlying method for subband decompositions. WT provides many desirable properties, for example, joint spacespatial frequency localization with orientation selectivity, selfsimilarity of wavelet coefficients across subbands of the same orientation, and energy clustering within each subband [10]. Among various WTbased image features, the commonly used are magnitude, energy, the generalized Gaussian distribution signature, and the cooccurrence measures [11–14].
EBCOT is a twotier algorithm. Tier1 consists of bitplane coding (BPC) followed by arithmetic coding (AC). Tier2 aims for optimal rate control. In BPC, three coding passes, namely, the significance propagation (SP) pass, the magnitude refinement (MR) pass, and the clean up (CU) pass, are involved with four primitive coding operations, namely, the significance coding operation, the sign coding operation, the magnitude refinement coding operation, and the cleanup coding operation. For a wavelet coefficient that is currently insignificant, if any of the 8 neighboring coefficients are already significant, it is coded in the SP pass using the significance coding operation; otherwise, it is coded in the CU pass using the cleanup coding operation. If this coefficient becomes significant, the sign is coded immediately using the sign coding operation. In the MR pass, magnitudes of the significant coefficients are updated using the magnitude refinement coding operation. The output bit streams of coding passes can be further coded by using a contextbased arithmetic coder known as the MQ coder to improve the compression performance. Based on the 8 neighboring coefficients, the MQ coder defines 18 context labels with their respective probability modes stored in the MQ table [29].
In JPEG2000, a large image can be partitioned into nonoverlapped subimages called tiles, each tile is decomposed into subbands by WT, each subband is divided into small blocks called code blocks, and each code block is independently coded from the most significant bitplane to the least significant bitplane. For optimal rate control, JPEG2000 adopts the postcompression rate distortion (PCRD) algorithm. Specifically, let be the code blocks of an image. The embedded code stream of can be terminated at some point, say , with a bit rate denoted by ; all the end points of coding passes are possible truncation points. PCRD selects the optimal truncation points to minimize the overall distortion: subject to the rate constraint: , where denotes the distortion incurred by discarding the coding passes after , and is the target bit rate. It is noted that the coding passes with nonincreasing rate distortion slopes (RDS) are candidates for the optimal truncation points. Based on the above, we propose an efficient scheme to segment JPEG2000 images in the following section.
3. Image Segmentation in the JPEG2000 Domain
In this section, we modify the Compressionbased Texture Merging (CTM) algorithm [32] to segment the MQ coderbased image features [28] in an adaptive manner. As a result, the image segmentation task can be conducted in the JPEG2000 domain directly, and the burden of decompressing computation can be avoided.
3.1. The MQ CoderBased Image Feature
The distribution of wavelet coefficients known as the wavelet histogram has been widely used for image segmentation. As the binary variables of wavelet coefficients are almost independent across bitplanes [14], the joint probability mass function (PMF) representing the wavelet histogram can be approximated as where is the absolute value of a wavelet coefficient, , which can be written by is the number of bitplanes, and is the bitplane’s PMF. Based on the MQ table defined in JPEG2000, we proposed a simple scheme to estimate the local PMF [28]. Specifically, let be the probability of 1 bit for variable on the bitplane, which can be obtained from the MQ table as follows: where is the probability of less probable symbol (LPS) stored in the MQ table, and MPS stands for more probable symbol. Note that the set obtained from the MQ table can be used to estimate the local PMF. As the MQ table is available at both encoder and decoder, there is no need to transmit the overhead information to construct the MQ features.
3.2. The Modified CTM Algorithm
In this section, we modify the CTM algorithm [32] to segment the MQ feature vectors of an image. With a set of MQ feature vectors, the number of coding bits can be approximated as where is the mean vector, is the covariance matrix, is the distortion incurred, is the feature dimension, and is the number of feature vectors. For sets of MQ feature vectors, the total number of coding bits is given by where and are the number of coding bits obtained by (3.4) and the number of MQ feature vectors in the set, respectively, and is the total number of MQ feature vectors, that is, . The idea behind CTM is to merge two sets of feature vectors such that the coding bits can be reduced maximally. The pairwise merging procedure of CTM is performed iteratively until no merge can reduce the coding bits any more. As mentioned in [32], the termination of CTM is dependent on the distortion parameter, , which can be determined by where is the distance between a pair of segments with respect to .
Motivated by the rate distortion theory, which has been widely used in embedded image coding for optimal rate control, we propose a simple scheme to determine the candidates of . Specifically, for a sequence of increasing distortion values: , the number of segments and the total number of coding bits are monotonically decreasing, that is, and . The rate distortion slope (RDS) is thus defined as where and is the number of MQ feature vectors. As RDS should be nondecreasing, that is, if , can be considered as a candidate to terminate the merging process of CTM. Thus, we modify the selection of as follows:
Figure 1 depicts flowchart of the modified CTM with the RDSbased adaptive selection of , where the MQ coderbased image features are projected into a lowdimensional space via principal component analysis (PCA) in order to reduce the computational cost further, and the initial superpixels are obtained by using the simple means algorithm. Take the image shown in Figure 2(a) as an example; the candidates of are shown in Figure 2(e), where the horizontal and vertical axes are the distortion and the RDS values, respectively. Figures 2(b)–2(d) show the segmentation results with the first, second, and third candidates of . As one can see, the rate distortion information can be used to avoid overmerging of CTM.
(a)
(b)
(c)
(d)
(e)
4. Experimental Results
The proposed algorithm has been extensively evaluated on the Berkeley database [33]. The 9/7wavelet filters adopted by JPEG2000 are used to extract the MQ coderbased image features. The number of initial superpixels is set to 50. In addition to visual inspection, the boundary displacement error (BDE) and the probabilistic Rand index (PRI) [34] are used for quantitative evaluation. The segmentation results are compared with CTM, MeanShift and NCuts. In MeanShift, the parameters and are set to 13 and 19, respectively; in NCuts, the number of segments is 20. The threshold of CTM is set to 0.1, as suggested in [32].
We first evaluate the segmentation performance at various compression rates. Figure 3(a) shows a test image with two Brodatz textures, namely, wood and grass. Figure 3(b) depicts percentages of errors at various rates of bits per pixel (bpp). It is noted that the segmentation results even at low bpp rates are satisfactory; thus, a small portion of code stream is sufficient for the segmentation task. It has the advantage of saving transmission time, computational cost, and memory space, which are desirable especially for the Internet applications.
(a)
(b)
Table 1 shows the BDE and PRI performances compared to CTM, MeanShift, and NCuts. The proposed algorithm is preferable in terms of the average BDE.

The improvements in PRI and BDE using (3.10) are shown in Figures 4(a) and 4(b), respectively, where the horizontal axis is the threshold: . It is shown that the proposed algorithm is more robust by taking account of the rate distortion information to avoid overmerging.
(a)
(b)
Figures 5, 6, 7, 8, 9, and 10 are representative of the Landscape, Objects, Urban, Water, Portraits, and Animals images in the Berkeley database. The original images are shown in the left column. The segmentation results by using the proposed algorithm and the CTM algorithm are given in the middle and right columns, respectively. It is noted that, for images with highdetail contents, the proposed algorithm improves the segmentation results visually.
5. Conclusion
The MQ coder provides effective probability models, which is available at both encoder and decoder and therefore can be used to extract image features in the JPEG2000 domain directly. As a result, no overhead transmission is necessary to extract the feature vectors, and moreover the burden of decompressing a JPEG2000 image can be avoided. Based on the MQ coder, an efficient scheme of segmenting an image has been proposed. In order to avoid overmerging, the CTM algorithm has been modified by taking account of the rate distortion information. The proposed algorithm has been evaluated on images with Brodatz textures and the Berkeley image database. It is shown that the segmentation results at lowmiddle bpp rates are rather promising. In addition, for images with highdetail contents, the proposed algorithm is preferable in terms of the average BDE measure and visual comparison.
Acknowledgments
The authors are grateful to the maintainers of the Berkeley image database. The National Science Council of Taiwan under Grants NSC1002628E239002MY2 supported this work.
References
 A. K. Jain, “Data clustering: 50 years beyond Kmeans,” Pattern Recognition Letters, vol. 31, no. 8, pp. 651–666, 2010. View at: Publisher Site  Google Scholar
 N. A. M. Isa, S. A. Salamah, and U. K. Ngah, “Adaptive fuzzy moving Kmeans clustering algorithm for image segmentation,” IEEE Transactions on Consumer Electronics, vol. 55, no. 4, pp. 2145–2153, 2009. View at: Publisher Site  Google Scholar
 S. K. Mitra and J. F. Kaiser, Handbook for Digital Signal Processing, John Wiley & Sons, New York, NY, USA, 1993.
 J. S. Bendat and A. G. Piersol, Random Data: Analysis and Measurement Procedure, John Wiley & Sons, New York, NY, USA, 3rd edition, 2000.
 M. Li, W. S. Chen, and L. Han, “Correlation matching method for the weak stationarity test of LRD traffic,” Telecommunication Systems, vol. 43, no. 34, pp. 181–195, 2010. View at: Publisher Site  Google Scholar
 M. Li and W. Zhao, “Variance bound of ACF estimation of one block of fGn with LRD,” Mathematical Problems in Engineering, vol. 2010, Article ID 560429, 14 pages, 2010. View at: Publisher Site  Google Scholar
 M. Li, “A method for requiring block size for spectrum measurement of ocean surface waves,” IEEE Transactions on Instrumentation and Measurement, vol. 55, no. 6, pp. 2207–2215, 2006. View at: Publisher Site  Google Scholar
 T. N. Tan and A. G. Constantinides, “Texture analysis based on a human visual model,” in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP '90), vol. 4, pp. 2137–2140, Albuquerque, NM, USA, April 1990. View at: Publisher Site  Google Scholar
 D. E. Ilea and P. F. Whelan, “CTex—an adaptive unsupervised segmentation algorithm based on colortexture coherence,” IEEE Transactions on Image Processing, vol. 17, no. 10, pp. 1926–1939, 2008. View at: Publisher Site  Google Scholar
 S. Mallat, A Wavelet Tour of Signal Processing, Elsevier/Academic Press, Amsterdam, The Netherlands, 3rd edition, 2009.
 M. K. Bashar, N. Ohnishi, and K. Agusa, “A new texture representation approach based on local feature saliency,” Pattern Recognition and Image Analysis, vol. 17, no. 1, pp. 11–24, 2007. View at: Publisher Site  Google Scholar
 C. M. Pun and M. C. Lee, “Extraction of shift invariant wavelet features for classification of images with different sizes,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 26, no. 9, pp. 1228–1233, 2004. View at: Publisher Site  Google Scholar
 J. Sun, D. Gu, S. Zhang, and Y. Chen, “Hidden Markov Bayesian texture segmentation using complex wavelet transform,” IEE Proceedings: Vision, Image and Signal Processing, vol. 151, no. 3, pp. 215–223, 2004. View at: Publisher Site  Google Scholar
 M. H. Pi, C. S. Tong, S. K. Choy, and H. Zhang, “A fast and effective model for wavelet subband histograms and its application in texture image retrieval,” IEEE Transactions on Image Processing, vol. 15, no. 10, pp. 3078–3088, 2006. View at: Publisher Site  Google Scholar
 M. Li, W. Zhao, and S. Chen, “MBmbased scalings of traffic propagated in internet,” Mathematical Problems in Engineering, vol. 2011, Article ID 389803, 21 pages, 2011. View at: Publisher Site  Google Scholar
 M. Li and W. Zhao, “Representation of a stochastic traffic bound,” IEEE Transactions on Parallel and Distributed Systems, vol. 21, no. 9, pp. 1368–1372, 2010. View at: Publisher Site  Google Scholar
 M. Li, “Generation of teletraffic of generalized Cauchy type,” Physica Scripta, vol. 81, no. 2, Article ID 025007, 2010. View at: Publisher Site  Google Scholar
 H. Kantz and T. Schreiber, Nonlinear Time Series Analysis, Cambridge University Press, Cambridge, UK, 2nd edition, 2004.
 T. Schreiber, “Interdisciplinary application of nonlinear time series methods,” Physics Report, vol. 308, no. 1, pp. 1–64, 1999. View at: Google Scholar
 B. J. West, E. L. Geneston, and P. Grigolini, “Maximizing information exchange between complex networks,” Physics Reports, vol. 468, no. 1–3, pp. 1–99, 2008. View at: Publisher Site  Google Scholar
 M. Li, “Fractal time series—a tutorial review,” Mathematical Problems in Engineering, vol. 2010, Article ID 157264, 26 pages, 2010. View at: Publisher Site  Google Scholar
 S. Y. Chen, H. Tong, and C. Cattani, “Markov models for image labeling,” Mathematical Problems in Engineering, vol. 2011, Article ID 814356, 18 pages, 2011. View at: Publisher Site  Google Scholar
 S. Y. Chen, H. Tong, Z. Wang, S. Liu, M. Li, and B. Zhang, “Improved generalized belief propagation for vision processing,” Mathematical Problems in Engineering, vol. 2011, Article ID 416963, 12 pages, 2011. View at: Publisher Site  Google Scholar
 S. Y. Chen, J. Zhang, Q. Guan, and S. Liu, “Detection and amendment of shape distortions based on moment invariants for active shape models,” IET Image Processing, vol. 5, no. 3, pp. 273–285, 2011. View at: Publisher Site  Google Scholar
 F. R. Alexander, P. V. Juan, and S. Hichem, “Estimating quality bounds of JPEG 2000 compressed leukocytes images,” in Proceedings of the 2nd Mexican Conference on Pattern Recognition, pp. 27–29, The National Institute of Astrophysics, Puebla, Mexico, 2010. View at: Google Scholar
 T. Y. Sung and H. C. Hsin, “A hybrid image coder based on SPIHT algorithm with embedded block coding,” IEICE Transactions on Fundamentals of Electronics, Communications and Computer Sciences, vol. E90A, no. 12, pp. 2979–2984, 2007. View at: Publisher Site  Google Scholar
 F. Zargari, A. Mosleh, and M. Ghanbari, “A fast and efficient compressed domain JPEG2000 image retrieval method,” IEEE Transactions on Consumer Electronics, vol. 54, no. 4, pp. 1886–1893, 2008. View at: Publisher Site  Google Scholar
 H.C. Hsin, “Texture segmentation in the joint photographic expert group 2000 domain,” IET Image Processing, vol. 5, no. 6, pp. 554–559, 2011. View at: Publisher Site  Google Scholar
 T. Acharya and P. S. Tsai, JPEG2000 Standard for Image Compression, John Wiley & Sons, New York, NY, USA, 2005.
 F. AuliLlinas, J. SerraSagristà, and J. BartrinaRapesta, “Enhanced JPEG2000 quality scalability through blockwise layer truncation,” EURASIP Journal on Advances in Signal Processing, vol. 2010, Article ID 803542, 11 pages, 2010. View at: Publisher Site  Google Scholar
 A. K. Gupta, S. Nooshabadi, and D. Taubman, “Novel distortion estimation technique for hardwarebased JPEG2000 encoder system,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 17, no. 7, pp. 918–923, 2007. View at: Publisher Site  Google Scholar
 A. Y. Yang, J. Wright, Y. Ma, and S. S. Sastry, “Unsupervised segmentation of natural images via lossy data compression,” Computer Vision and Image Understanding, vol. 110, no. 2, pp. 212–225, 2008. View at: Publisher Site  Google Scholar
 http://www.eecs.berkeley.edu/~yang/software/lossy_segmentation/.
 R. Unnikrishnan, C. Pantofaru, and M. Hebert, “Toward objective evaluation of image segmentation algorithms,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 29, no. 6, pp. 929–944, 2007. View at: Publisher Site  Google Scholar
Copyright
Copyright © 2012 HsiChin Hsin et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.