Research Article | Open Access

Jian Zheng, Pei-Rong Lu, Dehui Xiang, Ya-Kang Dai, Zhao-Bang Liu, Duo-Jie Kuai, Hui Xue, Yue-Tao Yang, "Retinal Image Graph-Cut Segmentation Algorithm Using Multiscale Hessian-Enhancement-Based Nonlocal Mean Filter", *Computational and Mathematical Methods in Medicine*, vol. 2013, Article ID 927285, 7 pages, 2013. https://doi.org/10.1155/2013/927285

# Retinal Image Graph-Cut Segmentation Algorithm Using Multiscale Hessian-Enhancement-Based Nonlocal Mean Filter

**Academic Editor:**Yujie Lu

#### Abstract

We propose a new method to enhance and extract the retinal vessels. First, we employ a multiscale Hessian-based filter to compute the maximum response of vessel likeness function for each pixel. By this step, blood vessels of different widths are significantly enhanced. Then, we adopt a nonlocal mean filter to suppress the noise of enhanced image and maintain the vessel information at the same time. After that, a radial gradient symmetry transformation is adopted to suppress the nonvessel structures. Finally, an accurate graph-cut segmentation step is performed using the result of previous symmetry transformation as an initial. We test the proposed approach on the publicly available databases: DRIVE. The experimental results show that our method is quite effective.

#### 1. Introduction

The retina is the only tissue in human body from which the information of blood vessel can be directly obtained in vivo. The information of retinal vessel plays an important role in the diagnosis and treatment of various diseases such as glaucoma [1], age-related macular degeneration [2], degenerative myopia, and diabetic retinopathy [3]. Recently, it is also found that the detection of vascular geometric change might be meaningful to judge whether people have high blood pressure or cardiovascular disease [4]. Retinal vessel extraction is quite essential for ophthalmologists to diagnose various eye diseases. Moreover, accurately segmented vessels could be very helpful for feature-based retinal image registration.

Generally, existing retinal vessel segmentation methods can be roughly divided into two categories. The first category is supervised learning-based method. Such methods require tremendous manual segmentations of vasculature to train the classifier. Staal et al. [5] first adopt a ridge-extraction method to separate the image into numerous patches. For each pixel in the patch, feature vector that contains profile information such as width, height, and edge strength is computed. Feature vectors are then classified using a *k* nearest neighbors classifier with sequential forward feature selection strategy. Soares et al. [6] selected pixel value and multiscale 2D Gabor wavelet coefficients to construct the feature vector. A Bayesian classifier based on class-conditional probability density functions was adopted to perform a fast classification and model complex decision surfaces. Marín et al. [7] proposed a new supervised method to segment retinal vessels. For each pixel, 5 gray-level descriptors and 2 moment invariants-based features of a squared window were computed as the feature vector. A multilayer neural network scheme was then adopted to classify each pixel. The performance of such method depends on the correlation between training data and test data. If these two datasets are quite different, the segmentation results may be less than ideal. Besides that, the manual segmentation step would bring additional burden to ophthalmologists. Therefore, such methods have not been widely used in clinic yet.

The second category is rule-based method. Such methods usually need to extract neighborhood information of each pixel. The information is then used to label the pixel according to some f preset rules. Matched filter [8], model-based method [9], and morphology-based method [10] all belong to this category. In the literature [8], Sofka and Stewart proposed a multiscale Gaussian and Gaussian-derivative profile kernel to detect vessels at a variety of widths. Lam et al. [9] proposed a multiconcavity modeling approach to handle unhealthy retinal images. In Lam’s approach, a lineshape concavity measure was used to remove dark lesions and a locally normalized concavity measure is designed to remove spherical intensity variation. The two concavity measures are combined together according to their statistical distributions to detect vessels in general retinal images. Mendonça and Campilho [10] proposed a region growing algorithm in the morphological processed image to extract vascular centerline. First, four-direction-based differential operator was adopted to detect vascular centerlines. These centerlines were then merged to complete vascular centerlines and selected as seed points. Finally, region growing step was performed to reconstruct the vasculature. These methods do not require training steps and the interactions with doctors are also minimized; thus rule-based methods have been widely used in clinical application. However, ideal automatic retinal vessel segmentation is still not easy. This can mainly be ascribed to two reasons. One is that the contrast of some tiny vessels may be quite low, especially in some pathologies affected images. The other reason is that the image noise such as edge blurring may disturb the final segmentation results.

Recently, graph-cut methods [11–15] have been very popular in image segmentation. This is because graph-cut methods could reach the global optimal value of the predefined energy function. Besides that, the user interactions of such methods are also very simple. Several graph-cut-based methods have been proposed to solve retina image segmentations. Chen et al. [16] proposed a 3D graph-search-graph-cut method to segment multilayers of 3D OCT retinal images. The multi-layers of retina and the symptomatic exudate-associated derangements (SEAD) are successfully segmented. Inspired by the superior performance of graph-cut method, we propose a new method to extract the blood vessels in retinal images following our previous work [17]. First, we perform a novel multiscale Hessian-based filter to compute the maximum response of vessel likeness function for each pixel, which is used to enhance the blood vessels of gray retinal images. Then, we adopt a nonlocal mean filter to suppress the noise of the enhanced image. After that, a radial gradient symmetry transformation is adopted to improve the detection of vessel structures and suppress the nonvessel structures. Finally, an accurate graph-cut segmentation is performed using previous symmetry transformation as an initial.

#### 2. Methods

##### 2.1. Multiscale Hessian-Based Enhancement

In order to improve the contrast of retinal vasculature with different widths, we propose a multiscale Hessian-based enhancement. Frangi et al. [18] have proposed a method to detect the tubular structure based on the eigenvalues of Hessian matrix. We denote by and the eigenvalues of scale-related Hessian matrix, which is defined as follows: where is the second order differential of input image. For an ideal tubular structure, the eigenvalues generally will meet the following conditions: We then define a new scale-related vessel likeness function: where is defined as follow: where is an experiential parameter and we set it to .

The function value of indicates the saliency of tubular structure for each pixel. We search in the scale range to find the maximum response of vessel likeness function. Figure 2 shows the optimal scale property of the input image. The pixel value stands for the scale that is corresponding to the maximal function value of . As seen from Figure 2, the positive correlation between vessel width and optimal scale is obvious. The primary vessel owns a large scale property while the tiny vessel owns a small scale property.

Figure 3 gives an example of multiscale Hessian-based enhancement, in which the pixel value stands for the vessel likeness function value. As shown in Figure 3, the whole retinal vasculature is prominently enhanced. However, some nonvessel structures are also enhanced including optic disk, yellow spots, and speckles. These nonvessel structures need to be removed in the following step.

##### 2.2. Nonlocal Mean Filtering

The effect of multiscale hessian based enhancement is obvious. As shown in Figure 3, the whole vasculature is significantly strengthened. However, some nonvessel structures are also enhanced and cause the enhanced image to be noisy. In order to suppress image noise and maintain the structural information, we employ a nonlocal mean filtering [19] step. We denote the enhanced image as and the filtered image as . The detailed filtering step can be described as in the following equation: where denotes the filtering neighborhood and is the weighting factor, which is defined as follow: where denotes a feature vector which is composed of the pixel values of the nearby area around pixel and is a filtering parameter. The parameters of filtering neighborhood and nearby area need to be selected suitably so as to achieve a balance between filtering performance and computation cost. Figure 4 shows a result of nonlocal mean filtering. We can find that the noise of enhanced image is suppressed effectively while the vasculature is maintained at the same time.

##### 2.3. Radial Gradient Symmetry Transform

In order to remove the nonvessel structures, we propose a radial gradient symmetry transform method based on Loy’s work [20]. An ideal vessel structure is shown in Figure 5. As shown in Figure 5, we notice that the gradient vectors own a symmetric property in both the magnitude and direction. For those nonvessel structures, there is no such property. Therefore, we propose a symmetric vessel likeness function as follows: where stands for a vessel likeness function of a given point along the direction of gradient vector and is an indicator function that indicates whether point owns the gradient symmetric property. The computation of and consists of the following steps.(1) For each pixel in the filtered image as shown in Figure 4, we compute its vessel contribution along the gradient direction. The normalized gradient vector is denoted by . The coordinates of pixels that are affected by pixel are computed as follows: where , and is the optimal scale parameter of pixel , as shown in Figure 2.(2) For these affected pixels , we compute the vessel accumulation image by the following equation: (3) Then the is given by where is a normalization factor and is a radial strictness parameter.(4) For each pixel , we search in its neighborhood along the gradient direction with . As shown in Figure 5, if there exist points and that meet the following condition, , otherwise; : (5) For symmetric vessel likeness function , we set a vessel likeness threshold to extract a coarse vasculature. After that, we take an erosion step to get a narrowed retinal vasculature. We then compute the pixel number of connected vessels. If the pixel number is smaller than the preset threshold , we consider it as the background noise and it should be removed.

##### 2.4. Graph-Cut Step

Throughout the previous processing steps, we have got a coarsely extracted vasculature. The final step is to accurately segment vessels from previous results. This can be described as a pixel labeling problem which can be formulated using the energy function: where is a labeling set and is the data prior energy, which measures the cost of giving a label to a given pixel according to prior information. is the potential energy, which measures the smoothness of a neighboring pixel system and is a weighting parameter.

We adopt the graph-cut algorithm to optimize the energy function (12). The centerline is used as shape prior to guide the extraction process. In our framework, a graph is created with nodes corresponding to pixels of a retinal image, where is the set of all nodes and is the set of all links connecting neighboring nodes. The neighboring pixel system is constructed with eight neighboring pixels. The terminal nodes are defined as source and sink . As an initial, we extract the centerline of previously extracted vasculature. The pixels on the centerline are considered as definite foreground and the pixels with are classified as candidate foreground . The pixels with are classified as background and others are classified as the candidate background . That is, and .

For each pixel , we compute its minimum distances to and according to the literature [22], which are denoted as and , respectively. The cost of *t*-links can be computed as follows:
where , , and . The nodes in and are definitely labeled as and , respectively. The weight of *n*-link describes the labeling coherence of a pixel with its neighbors. We utilize the pixel value information as the neighborhood penalty item. The cost of *n*-links can be defined as follows:
where is a tiny number to avoid division by 0. When the graph-cut algorithm terminates, we encourage candidate foreground pixels to be labeled as foreground and discourage candidate background to be classified as background pixels.

#### 3. Experiments and Conclusion

We test our method on the publicly available databases: DRIVE [5]. Three measures sensitivity (SE), specificity (SP), and accuracy (AC) are used to evaluate the performance of our method in the image field of view. They are defined as follows: where is the number of correctly classified vessel pixels and is the number of the vessel pixels in ground truth. is the number of correctly classified nonvessel pixels and is the number of nonvessel pixels. is the total number of correctly classified pixels and is total number of pixels.

We test the proposed method on 40 images and compare it with the methods developed by Staal et al. [5], Mendonça and Campilho [10], Wang et al. [21], and Marín et al. [7]. The mean values and standard deviations (sd.) of these methods are shown in Table 1. Some values that cannot be obtained from the literature are denoted by NA.

From Table 1 we can see that there is a prominent improvement in sensitivity. This means that our method is quite effective in extracting some tiny vessels. On the other hand, our method is not as well as others’ in both specificity and accuracy. For clinical application, the accuracy should be better than 90%. Our method is over the qualified standard, but there are still large improvements need to be done.

Some of the extraction results are shown in Figures 6 and 7. As it can be seen from Figure 6, the majority of the vessels in the retina can be finely extracted; however, some nonvessel structures are also extracted and some tiny vessels are not correctly segmented. This is mainly because our method focused more on tiny vessel extraction, which is not easy for ophthalmologists to extract manually. Therefore some boundaries between retinal region and the background are mistakenly recognized as vessels. Besides that, the optic disk also cannot be fully excluded from the final result. Figure 7 shows two segmentation results of ill-conditioned retinal images, where some speckles are hard to be differentiated from tiny vessels. Our method fails to extract an accurate vasculature, where the AC values are 0.8720 and 0.8814, respectively. Furthermore, we exchange the order of nonlocal mean filtering and multiscale hessian-based enhancement. Experimental results show that there is a little difference between the performances. This also indicates that nonlocal mean filter is very robust and could be widely used in image processing area.

**(a)**

**(b)**

**(c)**

**(d)**

**(e)**

**(f)**

**(a)**

**(b)**

**(c)**

**(d)**

In summary, this paper first presents a multiscale hessian-based enhancement for retinal images. Next, we adopt an effective nonlocal mean filtering step to suppress noise of the enhanced image. Then, we propose a radial gradient symmetry transform method to suppress the nonvessel artifacts. Finally, a graph-cut step is taken to accurately segment the retinal vessels. Experiments show that our method is very sensitive for the vessels segmentation, but the performance for some tiny vessel extraction and speckles exclusion is still needed to be improved. This will be our further work. We will make further studies to improve the performance.

#### Acknowledgments

This work is supported in part by the Natural Science Foundation of Jiangsu Province (Grant no. BK2011331), National Science Foundation of China (Grant no. 61201117), Special Funded Program on National Key Scientific Instruments and Equipment Development (Grant no. 2011YQ040082), and the 2nd Phase Major Project of Suzhou Institute of Biomedical Engineering and Technology (Grant no. Y053011305).

#### References

- C. Sanchez-Galeana, C. Bowd, E. Z. Blumenthal, P. A. Gokhale, L. M. Zangwill, and R. N. Weinreb, “Using optical imaging summary data to detect glaucoma,”
*Ophthalmology*, vol. 108, no. 10, pp. 1812–1818, 2001. View at: Publisher Site | Google Scholar - S. Wang, L. Xu, Y. Wang, Y. Wang, and J. B. Jonas, “Retinal vessel diameter in normal and glaucomatous eyes: the Beijing eye study,”
*Clinical and Experimental Ophthalmology*, vol. 35, no. 9, pp. 800–807, 2007. View at: Publisher Site | Google Scholar - G. Garhöfer, C. Zawinka, H. Resch, P. Kothy, L. Schmetterer, and G. T. Dorner, “Reduced response of retinal vessel diameters to flicker stimulation in patients with diabetes,”
*British Journal of Ophthalmology*, vol. 88, no. 7, pp. 887–891, 2004. View at: Publisher Site | Google Scholar - H. Yatsuya, A. R. Folsom, T. Y. Wong, R. Klein, B. E. K. Klein, and A. R. Sharrett, “Retinal microvascular abnormalities and risk of lacunar stroke: Atherosclerosis Risk in Communities Study,”
*Stroke*, vol. 41, no. 7, pp. 1349–1355, 2010. View at: Publisher Site | Google Scholar - J. Staal, M. Abràmoff, M. Niemeijer, M. A. Viergever, and B. van Ginneken, “Ridge-based vessel segmentation in color images of the retina,”
*IEEE Transactions on Medical Imaging*, vol. 23, no. 4, pp. 501–509, 2004. View at: Publisher Site | Google Scholar - J. V. B. Soares, J. J. G. Leandro, R. M. Cesar Jr., H. F. Jelinek, and M. J. Cree, “Retinal vessel segmentation using the 2-D Gabor wavelet and supervised classification,”
*IEEE Transactions on Medical Imaging*, vol. 25, no. 9, pp. 1214–1222, 2006. View at: Publisher Site | Google Scholar - D. Marín, A. Aquino, M. E. Gegúndez-Arias, and J. M. Bravo, “A new supervised method for blood vessel segmentation in retinal images by using gray-level and moment invariants-based features,”
*IEEE Transactions on Medical Imaging*, vol. 30, no. 1, pp. 146–158, 2011. View at: Publisher Site | Google Scholar - M. Sofka and C. V. Stewart, “Retinal vessel centerline extraction using multiscale matched filters, confidence and edge measures,”
*IEEE Transactions on Medical Imaging*, vol. 25, no. 12, pp. 1531–1546, 2006. View at: Publisher Site | Google Scholar - B. S. Y. Lam, Y. Gao, and A. W. C. Liew, “General retinal vessel segmentation using regularization-based multiconcavity modeling,”
*IEEE Transactions on Medical Imaging*, vol. 29, no. 7, pp. 1369–1381, 2010. View at: Publisher Site | Google Scholar - A. M. Mendonça and A. Campilho, “Segmentation of retinal blood vessels by combining the detection of centerlines and morphological reconstruction,”
*IEEE Transactions on Medical Imaging*, vol. 25, no. 9, pp. 1200–1213, 2006. View at: Publisher Site | Google Scholar - Y. Boykov, O. Veksler, and R. Zabih, “Fast approximate energy minimization via graph cuts,”
*IEEE Transactions on Pattern Analysis and Machine Intelligence*, vol. 23, no. 11, pp. 1222–1239, 2001. View at: Publisher Site | Google Scholar - Y. Boykov and V. Kolmogorov, “An experimental comparison of mincut/max-flow algorithms for energy minimization in vision,”
*IEEE Transactions on Pattern Analysis and Machine Intelligence*, vol. 26, no. 9, pp. 1124–1137, 2004. View at: Publisher Site | Google Scholar - Y. Boykov and G. Funka-Lea, “Graph cuts and efficient N-D image segmentation,”
*International Journal of Computer Vision*, vol. 70, no. 2, pp. 109–131, 2006. View at: Publisher Site | Google Scholar - X. Chen, J. K. Udupa, U. Bagci, Y. Zhuge, and J. Yao, “Medical image segmentation by combining graph cuts and oriented active appearance models,”
*IEEE Transactions on Image Processing*, vol. 21, no. 4, pp. 2035–2046, 2012. View at: Publisher Site | Google Scholar - D. Freedman and T. Zhang, “Interactive graph cut based segmentation with shape priors,” in
*Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR '05)*, pp. 755–762, June 2005. View at: Publisher Site | Google Scholar - X. Chen, M. Niemeijer, L. Zhang, K. Lee, M. D. Abramoff, and M. Sonka, “Three-dimensional segmentation of fluid-associated abnormalities in retinal OCT: probability constrained graph-search-graph-cut,”
*IEEE Transaction on Medical Imaging*, vol. 31, no. 8, pp. 1521–1531, 2012. View at: Publisher Site | Google Scholar - D. Xiang, J. Tian, K. Deng, X. Zhang, F. Yang, and X. Wan, “Retinal vessel extraction by combining radial symmetry transform and iterated graph cuts,” in
*Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society*, pp. 3950–3953, 2011. View at: Google Scholar - A. F. Frangi, W. J. Niessen, K. L. Vincken, and M. A. Viergever, “Multiscale vessel enhancement filtering,” in
*Medical Imaging Computing and Computer Assisted Interventions*, vol. 1496 of*Lecture Notes in Computer Sciences*, pp. 130–137, 1998. View at: Google Scholar - A. Buades, B. Coll, and J. M. Morel, “A non-local algorithm for image denoising,” in
*Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR '05)*, pp. 60–65, June 2005. View at: Google Scholar - G. Loy and A. Zelinsky, “Fast radial symmetry for detecting points of interest,”
*IEEE Transactions on Pattern Analysis and Machine Intelligence*, vol. 25, no. 8, pp. 959–973, 2003. View at: Publisher Site | Google Scholar - L. Wang, A. Bhalerao, and R. Wilson, “Analysis of retinal vasculature using a multiresolution hermite model,”
*IEEE Transactions on Medical Imaging*, vol. 26, no. 2, pp. 137–152, 2007. View at: Publisher Site | Google Scholar - Y. Li, J. Sun, C. Tang, and H.-Y. Shum, “Lazy snapping,”
*ACM Transactions on Graphics*, vol. 23, no. 3, pp. 303–308, 2004. View at: Publisher Site | Google Scholar

#### Copyright

Copyright © 2013 Jian Zheng et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.