International Journal of Biomedical Imaging

International Journal of Biomedical Imaging / 2017 / Article

Research Article | Open Access

Volume 2017 |Article ID 8147632 | 8 pages | https://doi.org/10.1155/2017/8147632

Saliency-Based Bleeding Localization for Wireless Capsule Endoscopy Diagnosis

Academic Editor: Victor Sanchez-Silva
Received27 Jun 2017
Revised03 Oct 2017
Accepted31 Oct 2017
Published28 Nov 2017

Abstract

Stomach bleeding is a kind of gastrointestinal disease which can be diagnosed noninvasively by wireless capsule endoscopy (WCE). However, it requires much time for physicians to scan large amount of WCE images. Alternatively, computer-assisted bleeding localization systems are developed where color, edge, and intensity features are defined to distinguish lesions from normal tissues. This paper proposes a saliency-based localization system where three saliency maps are computed: phase congruency-based edge saliency map derived from Log-Gabor filter bands, intensity histogram-guided intensity saliency map, and red proportion-based saliency map. Fusing the three maps together, the proposed system can detect bleeding regions by thresholding the fused saliency map. Results demonstrate the accuracy of 98.97% for our system to mark bleeding regions.

1. Introduction

Bleeding is one of the most common lesions about gastrointestinal (GI) disease [1]. The conventional detection method requires physicians to scan real-time images through an endoscope-attached wire, which is clinically invasive. Capsule endoscope camera, on the distal side of which a miniaturized sensor is mounted, can sweep past patients’ gut wall. Such noninvasive endoscopy, called wireless capsule endoscopy (WCE), can capture a video of nearly 57,000 frames, where the bleeding locations are recorded. Detecting and interpreting the lesions from numerous frames need physicians to be concentrated [2]. To ease the task of subjective diagnosis, computer-assisted bleeding localization systems are proposed [3].

The state-of-the-art bleeding localization systems depend mainly on color features. For instance, authors in [4] extracted mean and variance in the HSI color space: hue, saturation, and intensity. In [5], high-order statistical moments including kurtosis were added to the RGB color space. Similarly, statistical features based on intensity histogram [6] and local binary pattern are computed in the RGB and HSV color space. Unlike color features, edge was discarded since it increased false positive (FP) rate [7]. More importantly, these systems generally compose these statistical features from different color spaces into a vector, followed by a supervised pattern classification tool such as artificial neural network or support vector machine. As a result, the task of localization is degraded since classification fails to predict accurate positions of a lesion.

This paper proposes a saliency detection method that is devoted to localizing the bleeding regions in WCE images. Visual saliency models that extract the most interesting part from an image have been proposed in the nature image recognitions [811]. However, it is undesirable to apply the models directly to medical image analysis. Firstly, most existing saliency models assume that the visual fixation tends to stay within the center of images [12], while, in medical images, the fixation mainly refers to lesions which appear anywhere in the field of view. Secondly, salient regions in the natural images tend to be in the foreground where edge and color information is quite distinguishable; however, lesions in the WCE images vary according to the specific diseases. For instance, the bleeding lesions are small and dark red, while the colon lesions appear large and highly laminated. Thirdly, in the bleeding localization system, the salient region must exclude not only normal tissue, but also gastric fluid and undigested residue. The latter two are quite different from the normal, which can prevent the system from recognizing FP samples.

To detect the bleeding regions with both high sensitivity and specificity, the traditional methods take full advantage of color features and ignore other features. However, when diagnosing the bleeding area, physicians use multiple features, such as edge, intensity, and color. In order to simulate the diagnosis process, a visual saliency-based WCE bleeding detection system is proposed, which includes three kinds of saliency maps, color, intensity, and edge. The main contribution of this paper is that intensity and edge saliency maps are introduced to better imitate the diagnosis of physician. Salient intensity is extracted from the intensity histograms which have rotation invariance while edges are selected based on the phase congruency [9] of Log-Gabor filter bank which can solve the problem of nonuniform illumination of the WCE. Experimental results demonstrate that our algorithm has a very good performance for detecting the bleeding area. The rest of this paper is organized as follows. Materials and methods are presented in Section 2; Section 3 exhibits the experimental results; finally, some conclusion is drawn in Section 4.

2. Materials

A total of 200 WCE frames, with the pixel resolution of 480-by-480, are used with a warrant from Ankon Incorporation, Wuhan, China. Two physicians, with three-year clinical experience, drew bleeding masks independently; the two masks of each image are fused together by pixel-to-pixel AND operation. Randomly selected six bleeding images and masks are presented in Figure 1.

3. Saliency Map Extraction

Clinicians discriminate the bleeding lesions from a WCE image mainly based on some salient information. Here we classify the information into three categories: edge information, intensity information, and color information, each of which is quantified using a saliency map.

3.1. Edge Saliency Map

When clinicians calibrate the bleeding location, the clear contour of bleeding area is an important reference condition. Therefore, the outlines of bleeding area will occupy a large proportion in the clinician's attention mechanism to distinguish the bleeding area from the normal. In order to simulate the diagnosis of clinician, we propose the edge saliency map.

Generally, edges can be obtained from spatial edge detectors such as Roberts, Laplace, or Canny, but all these detectors are vulnerable to the image noise. Edge detection based on PC has the brightness invariance properties, which can compensate for the instability luminance of color channels [13]. The solution of extracting noise-insensitive edge is to convert an image into the spectrum domain and to calculate the phase congruency (PC [9]). Still, there are many frequency transformation tools like Fast Fourier Transformation (FFT), discrete cosine transformation (DCT), and Gabor filters. However, the FFT and DCT fail to be analyzed in the multiresolution of scales and orientations, which make the PC too coarse. Though Gabor filters can be used in the different scales and orientation, they tend to be bias frequency components towards the lowest band. Hence, the PCs calculated from Gabor frequency bands are not isolated. Here we choose the Log-Gabor wavelet transform because it is competent to extract the PC at the isolated center frequencies with symmetry attenuation responses [14]. The scales and orientations parameters are empirically set to 5 and 6, respectively. According to the spectrum analysis, an image can be decomposed into a combination of amplitude spectrum and phase spectrum. Accordingly, the Log-Gabor filter is constructed by multiplying the frequency response of the two components together in the polar coordinates system as follows:where represents the polar coordinates and is the center frequency of the filter and it is related to our current scale by in which minWave is the wavelength of smallest scale filter and mult is the scaling factor between the successive filters. is the orientation angle of the filter and and determine the scale bandwidth and the angular bandwidth, respectively. In our experiments, the parameters for the Log-Gabor filters are set as follows: , , ,  .

At each orientation, a PC map in the th scales is fused bywhere and denote the even and odd spectrums at the th scale. The constant stabilizes the denominator. The value of 0.01 is used for all the results presented in this paper.

Now five interscale PC maps at each orientation are fused together. The edge saliency map is obtained by maximizing the singular value-based moment of PC maps at six orientations:where denotes the orientation angle. is the edge saliency map. Figure 2 provides the flowchart of extracting the PC-based edge saliency map, and Figure 4(b) visualizes the final high-resolution edge saliency maps.

3.2. Intensity Saliency Map

The PC-based edge saliency map is sensitive to bleeding contours. On the other hand, the luminance intensities inside the contours are also prone to saliency. The difference in luminance is an indispensable condition for clinicians to distinguish between normal tissue and diseased tissue. Based on this principle, the gray-scale image histogram is exploited to compute an intensity saliency map. The histogram shown in Figure 3 is built by counting the gray-scale intensities within the bins. The counts of each bin are normalized to be a probabilistic value such thatwhere denotes the th bin. There are totally bins, and, in this paper, , which means that one gray-scale corresponds to one bin.where and λ are set to 1/72 and 0.6, empirically; denotes the pixel position in the gray-scale image . The function values constitute the intensity saliency map, which is visualized in Figure 4(c).

3.3. Color Saliency Map

Besides the above two features, the color feature which is the most widely used in the bleeding area detection cannot be ignored. The obvious color information of the bleeding area is the most intuitive reference for the clinicians, so the state-of-the-art methods utilize the color feature to distinguish the bleeding from other regions [5, 1517]. In perspective of the visual saliency, color is another visual stimulus that is isotropic to the contour and luminance. Hence color saliency map is also computed and fused together with the above two maps. As most bleeding lesions appear to be red, we calculate the proportion of the red of the color saliency map, where is a color vector at the pixel position of and is a small constant for the purpose of stabilization.

The aforementioned three saliency maps localize the bleeding ROIs in WCE images from different aspects. More importantly, the quantities in all the three saliency maps are between 0 and 1; thus we can derive the fused saliency map by fusing them together by the following formula, and the three saliency maps are pooled by

As an illustration, Figure 4 presents the three test WCE images in Figure 4(a). For each image, three isolated saliency maps and the contrastive fused saliency maps are arranged from Figures 4(b)4(h). Figure 4(i) localizes the suspected bleeding ROIs by thresholding the fused saliency map.

4. Experimental Results and Discussion

To evaluate our saliency-based bleeding detection system, we compared the saliency maps with the golden-standard masks that are marked by two physicians. As each saliency map implies the probability of bleeding positions, a receiver operator characteristic (ROC) curve can illustrate both sensitivity and specificity of a map. Both criteria are defined as follows:where TP denotes true positive rates and FP false positive rates; TN and FN denote true negative and false negative, respectively. TP means that a bleeding pixel is correctly classified by a saliency map while FP means that a nonbleeding pixel is incorrectly regarded as bleeding.

4.1. Quantitative Analysis

Given that a threshold increased from 0 to 1, a series of sensitivity and specificity values are obtained. Relying on these values, a ROC curve can be plotted. As an illustration, Figure 5 plots the ROC curves of fused saliency maps with regard to the six images in Figure 1.

From Figure 5, it can be observed that the fused saliency map shows unbalanced performance on the six representative bleeding images. Area under curves (AUC) is 0.994, 0.989, 0.986, 0.926, 0.895, and 0.869, respectively. Note that the fused saliency map tends to be more sensitive to tiny and sharp bleeding region (Figure 1(f)) than opaque one (Figure 1(a)). This may result from the fusion strategy which weights three isotropic maps equivalently.

To further evaluate the performance of the proposed saliency map, the three kinds of saliency maps, saliency maps of different combinations, and the peer algorithm [17] are listed in Table 1, where accuracy, sensitivity, and specificity values for different saliency maps are presented. The accuracy criterion is computed aswhere TP, TN, FP, and FN are obtained based on the threshold that is the mean value.


Accuracy (%)Sensitivity (%)Specificity (%)

Edge0.63720.98780.6317
Intensity0.96840.86720.9736
Color0.94960.71430.9615
Edge + intensity0.97690.93290.9778
Edge + color0.96350.86450.9774
Intensity + color0.97790.84550.9919
[17]0.93710.89580.9371
Fused0.98970.94070.9915

It can be observed that edge and intensity saliency maps are helpful to enhance bleeding detection accuracy. The edge saliency map has high degree of sensitivity to improve the sensitivity of the fused saliency map and the higher specificity of the fused saliency map benefits from the intensity saliency map; it is clear that the proposed saliency map is better than the saliency map in [17], which uses single color features. In order to discuss the robustness of the proposed saliency map, the histogram is drawn in Figure 6. From Figure 6(a), it can be seen that the sensitivity of edge saliency map is better than the other saliency maps; the specificity in Figure 6(b) illustrates that effect of the fused saliency map is very ideal and the values are very close to 1, while the main contribution of specificity of the fused saliency map is from the intensity saliency map; Figure 6(c) shows that the accuracy of the fused saliency map is the best, which meets the expectations of our design.

In consideration of the impact of the color space [1820], the fused saliency map is applied to the different color spaces in Table 2. From Table 2, the fused saliency map in the RGB color space has better performance than the HSI and the HSV.


RGBHSIHSV

Sensitivity (%)0.94070.94200.9393
Specificity (%)0.99150.97550.9804
Accuracy (%)0.98970.96950.9744

In addition, the proposed saliency map provides the probability of bleeding, which is suitable to be used as a postprocessing step in a superpixel detection flowchart [21]. Based on the saliency map, unimportant superpixels can be discarded.

4.2. Qualitative Analysis

In order to intuitively see the effect of the saliency map, we delineate the ROI of bleeding point. As shown in Figure 7, there is the binary image next to each saliency map. The saliency map is binarized according to single thresholding. The threshold of a saliency map is determined by the mean value of all quantities in the saliency map.

It can be seen that color binary mask is oversized while the shape binary mask is undersized. The edge binary mask tends to reveal all potential bleeding regions. Instead of applying logic AND or OR operations to fusing the three isolated binary masks, the proposed fusion strategy performs probabilistic calculation on isolated saliency maps, followed by binarizing the fused saliency map, which increases the fidelity of TP regions. It is obvious to see in Figure 7 that the fused binary map reveals exactly where the bleeding point is.

4.3. Image Noise Influence

WCE images are sometimes vulnerable to the noise contamination such as Gaussian noise or Salt and Pepper noise. To evaluate the performance of the proposed saliency map, we add three levels of Gaussian noise and Salt and Pepper noise in the WCE image in Figure 8.

From Figure 9 and Table 3, we can find that the accuracy of the proposed saliency map decreases as the level of noise increases, but the accuracy is still in expected range.


LevelSalt and Pepper (%)Gaussian (%)

198.9499
298.5198.44
398.1698.15

4.4. Run-Time Evaluation

The computation time of the fused saliency map and the contrast algorithm [17] is listed in Table 4. The algorithms are run on the MATLAB R2015b, with Intel Core i5 CPU at 2.4 GHz. Compared with [17], the proposed algorithm reduces the algorithm complexity of color saliency map and extracts three saliency maps in parallel. Table 4 shows that both algorithms are very fast and the proposed one has a minor advantage.


The fused saliency map[17]

Time/per (s)0.07480.0855

5. Conclusion

In this paper, a novel saliency map is proposed for bleeding localization on the WCE diagnosis. Unlike existing methods that use color features as the dominant criterion, we combine edge, intensity, and color information in visual saliency scheme. The results demonstrated that both the edge saliency map based on phase congruency and the intensity map based on luminance histogram dominate the saliency detection performance. Fused saliency map can detect the bleeding in WCE images with the average accuracy of 98.97%. Future work involves segmenting bleeding region with precise contours at a high specificity.

Conflicts of Interest

The authors declare that there are no conflicts of interest regarding the publication of this paper.

Acknowledgments

This work was financially supported by the National Natural Science Foundation of China (Grant no. J20141237). The authors would like to thank Hao Zhang, the project manager of Ankon Incorporation, Wuhan, China.

References

  1. “The national digestive diseases information clearinghouse website,” 2014, http://digestive.niddk.nih.gov/ddiseases/pubs/bleeding. View at: Google Scholar
  2. G. Iddan, G. Meron, A. Glukhovsky, and P. Swain, “Wireless capsule endoscopy,” Nature, vol. 405, no. 6785, pp. 417-418, 2000. View at: Google Scholar
  3. T. Ghosh, S. A. Fattah, and K. A. Wahid, “Automatic bleeding detection in wireless capsule endoscopy based on RGB pixel intensity ratio,” in Proceedings of the 1st International Conference on Electrical Engineering and Information and Communication Technology, ICEEICT, pp. 1–4, April 2014. View at: Publisher Site | Google Scholar
  4. L. Cui, C. Hu, Y. Zou, and M. Q.-H. Meng, “Bleeding detection in wireless capsule endoscopy images by support vector classifier,” in Proceedings of the 2010 IEEE International Conference on Information and Automation, ICIA, pp. 1746–1751, June 2010. View at: Publisher Site | Google Scholar
  5. T. Ghosh, S. K. Bashar, M. S. Alam, K. Wahid, and S. A. Fattah, “A statistical feature based novel method to detect bleeding in wireless capsule endoscopy images,” in Proceedings of the 2014 International Conference on Informatics, Electronics and Vision, ICIEV, pp. 1–4, May 2014. View at: Publisher Site | Google Scholar
  6. V. S. Kodogiannis, M. Boulougoura, E. Wadge, and J. N. Lygouras, “The usage of soft-computing methodologies in interpreting capsule endoscopy,” Engineering Applications of Artificial Intelligence, vol. 20, no. 4, pp. 539–553, 2007. View at: Publisher Site | Google Scholar
  7. Y. Fu, W. Zhang, M. Mandal, and M. Q.-H. Meng, “Computer-aided bleeding detection in WCE video,” IEEE Journal of Biomedical and Health Informatics, vol. 18, no. 2, pp. 636–642, 2014. View at: Publisher Site | Google Scholar
  8. R. Achanta and S. Süsstrunk, “Saliency detection using maximum symmetric surround,” in Proceedings of the 17th IEEE International Conference on Image Processing (ICIP '10), pp. 2653–2656, September 2010. View at: Publisher Site | Google Scholar
  9. Z. Feng, Y. Xu, and X. Yang, “Quaternion Phase Congruency model for edge saliency map extraction and image blur measurement,” in Proceedings of the 2014 IEEE International Symposium on Broadband Multimedia Systems and Broadcasting, BMSB 2014, pp. 1–6, June 2014. View at: Publisher Site | Google Scholar
  10. S. Goferman, Z. M. Lihi, and A. Tal, “Context-aware saliency detection,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 34, no. 10, pp. 1915–1926, 2012. View at: Publisher Site | Google Scholar
  11. L. Zhang, Z. Gu, and H. Li, “SDSP: A novel saliency detection method by combining simple priors,” in Proceedings of the 2013 20th IEEE International Conference on Image Processing, ICIP, pp. 171–175, September 2013. View at: Publisher Site | Google Scholar
  12. R. Achanta, F. Estrada, P. Wils, and S. Susstrunk, “Salient region detection and segmentation,” in Computer Vision Systems, vol. 5008 of Lecture Notes in Computer Science, pp. 66–75, Springer, Berlin, Germany, 2008. View at: Publisher Site | Google Scholar
  13. P. Kovesi, “Image features from phase congruency,” Journal of Computer Vision Research, vol. 1, pp. 1–26, 1999. View at: Google Scholar
  14. Y. Xu, D. Liu, Y. Quan, and P. Le Callet, “Fractal analysis for reduced reference image quality assessment,” IEEE Transactions on Image Processing, vol. 24, no. 7, pp. 2098–2109, 2015. View at: Publisher Site | Google Scholar | MathSciNet
  15. T. Ghosh, S. A. Fattah, C. Shahnaz, and K. A. Wahid, “An automatic bleeding detection scheme in wireless capsule endoscopy based on histogram of an RGB-indexed image,” Annual International Conference of the IEEE Engineering in Medicine and Biology Society, vol. 2014, pp. 4683–4686, 2014. View at: Publisher Site | Google Scholar
  16. A. K. Kundu, M. N. Rizve, T. Ghosh, S. A. Fattah, and C. Shahnaz, “A histogram based scheme in YIQ domain for automatic bleeding image detection from wireless capsule endoscopy,” in Proceedings of the IEEE International WIE Conference on Electrical and Computer Engineering, WIECON-ECE, pp. 455–458, Dhaka, Bangladesh, December 2015. View at: Publisher Site | Google Scholar
  17. Y. Yuan, B. Li, and M. Q.-H. Meng, “Bleeding frame and region detection in the wireless capsule endoscopy video,” IEEE Journal of Biomedical and Health Informatics, vol. 20, no. 2, pp. 624–630, 2016. View at: Publisher Site | Google Scholar
  18. J. A. W. M. Van Der Laak, J. R. Westphal, L. J. M. Schalkwijk et al., “An improved procedure to quantify tumour vascularity using true colour image analysis. Comparison with the manual hot-spot procedure in a human melanoma xenograft model,” The Journal of Pathology, vol. 184, no. 2, pp. 136–143, 1998. View at: Publisher Site | Google Scholar
  19. C. C. Reyes-Aldasoro, L. J. Williams, S. Akerman, C. Kanthou, and G. M. Tozer, “An automatic algorithm for the segmentation and morphological analysis of microvessels in immunostained histological tumour sections,” Journal of Microscopy, vol. 242, no. 3, pp. 262–278, 2011. View at: Publisher Site | Google Scholar
  20. O. A. Maximova, R. E. Taffs, K. L. Pomeroy, P. Piccardo, and D. M. Asher, “Computerized morphometric analysis of pathological prion protein deposition in scrapie-infected hamster brain,” Journal of Histochemistry & Cytochemistry, vol. 54, no. 1, pp. 97–107, 2006. View at: Publisher Site | Google Scholar
  21. Y. Yuan, J. Wang, B. Li, and M. Q.-H. Meng, “Saliency based ulcer detection for wireless capsule endoscopy diagnosis,” IEEE Transactions on Medical Imaging, vol. 34, no. 10, pp. 2046–2057, 2015. View at: Publisher Site | Google Scholar

Copyright © 2017 Hongda Chen et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

821 Views | 266 Downloads | 2 Citations
 PDF  Download Citation  Citation
 Download other formatsMore
 Order printed copiesOrder

We are committed to sharing findings related to COVID-19 as quickly and safely as possible. Any author submitting a COVID-19 paper should notify us at help@hindawi.com to ensure their research is fast-tracked and made available on a preprint server as soon as possible. We will be providing unlimited waivers of publication charges for accepted articles related to COVID-19.