Table of Contents Author Guidelines Submit a Manuscript
BioMed Research International
Volume 2017 (2017), Article ID 2028946, 9 pages
https://doi.org/10.1155/2017/2028946
Research Article

Robust Retinal Blood Vessel Segmentation Based on Reinforcement Local Descriptions

Qianfoshan Hospital of Shandong Province, Jinan 250014, China

Correspondence should be addressed to Guang Zhang; moc.621@5102papgz

Received 8 October 2016; Revised 16 December 2016; Accepted 25 December 2016; Published 18 January 2017

Academic Editor: Jiang Du

Copyright © 2017 Meng Li et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

Retinal blood vessels segmentation plays an important role for retinal image analysis. In this paper, we propose robust retinal blood vessel segmentation method based on reinforcement local descriptions. A novel line set based feature is firstly developed to capture local shape information of vessels by employing the length prior of vessels, which is robust to intensity variety. After that, local intensity feature is calculated for each pixel, and then morphological gradient feature is extracted for enhancing the local edge of smaller vessel. At last, line set based feature, local intensity feature, and morphological gradient feature are combined to obtain the reinforcement local descriptions. Compared with existing local descriptions, proposed reinforcement local description contains more local information of local shape, intensity, and edge of vessels, which is more robust. After feature extraction, SVM is trained for blood vessel segmentation. In addition, we also develop a postprocessing method based on morphological reconstruction to connect some discontinuous vessels and further obtain more accurate segmentation result. Experimental results on two public databases (DRIVE and STARE) demonstrate that proposed reinforcement local descriptions outperform the state-of-the-art method.