Table of Contents Author Guidelines Submit a Manuscript
Journal of Healthcare Engineering
Volume 2017, Article ID 5645498, 16 pages
https://doi.org/10.1155/2017/5645498
Research Article

An Approach to Evaluate Blurriness in Retinal Images with Vitreous Opacity for Cataract Diagnosis

1School of Information and Electronics, Beijing Institute of Technology, Beijing 100081, China
2Beijing Institute of Ophthalmology, Beijing Tongren Hospital, Beijing 100730, China

Correspondence should be addressed to Huiqi Li; nc.ude.tib@iliqiuh

Received 24 October 2016; Revised 22 January 2017; Accepted 13 February 2017; Published 26 April 2017

Academic Editor: Fabrice Meriaudeau

Copyright © 2017 Li Xiong et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

Cataract is one of the leading causes of blindness in the world’s population. A method to evaluate blurriness for cataract diagnosis in retinal images with vitreous opacity is proposed in this paper. Three types of features are extracted, which include pixel number of visible structures, mean contrast between vessels and background, and local standard deviation. To avoid the wrong detection of vitreous opacity as retinal structures, a morphological method is proposed to detect and remove such lesions from retinal visible structure segmentation. Based on the extracted features, a decision tree is trained to classify retinal images into five grades of blurriness. The proposed approach was tested using 1355 clinical retinal images, and the accuracies of two-class classification and five-grade grading compared with that of manual grading are 92.8% and 81.1%, respectively. The kappa value between automatic grading and manual grading is 0.74 in five-grade grading, in which both variance and value are less than 0.001. Experimental results show that the grading difference between automatic grading and manual grading is all within 1 grade, which is much improvement compared with that of other available methods. The proposed grading method provides a universal measure of cataract severity and can facilitate the decision of cataract surgery.

1. Introduction

A cataract occurs when there is a buildup of protein in the lens. The protein buildup makes lens cloudy; thus, some of the light is prevented from passing through the lens and vision is impaired [1]. This is mainly why cataract is the principle cause of partial to complete blindness. It is reported that cataract patients account for almost half of global blindness [2]. Research also suggests that the number of people worldwide who have lost their vision is increasing by approximately 1 million per year and this number is expected to increase up to 40 million by the year 2025 [3]. Although surgery is an effective treatment available for these patients. An accurate diagnosis of the severity of cataract is still needed prior to any surgical interventions to ensure both safety of the patients and high-quality treatment.

Cataracts are classified into three types according to the location of opacity: nuclear cataract, cortical cataract, and posterior subcapsular cataract. Clinically, nuclear cataract is diagnosed using slit-lamp photography, while cortical and posterior subcapsular cataracts are diagnosed using retroillumination imaging devices. Clinical grading of cataract is performed via comparing the observed picture with a set of standard photos with different cataract severities. Lens Opacities Classification System III (LOCS III) [4] is widely applied in clinical diagnosis. Besides, the Oxford Clinical Cataract Classification [5] and the American Cooperative Cataract Research Group (CCRG) method [6] are similar grading criteria. The criteria mentioned above require different image modalities to diagnose cataract which can be graded separately for each individual type. A universal evaluation of cataract severity is essential to determine the timing of cataract surgery and the priority of surgery when surgery service is not sufficient.

Since the first fundus camera was invented in 1910 [7], the retinal image acquired by the fundus camera has been widely applied in diagnosis of ocular diseases such as glaucoma, age-related macular degeneration, and diabetic retinopathy [8]. Recently, retinal image is proposed as a new means for cataract diagnosis according to its blurriness [9]. Compared with that of slit-lamp images and retroillumination images, only one image modality is required for cataract diagnosis using retinal images, and a universal grading of cataract severity can be assessed. Two distinguished benefits are found in this method which include compatibility for telediagnosis and integration with screening of glaucoma and diabetic retinopathy. Therefore, we believe screening for cataracts using retinal images can be an efficient means not only for addressing healthcare shortage problem in developing countries via telemedicine but also for integrated screening of different retinal diseases.

Opacity of lens will lead to blur photography of the retina. According to blurriness shown in a retinal image, blur level of retinal images can be classified into five grades [10]. Figure 1 shows examples of retinal images with five different grades. Figure 1(a) shows a healthy retinal image with grade 0. In an image with grade 0, the retinal vessels (both the main vessels and capillary vessels) and optic disc can be clearly seen. In an image with grade 1, main vessels are visible while the capillary vessels are only faintly visible as shown in Figure 1(b). Figure 1(c) presents an image with grade 2 in which small vessels are invisible and main vessels are faintly visible. In an image with grade 3, only the optic disc and vessels around the optic disc can be seen indistinctly, which is illustrated in Figure 1(d). As Figure 1(e) shows, no retinal structures can be observed in an image with grade 4. Ophthalmologists define grade 0 as noncataract and grades 1–4 as cataract cases, while no surgery is needed for grades 0-1 and timely surgery is necessary for grades 2–4 [11].

Figure 1: Examples of retinal images with different blurriness grades: (a) grade 0; (b) grade 1; (c) grade 2; (d) grade 3; (e) grade 4.

Screening for cataract using retinal images according to blurriness will be influenced by two factors: vitreous opacity and the small pupils of patients. Vitreous opacity is a common abnormality which can be observed in a retinal image. Normal vitreous body is a transparent and extracellular gel, which contains a complicated structural framework of collagen, soluble proteins, and hyaluronic acid and a water content of 99% [12]. There are many types of vitreous opacity with different causes, and they usually appear as a circular-shaped, translucent bright area in a retinal image. Figure 2 shows some examples of a retinal image with vitreous opacity. Figures 2(a), 2(b), and 2(c) are diagnosed as cataract in grades 2, 3, and 4, respectively. It can be seen that vitreous opacity may emerge in various shapes, sizes, intensities, and positions and they are obvious even in blurry images, which will influence the detection of retinal structures and evaluation of blurriness. The small pupil of patients will also cause blur of retinal images, but it mostly affects the peripheral regions of the retinal images. Generally, the retinal images of patients will be photographed with the mode of the small pupil.

Figure 2: Examples of images with vitreous opacity: (a) image with blurriness of grade 2; (b) image with blurriness of grade 3; (c) image with blurriness of grade 4.

In developing countries, due to the regional economic gap, well-experienced ophthalmologists are not sufficient in some areas [13]. This fact impedes timely treatment for patients. At present, screening of cataract and other retinal diseases in these developing areas with insufficient ophthalmologists can be performed through remote consultation. Computer-aided analysis and a screening system can realize feature-based visualization and provide references for ophthalmologists, which are applicable to remote consultation. Moreover, there is difference of diagnosis between different ophthalmologists. An automatic screening system can quantify features, which is beneficial to unify diagnostic criteria.

Automatic analysis of retinal image has been investigated for many years. Most research focuses on automatic detection of retinal structures such as the optic disc [1417], vessels [1820], and fovea [2123]. Based on anatomical structure detection, features of each fundus structure are extracted to detect pathological changes. Extensive studies have been performed on computer-aided diagnosis systems for ocular diseases such as microaneurysms [24], glaucoma [2527], macular degeneration [28, 29], and diabetic retinopathy [30, 31].

Research efforts have been put into diagnosis of a specific type of cataract automatically using slit-lamp images or retroillumination images [3234]. With retinal image becoming more and more widely used in clinical applications, automatic grading of cataract based on retinal image is proposed in recent years. Fourier analysis was proposed to estimate cataract severity in digital retinal images [35]. It showed that Fourier value has a close correlation with the combined LOCS III score and logMAR visual acuity. In this study, the detection between cataract and noncataract was evaluated, but automatic grading of cataract severity was not performed. Top-bottom-hat transformation and a trilateral filter were applied in paper [36] to extract features which include luminance feature, gray co-occurrence matrix, and gray-gradient co-occurrence matrix. Based on these features, BP neural network classifier was utilized to grade cataract severity. The wavelet transform and sketch-based methods were investigated for feature extraction in [13]. Retinal images were classified into noncataract, mild, moderate, and severe grade with multiclass discriminant analysis algorithm. The method in [37] combined methods proposed in [13, 36]. The ensemble methods, majority voting, and stacking are investigated to combine the multiple base learning models for final cataract classification.

Although cataract classification using retinal image has been investigated in the previous studies, there are still two challenges. One issue is that existing methods cannot handle the retinal images with vitreous opacity well. All available methods of screening cataract using retinal images are only based on the features of contrast between retinal structures and background [13, 3537] without considering the influence of vitreous opacity. In these methods, vitreous opacities with strong contrast will be detected as retinal structures, which will lead to incorrect grading. Another challenge is that the performance of automatic grading needs to be improved. For example, there is still great grading difference (>1 grade) between automatic grading and clinical grading using available methods, which will hinder the clinical usage of automatic grading.

In this paper, an approach to evaluate blurriness for cataract diagnosis with vitreous opacity removal is proposed. The main contribution of this work can be summarized as follows: (1) A morphological method with multithresholds is proposed to detect vitreous opacity. The detected vitreous opacity is removed from retinal structure detection to improve the accuracy. This method is also applicable to eliminate exudates from visible retinal structure segmentation. (2) The designed approach is applicable to images with different resolutions. (3) Our grading difference between automatic grading and clinical grading is all within 1 grade, which shows that the proposed method outperforms the other existing methods. (4) Our approach is evaluated using a large amount of retinal images (>1000) with different resolutions.

2. Method

The proposed automatic approach for cataract grading includes the following steps: preprocessing for image-size normalization, image enhancement, feature extraction, and cataract grading. The entire process is illustrated in Figure 3. In order to process retinal images with different resolutions, image-size normalization based on the diameter of the acquired region of interest (ROI) is performed in preprocessing. Images are further enhanced with 2D reverse Gaussian-matched filter after preprocessing, and three types of features are extracted. A decision tree is trained to classify every two adjacent grades into one class before the final grading. Pathological changes may appear in a retinal image, and these pathological changes will affect accurate feature extraction. Vitreous opacity is one type of pathological change, which is commonly presented in retinal images of cataract patients. To avoid its disturbance, a morphological method with multithresholds is proposed to detect vitreous opacity in this paper and vitreous opacity is further removed from the retinal structure segmentation.

Figure 3: Flow diagram of the proposed cataract grading system.
2.1. Preprocessing

In order to handle the images with different resolutions, every image will be adjusted to the same size automatically. It should be noted that the diameter of an ROI (region of interest) mask is normalized rather than the size of an input image. To obtain this diameter, an ROI mask is obtained. Firstly, a retinal image is set at a threshold in the red channel. This method is essentially employed to separate the dark background and the retina. To avoid misclassifying areas with low illumination in the retina as background, the threshold value is set to 10, which is an empirical value. Then, morphological operators of opening, closing, and erosion are applied on the image obtained in the previous step in turn with a 3 × 3 square kernel [16]. Figure 4(b) shows an example of the generated ROI mask of Figure 4(a). When processing a new retinal image, the ROI mask is extracted and the maximal width in the horizontal direction is computed as the diameter of the fundus as shown in Figure 4(b). If the diameter is not equal to a given diameter, then the input image will be resized according to the ratio between the estimated diameter and the given diameter by bilinear interpolation. In our test, the given diameter is set to 2161 pixels. To reduce the detection error near the boundary, the ROI mask will be shrunk by 50 pixels. Next, all the parameters are set according to the size of retinal structures in normalized images. In the normalized images, the average diameter of the optic disc is about 200 pixels, and the width of the vessels are in the range of 24∼36 pixels.

Figure 4: ROI mask generation: (a) original image; (b) ROI mask.
2.2. Image Enhancement with the 2D Gaussian-Matched Filter

In order to enhance the retinal vessels and boundary of the optic disc, the 2D Gaussian-matched filter [18] is applied on green channel followed by initial processing. It is observed that retinal blood vessels usually appear darker than their surroundings and can be approximated as piecewise linear segments; we assume that the model of a vessel segment is a reverse Gaussian function. So, a Gaussian filter is employed as the matched filter. Mathematically, a Gaussian filter can be described as follows: where is the length of a vessel segment on a fixed orientation. The orientation of is initialized along y-axis. represents the scale of this filter, which is related to the width of the vessel. The mean value of the filter is determined as (2), and is the number of points in this filter.

The final convolutional filter is defined as follows:

In order to detect the vessel in different directions, 12 filters on 12 different orientations are generated using the rotation matrix mentioned as follows: where is the orientation of the ith filter which increases 15° from ith filter. Twelve filters are applied on each pixel in the green channel. For every pixel, only the maximum response among the twelve filters is retained.

Filter size and parameter and are all determined according to the vessel width of an input image. It has been proved in [18] that it is proper to set and to 2 and 9 pixels, respectively, in their database, and the filter size is 15 × 16 pixels. In our paper, the size of our normalized image is 4 times of the size of images in [18]. Therefore, the filter is designed with size 60 × 61, and and are set to 8 and 36, respectively. An example of the enhanced image is shown in Figure 5(b). It can be seen that the retinal vessels and boundary of the optic disc are enhanced in the retinal image.

Figure 5: The enhancement with 2D Gaussian filter convolution: (a) original image; (b) enhanced image.
2.3. Detection of Vitreous Opacity

Vitreous opacity often appears in the retinal image, which will interfere the grading of blurriness for cataract diagnosing. Due to the higher intensity compared with background, boundaries of vitreous opacity will be enhanced by the 2D Gaussian-matched filter as well. When a large area of vitreous opacity appears, the accuracy of grading will be seriously influenced because the boundaries of vitreous opacity are wrongly extracted as visible retinal structures.

To eliminate the detection error brought by vitreous opacity, a set of morphology operations are applied on the green channel of input images to enhance regions of vitreous opacity. To avoid the noises, the green channel is smoothed by a median filter at first. The size of this median filter should be less than the width of the vessels. In our paper, the size is set to 12 × 12 pixels. Close operations are performed with circle structuring elements and which are defined by the following: where represents the intensity in green channel after median filtering.

Next, is dilated with circle structuring element as expressed in

A subtraction operation is performed between and to enhance the boundary of bright lesions:

Equations (5) and (6) aim to erode dark vessels, so the selection of the size of circle structuring elements is based on the width of the main retinal vessels. In our experiments, the main vessels’ width of normalized image is in the range of 24∼36 pixels. Accordingly, the radius of and is set to 24 and 36, respectively. The radius of decides the width of the detected boundaries of the candidate vitreous opacity, and the width should be slightly wider than the average width of the vessels. Through experiments, it is suitable to set to 28. Figures 6(a), 6(b), and 6(c) show the enhancement of vitreous opacity, which correspond to the original images in Figures 2(a), 2(b), and 2(c), respectively.

Figure 6: The enhancement of vitreous opacity. (a), (b), and (c) correspond to the original images in Figures 2(a), 2(b), and 2(c), respectively.

By thresholding in image , boundaries of bright regions will be detected as candidate vitreous opacity, which contain the boundaries of vitreous opacity, the optic disc, and few vessels across the optic disc. To judge whether there is vitreous opacity in an input image, thresholding is performed on image according to (9). The number of segmented pixels and number of connected regions are computed. Through testing, when is set to 8, vitreous opacity in most images can be detected well in our database. If the number of segmented pixels is much larger than that of the optic disc perimeter or if the number of connected domains is more than 5, it is considered that there is possible vitreous opacity in this image.

As shown in Figure 6(a), when the optic disc is obvious in retinal images, the boundary of the optic disc and vessels around the optic disc will be enhanced in image as well. Therefore, the region around the optic disc should be excluded from vitreous opacity detection. A method of optic disc removal is proposed and applied only on the images with possible vitreous opacity.

Because the optic disc is usually the largest bright region in a retinal image, the 2% brightest pixels in the ROI are selected as candidate pixels of the optic disc. If is defined as the selected bright pixels, a clustering algorithm is applied on to cluster these pixels into groups based on Euclidean distance. The candidate optic disc is searched among these groups according to the size of the optic disc.

Features in image are taken into consideration to search the optic disc. For , the number of pixels in is summed up in a window centered on the center of as follows:

If is larger than , then this group is considered as the optic disc. Then, detected candidate vitreous opacities in the square with the size centered on the center of the optic disc are excluded. If there is no region satisfying the above criteria, it is considered that there is no visible optic disc in the image. The whole process is expressed as in Algorithm 1.

Algorithm 1

The parameters , , and are all determined by the size of the optic disc. The optic disc is roughly an ellipse and its long diameter is about 200 pixels in our experiments. The bright pixels with the distance less than are all gathered into one region, and is set to be the diameter of the optic disc. is utilized to judge whether the clustered group is a candidate of the optic disc. Considering that not all pixels in the optic disc can be selected as the brightest pixels, is set to . is the area of detected edges in window . should cover the optic disc, which is set to . is employed to detect whether the clustered region contains the edges of the optic disc. should be at least larger than the area of optic disc boundaries. Because the width of disc edges is close to the vessel width, is set to .

2.4. Feature Extraction

It is important to extract appropriate features from a retinal image for classifying cataract. Figure 1 illustrates that retinal structures that are clearer and can be seen in images with lower grades. According to the grading protocol mentioned in Section 1, the contrast between retinal structures and background is the prime feature for grading. Considering the above, three types of features which contain nine features are selected in this paper. These nine features are described in Table 1.

Table 1: Feature description of cataract grading.
2.4.1. Visible Retinal Structure Segmentation Using Multiple Thresholds

Vessels and boundary of the optic disc are enhanced with the Gaussian-matched filter as described in Section 2.2. More obvious structure has higher response. Retinal structure pixels are segmented by thresholding the enhanced image with a given threshold value , and a visible structure map can be generated as where is the position of every pixel in the image and is a binary image in which visible structure pixels are expressed as 1. Figure 7 shows some examples of visible structure results with different grades of cataract by setting to 0.9. The images in Figure 7 correspond to the original images in Figures 1(a), 1(b), 1(c), 1(d), and 1(e), respectively. It is evident that there is the largest number of visible structure pixels in a healthy image, and the number of segmented pixels can describe the severity of cataract. The number of visible structure pixels is computed as follows:

Figure 7: Visible structure segmentation. (a), (b), (c), (d), and (e) correspond to the original images in Figures 1(a), 1(b), 1(c), 1(d), and 1(e), respectively.

High threshold will result in missing vessels in the retinal images with serious blur, but on the other hand, low threshold will produce noises in the retinal images with lower grades. Visible structure segmentation with only one threshold is not enough to classify retinal images into 5 different grades accurately. Therefore, four increasing threshold values t1, t2, t3, and t4 are applied using (11) separately, and each threshold is only employed to distinguish two adjacent grades. t1 is set to classify grades 3 and 4, t2 is for the classification of grade 2 and grade 3, and so on. The value t1 should not be too low to avoid over segmenting nonvessel pixels in grade 3 and grade 4. The value t4 should not be set too high to detect small vessels in the retinal images with grade 0. With the blurriness aggravating, the difference between two adjacent grades is less. Accordingly, difference between the two adjacent thresholds is increasing. Through experiments of sample images from five grades, t1, t2, t3, and t4 are set to 0.3, 0.5, 0.9, and 1.4, respectively, in this paper. Four visible retinal structure segmentation V1, V2, V3, and V4 are obtained, which correspond to the four thresholds t1, t2, t3, and t4. Then, the number of segmented pixels in V1, V2, V3, and V4 is summed and denoted as v1, v2, v3, and v4, respectively, via (12), which is described in Table 1.

As mentioned in Section 2.3, responses of the 2D Gaussian filter for the pixels around vitreous opacity are strong, which will be detected as visible retinal structure pixels (see Figures 9(a-1), 9(b-1), and 9(c-1)). If an input image is diagnosed as vitreous opacity, boundaries of vitreous opacity should be eliminated from segmentation results V1, V2, V3, and V4.

It is hard to eliminate vitreous opacity directly from visible structure segmentation due to the large variation among the images with different blur levels. Input images are classified into two classes: class 1 containing images with visible retinal structures (grades 0, 1, and 2) and class 2 including images with invisible retinal structures (grades 3 and 4). Vitreous opacity is removed from visible structure segmentation using different algorithms for these two classes.

For this rough classification, a retinal image is divided into two annular regions: region 1 and region 2 (refer to Figure 8). Average enhanced values of and using the Gaussian filter in these two regions are computed as follows: where and represent region 1 and region 2, respectively. contains no retinal structures and contains the main structures. And n1 and n2 are the number of pixels in and correspondingly. is the enhanced image with Gaussian filter correlation.

Figure 8: The partition of a retinal image.

Average values in two regions are added up as follows:

According to the value E, an input image is categorized into two classes by thresholding: where is a value that is trained using bipartition [38] in the sample set.

To eliminate vitreous opacities from four different vessel maps, V1, V2, V3, and V4, boundaries of vitreous opacity are extracted by applying three thresholds on image via where . In this paper, are set to 2, 5, and 8, respectively, via experiments. The extracted results are expressed as binary images in turn and are further expanded using circle structuring element as follows:

The radius of is set to 5 pixels to ensure that the width of vitreous opacity boundary is wider than the vessel diameter.

It is observed that most vitreous opacities are obvious in the retinal images of class 1, but a lot of vitreous opacities are dim in retinal images of class 2. Due to higher contrast of the images in class 1, many nonvitreous opacity pixels are detected in . In the images of class 2, most vitreous opacities can be detected well in , but many dim vitreous opacities will be missed in . Therefore, different algorithms for class 1 and class 2 are proposed to remove vitreous opacities from V1, V2, V3, and V4. Due to the strong contrast between vitreous opacities, some images with large number of vitreous opacities of class 2 may be misclassified as class 1. So for images in class 1, if there are still many connected regions with small areas in V1 after removal of vitreous opacities, it is considered that this image is misclassified as class 1 and further elimination needs to be performed on V1 and V2. The whole process of eliminating vitreous opacity is as in Algorithm 2.

Algorithm 2

is a set of connected regions in V1, and N is the number regions in . Because single vitreous opacity is a region of a small area, is the set of connected regions with a small area that is less than half of the area of the optic disc and is the number of regions in .

Examples of vitreous opacity boundary removal from visible retinal structure segmentation are shown in Figure 9. Final visible retinal structure maps are illustrated in Figures 9(a-3), 9(b-3), and 9(c-3)), and Figures 9(a-2), 9(b-2), and 9(c-2)) is the extracted vitreous opacity boundaries. It can be noted from Figure 9 that the proposed method can detect vitreous opacity well in these images and the extraction of retinal feature is more accurate after vitreous opacity removal.

Figure 9: Examples of vitreous opacity removal from visible retinal structure segmentation. The first row is the segmented visible structures in three retinal images without removing vitreous opacity. The second row shows the extracted boundaries of vitreous opacity. The last row is the final visible structure detection after removing vitreous opacity boundaries. The images in columns 1, 2, and 3 correspond to the original image in Figures 2(a), 2(b), and 2(c), respectively.
2.4.2. Mean Gaussian Filter Response of Segmented Visible Retinal Structure Pixels

The value of a pixel in the enhanced image implies the contrast in the original image. Mean enhanced value of segmented pixels is described as follows:

In (19), is a group of the segmented vessel pixels with a specified threshold value. represents the value of the point (x, y) in the enhanced image. denotes the number of all segmented vessel pixels. Four mean enhanced values (m1, m2, m3, and m4 presented in Table 1) are computed with four different thresholds.

2.4.3. Local Standard Deviation

A clearer image owns a sharper contrast, which corresponds to greater intensity variance. Accordingly, local variance is computed on every pixel within the ROI mask. The neighborhood standard deviation is calculated within the neighborhoods around the corresponding pixels. Mean value of the local standard deviations (presented in Table 1) in the ROI mask is calculated as a feature. The value of can be set in a range of 0∼vessel width. Large value of will lead to long computing time. Experiments show that different values of will not affect the classification results too much. To save computing time, is set to 3 in this paper.

2.5. Grading of Blurriness

Because the distribution of features overlaps between neighboring grades, a decision tree is trained to classify a retinal image according to the blurriness. A sample set S selected from database is used to train grading parameters. The whole process is demonstrated in Figure 10.

Figure 10: The diagram of classification.

Firstly, feature v3 is selected to classify input images into three classes: a group of images with grades 0 and 1, a group of images with grade 1 and grade 2, and a group of images with grade 2, grade 3, or grade 4. It can be described as follows:

Here, the interval of feature v3 in the overlapped distribution between sample images in grade 1 and grade 2 is defined as an increasing order . is the average value of feature in the overlapped distribution between sample images in grade 1 and grade 2. is computed as follows:

Next, feature v2 is used to classify the retinal images into three classes similarly as feature v3. Finally, images are graded into five different grades according to different features as Figure 10 shows. As retinal structures are hard to be seen in the images with the cataract in grades 3 and 4, feature is not used for classification between grades 3 and 4.

For classifying two neighboring grades (grade i and grade i + 1), threshold of the employed feature is trained using bipartition proposed by Quinlan [38]. Information gain of the feature based on threshold is calculated as follows: where is the sample set, is the sample set in which is less than threshold , and is the sample set in which is larger than .

Ent is the information entropy, which can be expressed as follows: where is the proportion of class in the sample set and is the number of the employed features.

A voting scheme is applied for final grading, and the voting score is calculated as follows: and

If the score is larger than , the image is classified into grade . Otherwise, the grade is decided by , the feature with the maximum information gain.

3. Results and Discussion

3.1. Data Description

In our experiment, retinal images from the Beijing Institute of Ophthalmology, Tongren Hospital, are tested to evaluate the proposed grading algorithm. This database consists of images with different resolutions which include 1396 × 1124, 1572 × 1308, 1704 × 1496, 2036 × 1696, 2196 × 1740, 3048 × 2432, and 3380 × 2592 pixels. In total, 1355 retinal images of the right or left eyes from more than 1000 patients (all are Chinese) aged 20–80 in China were tested. In this database, most are 45-degree retinal images which are centered on the macula and only 15 retinal images are centered on the optic disc. These retinal images are graded by two experienced ophthalmologists according to the criterion [10] described in Section 1. Our automatic grading is compared with this clinical grading.

3.2. Experimental Results and Discussion

In total, 1355 retinal images are collected to evaluate our automatic blurriness grading system and the dataset includes 433, 415, 217, 133, and 157 images with blur level in grade 0, grade 1, grade 2, grade 3, and grade 4, respectively. The evaluation criteria and clinical diagnosis are presented in Table 2. The five-fold cross-validation method is used to evaluate the performance of our proposed method. The retinal images of our database are randomly divided into five subsets. In each fold, one subset is chosen as a testing set and the other four subsets are the training data.

Table 2: Evaluation criteria.

The results of all folds are counted. The joint accuracy for all folds of different types of clinical evaluation compared with that of manual grading are shown in Table 3, and confusion matrix of five gradings is presented in Table 4. Compared with that of the clinical grading, the joint accuracy of five-grade grading can reach 81.1%. Screening between noncataract and cataract using the proposed algorithms can achieve an accuracy of 92.8%; specific performance metrics of recognizing cataract are described in Table 5. In Table 5, TP, FN, FP, and TN represent true positives, false negatives, false positives, and true negatives, respectively. According to the grading criteria [13, 35], the images in grade 0 are considered noncataract, while the images in grades 1 and 2 correspond to mild and moderate cataract, respectively. The rest of the images in grades 3 and 4 are classified into severe cataract. The accuracy reached 83.8% when retinal images are classified into 4 grades (healthy, mild, moderate, and severe) [13, 35]. Blurriness in grade 1 will not cause severe vision loss, while blurriness in grades 2, 3, and 4 will lead to blindness which needs timely surgery [11]. Hence, it is important to recognize those images that need surgery. Ophthalmologists suggest that classification between grade 0 and grade 1 will not influence surgical screening. After neglecting the classification between grades 0 and 1, the accuracy compared with that of manual grading is 88.4%.

Table 3: Results of automatic grading using different types of clinical evaluation.
Table 4: Confusion matrix of grading.
Table 5: Performance of recognizing cataract.

By taking clinical grading as the gold standard, the agreement between the automatic cataract grading system and clinical grading in five-grade grading is measured by kappa statistic. Kappa coefficient is a statistic which measures inter-rater agreement for qualitative items [39]. It is applied extensively in clinical studies [40] and clinical skill assessments [41]. The formula for Cohen’s kappa coefficient is expressed as follows: where is the observed percentage of agreement which implies the proportion of ratings where the raters agree and is the expected percentage of agreement which denotes the proportion of agreements that are expected to occur by chance as a result of random rating.

The range of kappa is from 0∼1, with larger values indicating better agreement. If the kappa value is greater than 0.61, a study shows that such a classification algorithm is an applicable method for grading [42]. Table 6 shows that our kappa value is 0.7435 and variance and value are both less than 0.001. It indicates that there is no large deviation between automatic grading and clinical grading.

Table 6: Statistics of validation experiment in five-grade grading.

Vitreous opacity is a common disease which will influence feature extraction and grading of blurriness. As seen in Table 7, 23 retinal images are seriously influenced by this type of inaccurate feature extraction and their grading difference is more than two grades. A morphological method with multithresholds is proposed to extract and remove lesions such as vitreous opacity. Large grading difference is thus eliminated; 4.4% improvement made is achieved for five-grade grading of blurriness.

Table 7: Comparison results with vitreous opacity removal.

To compare with other algorithms of cataract classification using retinal images, we have implemented these algorithms and tested these algorithms using our database. The comparison results are summarized in Table 8. Due to small differences between adjacent grades, the Fourier score proposed in paper [35] describes that the high-frequency components of an input image cannot achieve a promising result in cataract screening or grading. Forty features are extracted in the method [36] to classify the cataract without considering other pathologies. Experimental results show that there are large grading differences using this method [36]. For example, maximal deviation by this method [36] is more than 3 grades. Compared with that of these methods, the proposed algorithm achieved more promising results and performed well in processing images with pathological changes.

Table 8: Performance comparison with other methods.

Analyzing the grading difference, we found that the proposed algorithm still has limitation in handling images with the following cases: uneven illumination, blurriness of the retina due to the small pupil, or large area of lesions. Figure 11 shows some examples of such cases. Figures 11(a) and 11(b) display the retinal images with uneven illumination. In Figure 11(a), the blur of the retina is due to patient’s small pupil, so ophthalmologists diagnosed this image as cataract in grade 1, while our automatic system considered it as cataract in grade 2. The uneven illumination of Figure 11(b) is caused by the operation of taking the picture. The small vessel can be seen in the bright region, but it is invisible in the dark area. This image is considered cataract in grade 1 in clinical grading while it is graded into grade 2 by the automatic system. Figures 11(c) and 11(d) show the examples of retinal blurriness caused by other lesions, so ophthalmologists tend to classify these images into lighter grades. As Figure 11(c) shows, the lesions cover the main retinal structures. Our algorithm classified Figure 11(c) as cataract in grade 3, but ophthalmologists think it is in grade 2. The blurriness of Figure 11(d) is caused by high myopia, so grade 2 is determined by ophthalmologists and grade 3 is obtained using our grading system.

Figure 11: Analysis of the grading difference.

Different retinal diseases may appear on a retinal image, such as macular edema, diabetic retinopathy, peripapillary atrophy, and vitreous opacity. In our tested database, most images only have cataract and parts of retinal images have other pathologies, such as mild diabetic retinopathy and vitreous opacity. Our proposed method performs well in grading cataract in these images. However, because a large area of severe pathologies will affect the detection of retinal structures, grading difference also exists among different ophthalmologists. Therefore, cataract screening of retinal images with severe pathologies is still a challenge for manual grading as well as automatic grading. Our future work will focus on handling the retinal images with complicated pathologies. Moreover, there are causes of blur other than cataract, such as media opacities in the anterior segment (e.g., cornea opacities) or posterior segment (e.g., vitreous haemorrhage), the small pupil, and process of photo taking. The small pupil of patients will cause blur of retinal images, but it mostly affects the peripheral regions of the retinal images. Generally, the retinal images of patients will be photographed with the mode of the small pupil. For blurriness caused by photo taking, these images are taken under strict protocol and images with bad quality will be taken again. At the current stage of research, our proposed method cannot recognize the causes of blurriness yet. The blur images caused by other media opacities may affect the grading of cataract. Identifying the cause of blurriness will be investigated in our future work. There are still some parameters that are currently set empirically, which may not be applicable to retinal images with different fields of view or other ethnicities. In the future work, automatic estimation of the parameters will be studied to improve the generalization of the proposed approach.

4. Conclusion

An approach to evaluate blurriness with vitreous opacity removal for cataract diagnosis using retinal images is proposed. Three types of features based on image enhancement using the 2D Gaussian filter are extracted to describe the blurriness in a retinal image. A decision tree is further proposed to classify the image into five grades. Compared with other research works in cataract grading using a retinal image, it is the first time that detection and removal of pathological changes are used to improve the grading accuracy. Using a large amount of clinical images as test data, our system can obtain 81.1% accuracy of exact cataract grading and 92.8% accuracy of two-class screening (noncataract and cataract). Kappa value is 0.7435 and variance and value are both less than 0.001 in the five-grade grading, which illustrates that there are no large differences between our grading and manual grading. The proposed approach provides an accurate and universal measure of cataract severity using a retinal image. The proposed system is currently in the stage of clinical validation, and it can be applied to mass screening and remote consultation in the future.

Conflicts of Interest

The authors declare that they have no financial and personal relationships with other people or organizations that can inappropriately influence their work; there is no professional or other personal interest of any nature or kind in any product, service, and/or company that could be construed as influencing the position presented in or the review of this manuscript.

References

  1. D. Allen and A. Vasavada, “Cataract and surgery for cataract,” British Medical Journal, vol. 333, no. 7559, pp. 128–132, 2006. View at Publisher · View at Google Scholar
  2. Organization, W. H, State of the World’s Sight: Vision 2020: The Right to Sight: 1999-2005, 2005.
  3. A. Foster, “Vision 2020: the cataract challenge,” Community Eye Health, vol. 13, no. 34, pp. 17–19, 2000. View at Google Scholar
  4. M. Karbassi, P. M. Khu, D. M. Singer, and C. L. Jr, “Evaluation of lens opacities classification system III applied at the slitlamp,” Optometry & Vision Science, vol. 70, no. 11, pp. 923–928, 1993. View at Google Scholar
  5. J. Sparrow, A. Bron, N. Brown, W. Ayliffe, and A. Hill, “The Oxford clinical cataract classification and grading system,” International Ophthalmology, vol. 9, no. 4, pp. 207–225, 1986. View at Google Scholar
  6. L. T. Chylack, M. Lee, W. Tung, and H. Cheng, “Classification of human senile cataractous changes by the American Cooperative Cataract Research Group (CCRG) method. I. Instrumentation and technique,” Investigative Ophthalmology & Visual Science, vol. 24, no. 4, pp. 424–431, 1983. View at Google Scholar
  7. A. Gullstrand, “Neue methoden der reflexlosen ophthalmoskopie,” Berichte Deutsche Ophthalmologische Gesellschaft, vol. 36, pp. 75–80, 1910. View at Google Scholar
  8. M. D. Abràmoff, M. K. Garvin, and M. Sonka, “Retinal imaging and image analysis,” IEEE Reviews in Biomedical Engineering, vol. 3, pp. 169–208, 2010. View at Publisher · View at Google Scholar · View at Scopus
  9. C. H. Parikh, S. Fowler, and R. Davis, “Cataract screening using telemedicine and digital fundus photography,” Investigative Ophthalmology & Visual Science, vol. 46, no. 13, pp. 1944–1944, 2005. View at Publisher · View at Google Scholar · View at Scopus
  10. C. W. Yang, X. U. Liang, S. Wang, and H. Yang, “The evaluation of screening for cataract needed surgery with digital nonmydriatic fundus camera,” Ophthalmology in China, vol. 19, no. 1, pp. 46–49, 2010. View at Publisher · View at Google Scholar · View at Scopus
  11. X. U. Liang, Q. F. Liang, and W. Shuang, “Cataract screening as a key step to vision restoring project in country,” Ophthalmology in China, vol. 19, no. 1, pp. 1–3, 2010. View at Google Scholar
  12. S. Coupland, “The pathologist’s perspective on vitreous opacities,” Eye, vol. 22, no. 10, pp. 1318–1329, 2008. View at Publisher · View at Google Scholar · View at Scopus
  13. L. Guo, J. J. Yang, L. Peng, J. Li, and Q. Liang, “A computer-aided healthcare system for cataract classification and grading based on fundus image analysis,” Computers in Industry, vol. 69, no. C, pp. 72–80, 2015. View at Publisher · View at Google Scholar · View at Scopus
  14. M. Foracchia, E. Grisan, and A. Ruggeri, “Detection of optic disc in retinal images by means of a geometrical model of vessel structure,” IEEE Transactions on Medical Imaging, vol. 23, no. 10, pp. 1189–1195, 2004. View at Publisher · View at Google Scholar · View at Scopus
  15. A. Hoover and M. Goldbaum, “Locating the optic nerve in a retinal image using the fuzzy convergence of the blood vessels,” IEEE Transactions on Medical Imaging, vol. 22, no. 8, pp. 951–958, 2003. View at Publisher · View at Google Scholar · View at Scopus
  16. A. R. Youssif, A. Z. Ghalwash, and A. R. Ghoneim, “Optic disc detection from normalized digital fundus images by means of a vessels’ direction matched filter,” IEEE Transactions on Medical Imaging, vol. 27, no. 1, pp. 11–18, 2008. View at Publisher · View at Google Scholar · View at Scopus
  17. C. Pereira, L. Gonçalves, and M. Ferreira, “Optic disc detection in color fundus images using ant colony optimization,” Medical & Biological Engineering, vol. 51, no. 3, pp. 295–303, 2012. View at Publisher · View at Google Scholar · View at Scopus
  18. S. Chaudhuri, S. Chatterjee, N. Katz, M. Nelson, and M. Goldbaum, “Detection of blood vessels in retinal images using two-dimensional matched filters,” IEEE Transactions on Medical Imaging, vol. 8, no. 3, pp. 263–269, 1989. View at Publisher · View at Google Scholar · View at Scopus
  19. A. Osareh and B. Shadgar, “Automatic blood vessel segmentation in color images of retina,” Iranian Journal of Science & Technology Transaction B Engineering, vol. 33, no. 2, pp. 191–206, 2009. View at Google Scholar
  20. F. Zana and J. C. Klein, “Segmentation of vessel-like patterns using mathematical morphology and curvature evaluation,” IEEE Transactions on Image Processing, vol. 10, no. 7, pp. 1010–1019, 2001. View at Publisher · View at Google Scholar · View at Scopus
  21. E. F. Kao, P. C. Lin, M. C. Chou, T. S. Jaw, and G. C. Liu, “Automated detection of fovea in fundus images based on vessel-free zone and adaptive Gaussian template,” Computer Methods & Programs in Biomedicine, vol. 117, no. 2, pp. 92–103, 2014. View at Publisher · View at Google Scholar · View at Scopus
  22. R. J. Qureshi, L. Kovacs, B. Harangi, B. Nagy, T. Peto, and A. Hajdu, “Combining algorithms for automatic detection of optic disc and macula in fundus images,” Computer Vision & Image Understanding, vol. 116, no. 1, pp. 138–145, 2012. View at Publisher · View at Google Scholar · View at Scopus
  23. D. Welfer, J. Scharcanski, and D. R. Marinho, “Fovea center detection based on the retina anatomy and mathematical morphology,” Computer Methods & Programs in Biomedicine, vol. 104, no. 3, pp. 397–409, 2011. View at Publisher · View at Google Scholar · View at Scopus
  24. B. Zhang, X. Wu, J. You, Q. Li, and F. Karray, “Detection of microaneurysms using multi-scale correlation coefficients,” Pattern Recognition, vol. 43, no. 6, pp. 2237–2248, 2010. View at Publisher · View at Google Scholar · View at Scopus
  25. M. S. Haleem, L. Han, J. V. Hemert, and B. Li, “Automatic extraction of retinal features from colour retinal images for glaucoma diagnosis: a review,” Computerized Medical Imaging & Graphics, vol. 37, no. 7-8, pp. 581–596, 2013. View at Publisher · View at Google Scholar · View at Scopus
  26. S. Dua, U. R. Acharya, P. Chowriappa, and S. V. Sree, “Wavelet-based energy features for glaucomatous image classification,” IEEE Transactions on Information Technology in Biomedicine, vol. 16, no. 1, pp. 80–87, 2012. View at Publisher · View at Google Scholar · View at Scopus
  27. S. Yousefi, M. H. Goldbaum, M. Balasubramanian et al., “Glaucoma progression detection using structural retinal nerve fiber layer measurements and functional visual field points,” IEEE Transactions on Biomedical Engineering, vol. 61, no. 4, pp. 1143–1154, 2014. View at Publisher · View at Google Scholar · View at Scopus
  28. L. Giancardo, F. Meriaudeau, T. P. Karnowski et al., “Exudate-based diabetic macular edema detection in fundus images using publicly available datasets,” Medical Image Analysis, vol. 16, no. 1, pp. 216–226, 2011. View at Publisher · View at Google Scholar · View at Scopus
  29. S. Kankanahalli, P. M. Burlina, Y. Wolfson, D. E. Freund, and N. M. Bressler, “Automated classification of severity of age-related macular degeneration from fundus photographs,” Investigative Ophthalmology & Visual Science, vol. 54, no. 3, pp. 1789–1796, 2013. View at Publisher · View at Google Scholar · View at Scopus
  30. M. R. K. Mookiah, U. R. Acharya, C. K. Chua, C. M. Lim, E. Y. K. Ng, and A. Laude, “Computer-aided diagnosis of diabetic retinopathy: a review,” Computers in Biology & Medicine, vol. 43, no. 12, pp. 2136–2155, 2013. View at Publisher · View at Google Scholar · View at Scopus
  31. C. Sinthanayothin, J. F. Boyce, T. H. Williamson et al., “Automated detection of diabetic retinopathy on digital fundus images,” Diabetic Medicine, vol. 19, no. 2, pp. 105–112, 2002. View at Google Scholar
  32. H. Li and J. Liu, “Image based grading of nuclear cataract by SVM regression,” in Proceedings of SPIE–The International Society for Optical Engineering, Bellingham, 2008.
  33. H. Li, L. Ko, J. H. Lim, J. Liu, D. W. K. Wong, and T. Y. Wong, “Image based diagnosis of cortical cataract,” in International Conference of the IEEE Engineering in Medicine & Biology Society, pp. 3904–3907, Los Alamitos, 2008. View at Publisher · View at Google Scholar
  34. H. Li, J. H. Lim, L. Jiang, and D. W. K. Wong, “Automatic detection of posterior subcapsular cataract opacity for cataract screening,” Conference Proceedings: Annual International Conference of the IEEE Engineering in Medicine and Biology Society, vol. 2010, no. 10, pp. 5359–5362, 2010. View at Publisher · View at Google Scholar · View at Scopus
  35. A. M. Abdul-Rahman, M. Tim, and C. B. Molteno Anthony, “Fourier analysis of digital retinal images in estimation of cataract severity,” Clinical & Experimental Ophthalmology, vol. 36, no. 7, pp. 637–645, 2008. View at Publisher · View at Google Scholar · View at Scopus
  36. M. Yang, J. J. Yang, Q. Zhang, and Y. Niu, “Classification of retinal image for automatic cataract detection,” in IEEE International Conference on E-Health Networking, Applications and Services, pp. 674–679, Los Alamitos, 2003.
  37. J. J. Yang, J. Li, R. Shen et al., “Exploiting ensemble learning for automatic cataract detection and grading,” Computer Methods & Programs in Biomedicine, vol. 124, pp. 45–57, 2016. View at Publisher · View at Google Scholar · View at Scopus
  38. J. R. Quinlan, C4. 5: Programs for Machine Learning, Elsevier, San Fransico, 2014. View at Publisher · View at Google Scholar · View at Scopus
  39. J. Cohen, “A coefficient of agreement for nominal scales,” Educational & Psychological Measurement, vol. 20, no. 1, pp. 37–46, 1960. View at Publisher · View at Google Scholar · View at Scopus
  40. N. G. Schaefer, T. F. Hany, C. Taverna et al., “Non-Hodgkin lymphoma and Hodgkin disease: coregistered FDG PET and CT at staging and restaging—do we need contrast-enhanced CT?” Radiology, vol. 232, no. 3, pp. 823–829, 2004. View at Publisher · View at Google Scholar · View at Scopus
  41. T. Stamm, A. Hohoff, A. V. Meegen, and U. Meyer, “On the three-dimensional physiological position of the temporomandibular joint,” Journal of Orofacial Orthopedics, vol. 65, no. 4, pp. 280–289, 2004. View at Publisher · View at Google Scholar
  42. J. R. Landis and G. G. Koch, “The measurement of observer agreement for categorical data,” Biometrics, vol. 33, no. 1, pp. 159–174, 1977. View at Google Scholar