Advancements of Medical Image Enhancement in Healthcare Applications
View this Special IssueResearch Article  Open Access
The Edge Detectors Suitable for Retinal OCT Image Segmentation
Abstract
Retinal layer thickness measurement offers important information for reliable diagnosis of retinal diseases and for the evaluation of disease development and medical treatment responses. This task critically depends on the accurate edge detection of the retinal layers in OCT images. Here, we intended to search for the most suitable edge detectors for the retinal OCT image segmentation task. The three most promising edge detection algorithms were identified in the related literature: Canny edge detector, the twopass method, and the EdgeFlow technique. The quantitative evaluation results show that the twopass method outperforms consistently the Canny detector and the EdgeFlow technique in delineating the retinal layer boundaries in the OCT images. In addition, the mean localization deviation metrics show that the twopass method caused the smallest edge shifting problem. These findings suggest that the twopass method is the best among the three algorithms for detecting retinal layer boundaries. The overall better performance of Canny and twopass methods over EdgeFlow technique implies that the OCT images contain more intensity gradient information than texture changes along the retinal layer boundaries. The results will guide our future efforts in the quantitative analysis of retinal OCT images for the effective use of OCT technologies in the field of ophthalmology.
1. Background
Optical coherence tomography (OCT) is the optical equivalent of ultrasonography, with the capability of capturing the depthresolved crosssectional images of biological tissues in vivo at nearhistologic resolution [1]. Due to its noninvasiveness and high resolution, in combination with the characteristics of the eye and retinal anatomy, OCT has a rapid development of clinical applications in ophthalmology in recent years.
Quantitative analysis of retinal OCT image has been critical for reliable and efficient diagnosis of diseases such as glaucoma, agerelated macular degeneration, and macular edema caused by diabetic retinopathy and for the evaluation of development of diseases, medical treatment responses, drug effectiveness, visual functions, and so forth [2–4]. Among others [5–7], automatic and semiautomated measurement of retinal layer thickness is considered as a class of key quantitative analysis. Numerous research efforts have been devoted to this topic [8–11], and these efforts have significantly promoted the clinical understanding of ocular diseases and improved the OCT technologies and their applications.
Retinal layer thickness measurement relies on accurate OCT image segmentation. For many automated segmentation algorithms, edge detection is an essential foundation [12–19], notwithstanding some methods resort to other features of the images [18]. Literature shows that diverse types of edge detection algorithms can be employed as a key step in image segmentation. Table 1 summarizes the commonly used algorithms for retinal OCT image segmentation, including the Canny edge detector [12, 13], twopass edge detection method [14, 15], local mean gradientbased edge tracking [16], peak detection method, Gaussian smoothing in combination with the Sobel kernel method [17, 18], and EdgeFlow technique [19]. Based on the nature of the information used in their algorithms, we can classify these different edge detection techniques into two categories, namely, the intensitybased and texturebased methods. The former category utilizes the intensity gradient in the images, whereas the latter tracks the texture changes rather than the intensity gradient.

Given the importance of edge detection for retinal image segmentation and the diversity of edge detectors, a natural question is which method gives the best edge detection outcome for current application. To our best knowledge, there is no study so far systematically evaluating the performance of the edge detectors applied for OCT image analysis. General performance evaluation of edge detection has long been interesting many researchers, but prior studies have not reached a unanimous conclusion yet because of the complexity of the problem. According to Heath et al., the challenge lies in that the edge detection performance depends not only on the algorithms themselves but also on the images applied to, the parameters used in the specific case, and the evaluation metrics [20].
Our goal in this paper is to evaluate what type of edge detectors best suit for retinal OCT image segmentation with given equivalent parameters, when measured using performance metrics that are meaningful for retinal layer thickness quantification. To this end, we first collect the commonly used edge detection methods in the literature on OCT image analysis and choose the most representative ones for comparison. We then research the edge detection performance evaluation literature to select the most relevant performance metrics that are meaningful for OCT image segmentation and adapt them when necessary. Using these metrics, we examine which of the selected edge detectors gives the best edge detection outcome when they are applied to the OCT images that we have collected from healthy subjects.
The remaining parts of the article are organized as follows. In the next section, we describe our research materials and methods, including retinal OCT image data collection, screening edge detectors for comparison by reviewing and analyzing the edge detection techniques used in prior studies on the retinal OCT image segmentation, and determining the most relevant performance evaluation metrics. In the third section, we present the comparison of the representative advanced approaches to retinal OCT image edge detection against different performance evaluation metrics. Finally, we discuss the findings and research opportunities.
2. Materials and Methods
We conducted this study in accordance with the Tenets of the World Medical Association’s Declaration of Helsinki [21]. Ethical approvals were obtained from the Ethical Review Board of Southern Medical University, the Ethical Review Board of Sun Yatsen University, and the Research Ethical Committee of Zhongshan Ophthalmic Center. After an introduction about the purpose of the study and explanation of the process and risks, the voluntary participants signed the informed consent for this data collection.
2.1. Image Data Choice and Data Collection
In this article, our goal is to evaluate the performance of edge detection algorithms for retinal OCT image segmentation. Because of their image dependence [20], edge detectors perform well for other types of images but may not give as good results when applied to retinal OCT image. Therefore, all interested algorithms are tested on retinal OCT image in this study.
We collected the image data from 11 healthy volunteers (age ranges from 21 to 29, 7 males) using Topcon 3D OCT2000 (Topcon Corporation, Tokyo, Japan) at Zhongshan Ophthalmic Center, a tertiary specialized hospital affiliated to Sun Yatsen University, Guangzhou, China. Using 7 line 6.0 scanning mode to scan the macular area with a resolution of 1024 Ascans, we obtained the raw retinal OCT images. Due to the limitation of the software (software version: 8.20.003.04, Topcon Corporation, Tokyo, Japan), the raw images saved into bmp files have a size of 759 × 550 pixels. As the best available from our OCT equipment, this resolution is above average within those of OCT images reported in the most recent literature [3, 6, 7, 17–19]. Figure 1 shows a typical OCT image in our dataset. In this image, the 6 layer boundaries, namely, ILM, NFLGCL, IPLINL, OPLONL, ONLIPS, and RPEchoroid are readily observable.
2.2. Edge Detection Algorithms
Three criteria for our choice of edge detectors for evaluation were (1)to include the algorithms that have been most commonly used in the OCT image segmentation literature,(2)to give preference to the ones representing the state of the art in edge detection, which were usually used to detect more than 3 retinal layer boundaries,(3)to include a diverse mix of algorithms utilizing different image feature information.
Using these criteria, three algorithms reviewed in Table 1 were chosen for further analysis. They are the Canny edge detector [21], twopass algorithm [14, 15], and EdgeFlow method [22]. The former two are based on the image intensity gradient, whereas the last one is based on the image texture changes. We refer to the first two as the intensitybased edge detection methods and the third as the texturebased edge detection method. The principles of the three algorithms will be outlined below. The rest three edge detection methods, all intensitybased, were excluded from further analysis for different reasons. Among them, the edgetracking algorithm based on the maximization of the local mean gradient was used only to detect ILM [16]; the peak detection method [17] and the Sobel kernel in combination of Gaussian smoothing [18] only detect the easily detectable boundaries ILM and OS/RPE.
2.2.1. Canny Edge Detector
The Canny edge detection algorithm [23] is now generally regarded as the “standard” for edge detection in the field of digital image processing ([24], Chapter 10). The Canny edge detector works in a multistep process to detect a wide range of edges in images. At first, the image is smoothed using a linear Gaussian filter. Then, a 2D first derivative operator is utilized on the smoothed image to compute the derivatives in both the vertical and the horizontal orientations. The gradient magnitude is calculated as the root sum of squares of the derivatives in two orthogonal directions and the gradient phase as the arctangent of their ratio. Candidate edge pixels are identified as the pixels that survive after a thinning process called nonmaximal suppression. In this process, the edge strength of each candidate edge pixel is set to zero if its gradient magnitude is not larger than the gradient magnitude of the two adjacent pixels in the gradient direction, and the pixel whose gradient magnitude is the local maximum is preserved. At last, hysteresis thresholding is used to eliminate weak edge points and track the possible edge pixels. In this step, doublethreshold T1 and T2 with T2 > T1 are applied; all candidate edge pixels below the lower threshold T1 are set to zero, and all pixels above the lower threshold T1 can be connected to any pixels above the higher threshold T2 through a chain of edge pixels which are labeled as edge pixels. The hysteresis helps in ensuring that the noisy edges are not broken into multiple edge fragments. In the Canny edge detection algorithm, three parameters are incorporated, which play a decisive role for detecting the result. One is the width of Gaussian filter (i.e., standard deviation of the Gaussian, σ). An increase in the width of Gaussian filter reduces the detector’s sensitivity to noise, but blurs the image and results in loss of finer edge details. The other two are the lower threshold (T1) and the higher threshold (T2), respectively. The higher threshold should be set reasonably high and the lower threshold quite low for good detection results, because if it is too high, the lower threshold causes edge fragments and if too low, the higher threshold increases false alarms and undesirable edge fragments in the edge detection output.
2.2.2. TwoPass Edge Detection Algorithm
Twopass edge detection algorithm is designed exclusively for detection of retinal layer in OCT images by Bagci et al. [14, 15]. The feature of edges in retinal OCT image, extending along the horizontal direction with a gentle up and downslope, was taken into account in the algorithm. The edge detection kernel L(x, y) is based on the first derivative of Gaussian in the vertical direction:
The parameter p determines the polarity of edges and takes values either 1 or −1. The edge detection kernel is applied twice with alternating values of p. On the first pass, the boundaries between each pair of adjacent bright and dark regions, with bright on the top, such as NFLGCL, IPLINL, and OPLONL, are extracted with . On the second pass, boundaries between each pair of adjacent bright and dark regions, with dark on the anterior, such as ILM, INLOPL, and ONLIPS, were detected with . The peak values are marked as edges, using nonmaximal suppression and hysteresis thresholding. Satisfactory results can be obtained by adjusting the value of σ.
2.2.3. EdgeFlow Technique
The EdgeFlow technique is a novel boundary detection scheme proposed by Ma and Manjunath [22]. The technique for boundary detection based on EdgeFlow utilizes a predictive coding model to characterize the direction of change in color (intensity of grey image) and texture at each image location at a given scale and constructs an EdgeFlow vector. By propagating the EdgeFlow vectors, the boundaries can be detected at image locations which encounter two opposite directions of flow in the stable state. Differing from intensitybased detection methods that focus on finding the local gradient maximum, EdgeFlow technique computes the directions of edge energy according to intensity or texture in an image and associated probabilities. The edge energy and corresponding probabilities obtained from different image attributes are pooled together to form a single edge field for boundary detection: where and represent the energy and probability of the EdgeFlow computed from image attribute , ; is the weighting coefficient associated with image attribute . The edge flow direction is estimated as follows:
The EdgeFlow vector is then defined as where is a complex number with its magnitude representing the resulting edge energy and phase representing the flow direction. After the EdgeFlow vector of an image is computed, boundary detection can be performed by propagating the EdgeFlow vector and identifying the locations where two opposite flow directions encounter each other. The scheme facilitates integration of intensity and texture into a single framework for boundary detection.
2.3. Performance Evaluation Using Ground Truth
According to Heath et al. [20], edge detection performance evaluation can be classified into theoretical and empirical approaches. The former uses pure mathematical analysis without the algorithms ever being applied to an image. It has major limitation for not being able to deal with the complexity of modern edge detection algorithms. The latter can be further classified into (1) evaluation using ground truth and (2) evaluation without ground truth. Our goal in this study is to examine how different edge detectors give the best results for OCT retinal layer segmentation. Ultimately, we hope to identify the most reliable and efficient edge detectors to help doctors automate the measurement of retinal layer thickness in order to make quantitatively informed medical decisions. Our human vision systems are the most complex and efficient machine for image analysis, including edge detection. Therefore, for our intended application, the most appropriate evaluation approach should be evaluation using ground truth, which measures the difference between the algorithmdetected edges and the humandetected edges.
Due to the importance of edge detector performance evaluation using ground truth, researchers have developed numerous metrics. These metrics can be largely classified into three categories. The first category, which we refer to as the edge presence accuracy metrics (EPAM), focuses on to which extent the detected edges coincide with the ground truth without considering location shift. EPAM include mainly four metrics [25–28], namely, true positive rate, false positive rate, false negative rate, and total edge detection accuracy. The first three, respectively, measure the ratio of true edge pixels, falsely detected edge pixels, and missed edge pixels to the number of total edge pixels in the ground truth, and the fourth is the ratio between the total true edge pixels and true nonedge pixels and the total number of pixels in the region of interest. The second category, which we refer to as the edge location accuracy metrics (ELAM), focuses on the extent of edge shifts [29] introduced by the edge detection algorithms as compared to the ground truth. Metrics in this category include Hausdorff’s distance [30], which measures the similarity between two images, and mean localization deviation (MLD). The third category takes into account both location accuracy and edge presence. Typical metrics include figure of merit (FOM) [31] and its expanded version (expanded FOM) [32] and multifeature quality measurement [25].
2.3.1. Criteria for Performance Evaluation Metrics
OCT retinal layer segmentation aims to automate retinal layer thickness measurement in order to free ophthalmologists from laborious manual tracing of the layer boundaries. The ideal layer edge detector would give the same thickness measures to those from ground truth specified by human observers. However, even experts could not arrive at the same segmentation for a given retinal OCT image [33]. This is because manual segmentation is subject to human subjectiveness. The ground truth used for the evaluation is not really the ultimate truth. Thus, it is important to note that the traditional edge presence accuracy metrics, the probabilities of true positive, false positive (spurious edges), and missing edge, cannot offer the complete evaluation of edge detector performance. We propose that the performance metrics need to meet the criteria as follows: (1)To measure the edge presence accuracy by calculating the rates of true positive, false positive, true negative, and false negative (missing)(2)To measure edge location accuracy by calculating the signed and unsigned edge shift distance(3)To allow edge shift when calculating edge presence accuracy(4)To examine the computational costs.
2.3.2. Evaluation Metrics for this Study
Based on our analysis of existing performance metrics in the literature and the metric criteria discussed previously, we choose the figure of merit (FOM, Pratt) [31], true positive rate (TPR), false positive rate (FPR), accuracy (ACC), and mean localization deviation (MLD) [34] as the basis to develop our procedure for comprehensive evaluation of the chosen edge detectors. In the paragraphs to follow, we outline the principles of these metrics.
(1) Pratt’s Figure of Merit. FOM [31] is a classical metric utilized by numerous researchers for evaluating the performance of edge detection algorithms [25, 26, 32, 35, 36]. The definition of the FOM is given by where and represent the number of ideal and actual detected edge pixels, denotes the distance between the ith detected edge pixel and its correct position, and is the scaling constant (normally set at 1/9) that is applied to provide a relative penalty between smeared edges and isolated, but offset, edges.
(2) Edge Presence Accuracy. The criteria on which the FOM of Pratt is based include missed valid edges, localization errors, and false alarms. Different configurations of detected edges may yield equal FOM value [35]. In order to decompose the sources of difference, Yin et al. [27] developed three metrics (TPR, FPR, and ACC) that are defined as follows.
True positive rate ():
False positive rate ():
Accuracy ():
In these equations, (true positive) and (true negative) represent the numbers of correctly detected edge pixels and nonedge pixels. (false positive) is the number of pixels not belonging to edge but recognized as one by the algorithm, and (false negative) is the number of pixels belonging to edge but failed to be recognized by the algorithm. is the total number of pixels within the ROI of the image, and is the number of ideal edge pixels.
Given the large number (N−N_{I}) of nonboundary pixels in the images, FPR calculated in the form of (7) is close to zero, making the metrics insensitive to the change of edge detection algorithms. We redefined it as where denotes the number of pixels of the actually detected edges.
(3) Edge Location Accuracy Metrics. It is known that some image processing procedures cause the shift of detected edges ([29], Chapter 3, p. 56). In order to characterize the extent to which the results from edge detection algorithms deviate from the ground truth, we introduce the location accuracy metrics, the mean localization deviation (MLD) in the context of OCT image analysis: where is the number of edge pixels in searching neighborhood of a ground truth edge pixel, the number of pixels in the ideal edge, and the Euclidean distance of the current edge pixel in ground truth and edge pixels in searching neighborhood. For the retinal OCT images, we limited the searching neighborhood to be within 3 pixels of the true edge along each Ascan.
(4) Adjusted TPR, FPR, and ACC. Some procedures of image processing can introduce edge shift ([29], Chapter 3, p. 56–74). As a result, the detected edge may not match the position of actual edge. As the goal of retinal OCT image segmentation is to extract the contours of retinal layer boundaries and measure the thicknesses of different retinal layers, small and constant shifts do not have effective impact when the layer thicknesses are of the only interest. Therefore, edge pixels in the neighborhood detected by the algorithms may be accepted into true positive edge pixels when calculating the edge presence metrics. In this case, FOM, the true positive rate, and false positive rate and accuracy measures need also to be adjusted. We define these adjusted metrics as , , , and : where and represent the number of ideal and actual detected edge pixels, is the distance between the ith detected edge pixel and its correct position, is the number of edge pixels detected by an algorithm that are considered as edges within the neighborhood of the ground truth, is the number of false positive pixels after the neighborhood searching and edge pixel adjustment, and equals as it is not affected by the neighborhood adjustment. These adjusted metrics allowing the edge shift can better reflect the amount of detected edge points.
2.3.3. Evaluation Procedure
The major steps for performance measurement include the preparation of ground truth, the preprocessing of OCT images, the application of edge detectors with appropriate parameters to obtain the near optimal outcome for each detector, and using the performance metrics to evaluate the goodness of edge detectors against the ground truth. Figure 2 summarizes the flow of performance evaluation.
(1) Ground Truth Preparation. We asked an expert observer to manually delineate the edges for representative retinal OCT images to form a base dataset of ground truth, as noted by I_{ref} in Figure 2. Because the ILM and RPE are the outer boundaries of the retinal structure and they are strong edges that can usually be reliably detected, we define the images between ILM and RPE (included) as the region of interest (ROI). Only those edges within the ROI are extracted for comparison with the ground truth.
(2) Image Preprocessing. Before applying the computer algorithm for each edge detector, we conducted necessary image preprocessing. Due to constructive or destructive interference of the light waves from the object, spectral domain retinal OCT images suffer from the inherit speckle noise [37], which decreases the quality of image and causes unreliable retinal layer segmentation. In order to improve the quality of edge detection, preprocessing becomes a necessary step. We first converted the raw OCT image bmp files into grayscale images and cropped the images to the region of interest (ROI, 200 by 400 pixels) in this study. The literatures have suggested the use of filters like mean, median, and Gaussian [38–40] for noise removal. We choose median filtering to remove the speckle noise. The original retinal OCT image and the denoised image are shown in Figure 3.
(a)
(b)
(3) Edge Detection. We randomly chose 8 images from our database of raw OCT retinal images and apply the three edge detectors. As discussed earlier, the edge detection outcomes may be influenced not only by the algorithm itself but also by the input parameters [20]. We varied the parameters systematically to obtain the optimal possible edge outcomes for each of the edge detection algorithms.
(4) Performance Evaluation. In edge detection performance evaluation step, we compared the edge detection outcomes from the three computer algorithms against the human manually traced retinal layer boundaries. We applied the metrics that were broadly used in the literature and relevant to our specific research context and purpose. We also applied the adjusted metrics developed and discussed in the previous section. Finally, we examined the evaluation outcomes in terms of both differences and relationships.
3. Experiments and Results
We implemented all the data analysis in MATLAB R2012a (The Mathworks Inc., MA, USA) on a personal computer running Windows 7 operating system with an 3.60 GHz Intel® Core™ i7 CPU and 4 GB of memory. The raw image data was acquired and stored into bmp files. Data in the intermediate analysis steps were computed and stored in the double precision data format in order to minimize digitization errors. Three edge detection algorithms were carefully coded and double checked for the correctness.
3.1. Edge Detection
3.1.1. Input Parameters
Input parameters can significantly influence the resulting edge quality for given edge detection algorithm [20]. In selecting values for these parameters, we are interested in finding the set that provides good edge detection accuracy, that is, the boundaries coinciding the six retinal layers as shown in Figure 1 with high signaltonoise ratio. In terms of performance evaluation metrics, good parameters give the high values of FOM, TPR, ACC, and their adjusted forms and lower values of FPR and MLD. Through applying the three detection methods with multiple sets of parameters on OCT retinal images (), the edge detection results were obtained and compared. By observing the outcomes, we chose the parameter set for each algorithm as follows. In Canny edge detection, the best result can be obtained by setting the width of the Gaussian filter , the lower threshold , and the upper threshold ; for twopass edge detection algorithm, the width of Gaussian , consistent with Bagci et al. [14], the lower threshold , and the upper threshold , similar to those for Canny edge detector; for the EdgeFlow algorithm, we followed Ma and Manjunath [22] and chose the equal weighting coefficients for intensity and texture, that is, ω(intensity) = ω(texture) = 0.5.
3.1.2. Edge Detection Results
Figure 4(a) shows an original retinal OCT image. Figures 4(b), 4(c), and 4(d) show the edge detection outcomes from the Canny edge detector, twopass edge detection technique, and EdgeFlow algorithm, respectively. From these edge detection results, six retinal layer boundaries of our interest are readily identifiable, although with some noises caused by false positives and boundary breakages caused by the false negatives.
(a)
(b)
(c)
(d)
For our purpose in this study, we were mainly interested in how the three edge detectors performed in detecting the 6 retinal layer boundaries, which were also the key information in the literatures for retinal layer thickness measurement [8–11]. We defined the region of interest (ROI) to be the area between the ILM and RPE that are the most outer boundaries of the retinal structure. Figure 5(a) is the OCT retinal image with overlaid ground truth edges marked by an expert observer. Figures 5(b), 5(c), and 5(d) show the edges within ROI detected by the three algorithms, which will be the basis of performance evaluation in the next section. Visually, the result from the Canny edge detector in Figure 5(b) shows a welldefined six boundaries, although with some breakages and noises. The result from the twopass method shown in Figure 5(c) gives more than 6 layers in some locations, but in general, the six layer boundaries of interest are very clear with less breakages compared to those in Figure 5(b). The result from EdgeFlow algorithm depicted in Figure 5(d) shows more breakages and more noises, although all 6 layer boundaries are still recognizable.
(a)
(b)
(c)
(d)
3.2. Performance Evaluation
To quantify the performance of the three edge detectors, we use three sets of measurements discussed in Materials and Methods. The first set of metrics include FOM, TPR, FPR, and ACC, which have been broadly used in the literature [20, 25–28, 30–32] when evaluating edge detectors on images other than the OCT retinal images. In order to calculate TPR, FPR, and ACC, we use the ground truth as a template to screen the coincided edge pixels from the three edge detection algorithms. Figure 6 shows the edge points overlapping with the manually traced edges (ground truth).
(a)
(b)
(c)
(d)
The calculated performance metrics are summarized in Table 2. Based on the mean values of FOM for the three edge detectors, it seems that the best performer is the twopass method. Table 3 summarizes the results of the statistical analysis. Twosample ttests confirmed the impression that twopass method significantly outperforms both the Canny edge detector (), and EdgeFlow () in terms of FOM. In addition, the EdgeFlow algorithm outperforms the Canny edge detector (). The TPR for twopass is also significantly higher than that for Canny () and that for the EdgeFlow method (). The same pattern occurs when measured with ACC. ACC for the twopass method is significantly higher than that for the Canny () and EdgeFlow method (). On the other hand, FPR for twopass method is significantly lower than that for Canny () and EdgeFlow (). All the metrics suggest that twopass method is the best among the three. However, Canny and EdgeFlow methods are not significantly different when measured using TPR (), FPR (), or ACC ().


However, when comparing Figures 5(b), 5(c), and 5(d) and Figures 6(b), 6(c), and 6(d) one by one, the Canny edge detector results in much less edge pixels on locations of the ground truth, although its outcome seems good as well in Figure 5(b). This suggests that the Canny edge detector introduced edge shifts and resulted and lower performance score in Tables 2 and 3. To examine the possibility, we calculated the mean localization deviation (MLD) for the three edge detectors. The results in Table 4 indeed show the largest mean value of MLD for the Canny edge detector. Twosample ttests (Table 5) confirmed that the MLD value for the Canny edge detector is significantly higher than that for the twopass edge detection algorithm () and almost significantly higher than that for the EdgeFlow method ().


For OCT retinal image layer thickness measurement, if the edge shift is within a small range and to the same direction, the outcome may not be significantly influenced. Therefore, we examine how the edge detectors perform when measured with a second set of metrics; the adjusted measures include FOM_{ADJ}, TPR_{ADJ}, FPR_{ADJ}, and ACC_{ADJ}, which were developed in Materials and Methods. Figure 7(a) is the ground truth edges; Figures 7(b), 7(c), and 7(d), respectively, show detected edge pixels within 2 pixels searching neighborhood corresponding to its ground truth edge for the three edge detectors. Visually, the outcome for Canny edge detector is much improved when compared to that in Figure 6(b); the edges for EdgeFlow method are much noisy. Quantitative metrics for the edge detectors based on Figure 7 are summarized in Table 6.
(a)
(b)
(c)
(d)

The results for FOM, TPR, ACC, and FPR after the adjustment (Table 6) are all better than those before the adjustment (Table 2). This finding is reasonable in that the false alarm reduces, and FOM, TPR, and ACC increase, when more detected pixels are considered as correct edge. Note that in both tables, the higher values for FOM, TPR, and ACC mean the better performance, whereas for FPR, the lower the value, the better the performance.
Detailed statistical analysis on the performance of three edge detectors measured with the adjusted metrics is summarized in Table 5. For FOM_{ADJ} measure, the twopass is again significantly better than both Canny edge detector () and the EdgeFlow method (); EdgeFlow is also significantly better than Canny (). When measured with TPR_{ADJ}, FPR_{ADJ}, and ACC_{ADJ}, Canny does not differ significantly from the twopass method (, , and , resp.); a similar pattern (except for TPR_{ADJ}) occurs for Canny and the EdgeFlow method (, , and , resp.). However, these measures show that the twopass method is significantly better than the EdgeFlow technique (, , and , resp.).
Finally, we noticed in our experiments that the EdgeFlow technique took much longer time for each edge detection task. When the time was measured for processing a sample of OCT retinal images (200 by 400 pixels in the ROI), the average computational time is 2.77 ± 1.24 seconds, 3.85 ± 0.18 seconds, and 467.66 ± 1.33 seconds, respectively, for twopass, Canny, and EdgeFlow methods, confirming that the EdgeFlow approach took a significantly larger amount of time than the other two algorithms.
4. Discussion and Conclusions
In this study, we intended to search for the edge detectors that best suit for the OCT retinal image segmentation task. With the analysis of literature and our experiment, we have identified the most promising candidate algorithms, namely, Canny edge detector, the twopass method, and the EdgeFlow technique. Using the performance evaluation metrics (FOM, TPR, FPR, and ACC) and their adjusted versions (FOM_{ADJ}, TPR_{ADJ}, FPR_{ADJ}, and ACC_{ADJ}), we examined the three methods applied to the realistic OCT retinal images. Our results show that the twopass method consistently outperforms the other two. In addition, the MLD metrics shows that the twopass method caused smaller edge shifting problem. Although the computational cost for the twopass method is slightly higher than the Canny edge detector, it is over 100 times lower than that for the EdgeFlow technique. Based on the above analysis and findings, we conclude that the twopass method is among the three the best approach to edge detection for the OCT retinal layer image segmentation task. Furthermore, the outperformance of twopass method measured by the original and adjusted metrics and the advantage of Canny edge detector over EdgeFlow technique in terms of FOM_{ADJ} and TPR_{ADJ} and MLD lead to another conclusion that the intensitybased edge detectors outperform the texturebased edge detector for OCT retinal image analysis.
The findings in the study suggest that it is critical to use the most appropriate algorithms to detect the retinal layer boundaries in the OCT images in order to automate the quantitative analysis of retinal OCT images. Combined with the findings in the literature that EdgeFlow method significantly outperformed Canny algorithm in texture segregation tasks [22], this study offers support to the idea that the performance of edge detectors is image property dependent [20] as both Canny and twopass methods surpass EdgeFlow in the current application. In line with this thought and findings, it is necessary argue that the best performer for normal retinal OCT images also work best for pathological retinal images. Additionally, the intensitygradient based methods (twopass and Canny algorithms) outperforming texturebased method (EdgeFlow) might suggest that the OCT images contain more intensity gradient changes than texture changes along the longitudinal direction. The relative weight of intensity and texture information in OCT retinal image warrants further study in the future.
With the development of OCT technologies and their applications in the field of ophthalmology, more and more data is readily available. Extracting meaningful information from the everincreasing volume of clinical data reliably and efficiently forms the basis for modern medical decision making and research. Reliable and efficient OCT retinal image segmentation will contribute to the development of this trend. Future research efforts would need to overcome several limitations in this study. First, the input parameters used in our experiments were selected over a relatively small sample space and the decisions on the “optimal” parameters were subject to human subjectiveness. Although it is almost impossible to identify the absolutely optimal input parameters for each edge detector [20], the choice of optimal input parameters may be improved by conducting a large number of experiments and averaging opinions from more expert viewers. The second limitation in our study is the use of a single expert observer to define the ground truth. Individual subjectiveness may be reduced by averaging across multiple decisions for the ground truth. Moreover, our data were all collected from voluntary healthy subjects. If the edge detectors perform differently for different types of images, it is necessary to examine how they perform on pathological retinal images in future studies.
Conflicts of Interest
The authors declare that there are no competing interests regarding the publication of this paper.
Authors’ Contributions
Su Luo and Jing Yang contributed equally to this work.
Acknowledgments
This work was supported by the National Natural Science Foundation of China (no. 61271154 and no. 61172033). The authors would like to thank the peer reviewers for their insightful and constructive suggestions.
References
 D. Huang, E. A. Swanson, C. P. Lin et al., “Optical coherence tomography,” Science, vol. 254, no. 5035, pp. 1178–1181, 1991. View at: Google Scholar
 P. Hrynchak and T. Simpson, “Optical coherence tomography: an introduction to the technique and its use,” Optometry & Vision Science Official Publication of the American Academy of Optometry, vol. 77, no. 7, pp. 347–356, 2000. View at: Google Scholar
 W. Geitzenauer, C. K. Hitzenberger, and U. M. SchmidtErfurth, “Retinal optical coherence tomography: past, present and future perspectives,” British Journal of Ophthalmology, vol. 95, no. 2, pp. 171–177, 2011. View at: Publisher Site  Google Scholar
 M. Sonka and M. D. Abràmoff, “Quantitative analysis of retinal OCT,” Medical Image Analysis, vol. 33, pp. 165–169, 2016. View at: Publisher Site  Google Scholar
 G. Quellec, K. Lee, M. Dolejsi, M. K. Garvin, M. D. Abramoff, and M. Sonka, “Threedimensional analysis of retinal layer texture: identification of fluidfilled regions in SDOCT of the macula,” IEEE Transactions on Medical Imaging, vol. 29, no. 6, pp. 1321–1330, 2010. View at: Publisher Site  Google Scholar
 H. Bogunovic, M. Sonka, Y. H. Kwon, P. Kemp, M. D. Abramoff, and X. Wu, “Multisurface and multifield cosegmentation of 3D retinal optical coherence tomography,” IEEE Transactions on Medical Imaging, vol. 33, no. 12, pp. 2242–2253, 2014. View at: Publisher Site  Google Scholar
 X. Chen, P. Hou, and C. Jin, “Quantitative analysis of retinal layer optical intensities on threedimensional optical coherence tomography,” Investigative Ophthalmology & Visual Science, vol. 54, no. 10, pp. 6846–6851, 2013. View at: Google Scholar
 D. Koozekanani, K. Boyer, and C. Roberts, “Retinal thickness measurements in optical coherence tomography using a Markov boundary model,” IEEE Transactions on Medical Imaging, vol. 20, no. 9, pp. 900–916, 2001. View at: Publisher Site  Google Scholar
 A. Lang, A. Carass, and P. A. Calabresi, “An adaptive grid for graphbased segmentation in retinal OCT,” Proceedings of SPIE  the International Society for Optical Engineering, vol. 9034, no. 10, pp. 1314–1323, 2014. View at: Google Scholar
 R. J. Zawadzki, A. R. Fuller, and D. F. Wiley, “Adaptation of a support vector machine algorithm for segmentation and visualization of retinal structures in volumetric optical coherence tomography data sets,” Journal of Biomedical Optics, vol. 12, no. 4, article 041206, 2007. View at: Publisher Site  Google Scholar
 F. D. Cabrera, H. M. Salinas, and C. A. Puliafito, “Automated detection of retinal layer structures on optical coherence tomography images,” Optics Express, vol. 13, no. 25, pp. 10200–10216, 2005. View at: Google Scholar
 Q. Yang, C. A. Reisman, and Z. Wang, “Automated layer segmentation of macular OCT images using dualscale gradient information,” Optics Express, vol. 18, no. 20, pp. 21293–21307, 2010. View at: Google Scholar
 S. Lu, C. Y. Cheung, J. Liu, J. H. Lim, C. K. Leung, and T. Y. Wong, “Automated layer segmentation of optical coherence tomography images,” IEEE Transactions on Biomedical Engineering, vol. 57, no. 10, pp. 2605–2608, 2010. View at: Publisher Site  Google Scholar
 A. M. Bagci, R. Ansari, and M. Shahidi, “A method for detection of retinal layers by optical coherence tomography image segmentation,” in 2007 IEEE/NIH Life Science Systems and Applications Workshop, pp. 144–147, Bethesda, MD, USA, 2007. View at: Publisher Site  Google Scholar
 A. M. Bagci, M. Shahidi, R. Ansari, M. Blair, N. P. Blair, and R. Zelkha, “Thickness profiles of retinal layers by optical coherence tomography image segmentation,” American Journal of Ophthalmology, vol. 146, no. 5, pp. 679–687, 2008. View at: Publisher Site  Google Scholar
 F. Rossant, I. Ghorbel, and I. Bloch, “Automated segmentation of retinal layers in OCT imaging and derived ophthalmic measures,” in 2009 IEEE International Symposium on Biomedical Imaging: From Nano to Macro, pp. 1370–1373, Boston, MA, USA, 2009. View at: Publisher Site  Google Scholar
 Y. M. Cha and J. H. Han, “Highaccuracy retinal layer segmentation for optical coherence tomography using tracking kernels based on Gaussian mixture model,” IEEE Journal of Selected Topics in Quantum Electronics, vol. 20, no. 2, pp. 1–10, 2014. View at: Publisher Site  Google Scholar
 A. Lang, A. Carass, and M. Hauser, “Retinal layer segmentation of macular OCT images using boundary classification,” Biomedical Optics Express, vol. 4, no. 7, pp. 1133–1152, 2013. View at: Publisher Site  Google Scholar
 S. Niu, Q. Chen, and S. L. De, “Automated retinal layers segmentation in SDOCT images using dualgradient and spatial correlation smoothness constraint,” Computers in Biology & Medicine, vol. 54, no. C, pp. 116–128, 2014. View at: Publisher Site  Google Scholar
 M. D. Heath, S. Sarkar, and T. A. Sanocki, “A robust visual method for assessing the relative performance of edgedetection algorithms,” IEEE Transactions on Pattern Analysis & Machine Intelligence, vol. 19, no. 12, pp. 1338–1359, 1997. View at: Publisher Site  Google Scholar
 A. Nicogossian, O. Kloiber, and B. Stabile, “The Revised World Medical Association's Declaration of Helsinki 2013: enhancing the protection of human research subjects and empowering ethics review committees,” World Medical & Health Policy, vol. 6, no. 1, pp. 1–3, 2014. View at: Google Scholar
 W. Y. Ma and B. S. Manjunath, “EdgeFlow: a technique for boundary detection and image segmentation,” IEEE Transactions on Image Processing, vol. 9, no. 8, pp. 1375–1388, 2000. View at: Publisher Site  Google Scholar
 J. Canny, “A computational approach to edge detection,” IEEE Transactions on Pattern Analysis & Machine Intelligence, vol. 8, no. 6, pp. 679–698, 1986. View at: Publisher Site  Google Scholar
 R. C. Gonzalez and R. E. Woods, Digital Image Processing, Publishing House of Electronics Industry, 2nd edition, 2002.
 R. RománRoldán, J. F. GómezLopera, and C. AtaeAllah, “A measure of quality for evaluating methods of segmentation and edge detection,” Pattern Recognition, vol. 34, no. 5, pp. 969–980, 2001. View at: Publisher Site  Google Scholar
 I. E. Abdou and W. Pratt, “Quantitative design and evaluation of enhancement/thresholding edge detectors,” Proceedings of the IEEE, vol. 67, no. 5, pp. 753–763, 1979. View at: Publisher Site  Google Scholar
 X. Yin, B. W. Ng, and J. He, “Accurate image analysis of the retina using hessian matrix and binarisation of thresholded entropy with application of texture mapping,” PLoS One, vol. 9, no. 4, article e95943, 2014. View at: Publisher Site  Google Scholar
 E. Nadernejad, S. Sharifzadeh, and H. Hassanpour, “Edge detection techniques: evaluations and comparisons,” Applied Mathematics Sciences, vol. 2, no. 31, pp. 1507–1520, 2008. View at: Google Scholar
 E. R. Davies, Computer and Machine Vision: Theory, Algorithms, Practicalities, China Machine Press, 4th edition, 2013.
 M. Beauchemin, “On the Hausdorff distance used for the evaluation of segmentation results,” Canadian Journal of Remote Sensing, vol. 24, no. 1, pp. 3–8, 1998. View at: Google Scholar
 W. K. Pratt, Digital Image Processing, WileyInterscience, New York, NY, USA, 1978.
 K. C. Strasters and J. J. Gerbrands, “Threedimensional image segmentation using a split, merge and group approach,” Pattern Recognition Letters, vol. 12, no. 5, pp. 307–325, 1991. View at: Publisher Site  Google Scholar
 S. J. Chiu, X. T. Li, and P. Nicholas, “Automatic segmentation of seven retinal layers in SDOCT images congruent with expert manual segmentation,” Optics Express, vol. 18, no. 18, pp. 19413–19428, 2010. View at: Google Scholar
 Z. Xinyan and Z. Ranyang, “An adaptive edgedetection algorithm based on Canny and its performance evaluation,” Computer Technology and Development, vol. 25, no. 11, pp. 32–37, 2015. View at: Google Scholar
 F. van der Heyden, “Evaluation of edge detection algorithms,” in Third International Conference on Image Processing and its Applications, 1989, pp. 618–622, Warwick, UK, 1989. View at: Google Scholar
 S. Chabrier, H. Laurent, and B. Emile, “A comparative study of supervised evaluation criteria for image segmentation,” in 2004 12th European Signal Processing Conference, pp. 1143–1146, Vienna, Austria, 2004. View at: Google Scholar
 E. Götzinger, M. Pircher, B. Baumann et al., “Speckle noise reduction in high speed polarization sensitive spectral domain optical coherence tomography,” Optics Express, vol. 19, no. 15, pp. 14568–14585, 2011. View at: Google Scholar
 Y. Huang, R. P. Danis, and J. W. Pak, “Development of a semiautomatic segmentation method for retinal OCT images tested in patients with diabetic macular edema,” PLoS One, vol. 8, no. 12, article e82922, 2013. View at: Publisher Site  Google Scholar
 A. M. Abhishek, T. T. J. M. Berendschot, and S. V. Rao, “Segmentation and analysis of retinal layers (ILM & RPE) in optical coherence tomography images with edema,” in 2014 IEEE Conference on Biomedical Engineering and Sciences (IECBES), pp. 204–209, Kuala Lumpur, Malaysia, 2014. View at: Publisher Site  Google Scholar
 G. R. Wilkins, O. M. Houghton, and A. L. Oldenburg, “Automated segmentation of intraretinal cystoid fluid in optical coherence tomography,” IEEE Transactions on Biomedical Engineering, vol. 59, no. 4, pp. 1109–1114, 2012. View at: Publisher Site  Google Scholar
Copyright
Copyright © 2017 Su Luo et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.