BioMed Research International

BioMed Research International / 2021 / Article
Special Issue

Application of Intelligence Methods in Biosciences

View this Special Issue

Research Article | Open Access

Volume 2021 |Article ID 5597222 | https://doi.org/10.1155/2021/5597222

Saeid Jafarzadeh Ghoushchi, Ramin Ranjbarzadeh, Amir Hussein Dadkhah, Yaghoub Pourasad, Malika Bendechache, "An Extended Approach to Predict Retinopathy in Diabetic Patients Using the Genetic Algorithm and Fuzzy C-Means", BioMed Research International, vol. 2021, Article ID 5597222, 13 pages, 2021. https://doi.org/10.1155/2021/5597222

An Extended Approach to Predict Retinopathy in Diabetic Patients Using the Genetic Algorithm and Fuzzy C-Means

Academic Editor: Alireza Baghban
Received02 Mar 2021
Accepted19 Jun 2021
Published28 Jun 2021

Abstract

The present study is developed a new approach using a computer diagnostic method to diagnosing diabetic diseases with the use of fluorescein images. In doing so, this study presented the growth region algorithm for the aim of diagnosing diabetes, considering the angiography images of the patients’ eyes. In addition, this study integrated two methods, including fuzzy C-means (FCM) and genetic algorithm (GA) to predict the retinopathy in diabetic patients from angiography images. The developed algorithm was applied to a total of 224 images of patients’ retinopathy eyes. As clearly confirmed by the obtained results, the GA-FCM method outperformed the hand method regarding the selection of initial points. The proposed method showed 0.78 sensitivity. The comparison of the fuzzy fitness function in GA with other techniques revealed that the approach introduced in this study is more applicable to the Jaccard index since it could offer the lowest Jaccard distance and, at the same time, the highest Jaccard values. The results of the analysis demonstrated that the proposed method was efficient and effective to predict the retinopathy in diabetic patients from angiography images.

1. Introduction

Patients suffering from diabetes, in some cases, lose their eyesight, which is often because of diabetic retinopathy (DR), though, there are some other causes of vision loss or poor eyesight, including other retinal and nonretina vision conditions such as macular degeneration and glaucoma (which are associated with age) and neuropathy vascular vision (which is Nonarteritic Anterior Ischemic Optic Neuropathy (NAION)) and cataract [14]. In addition, when an individual (a patient) with diabetes complains of visual disturbance, despite the 6.6’s precision, the refractive errors, contrast sensitivity, direct light, and compliance range need to be taken into account. The doctors managing diabetic patients should consider these vision problems in order to make sure that timely referral and treatments are done to control vision impairment as much as possible. This control significantly affects peoples’ daily life, particularly for activities like driving [5, 6].

Diabetic retinopathy can be considered a disease of the type of diabetes mellitus (DM), and it has been recognized as the primary basis for blindness in healthy adults in advanced countries. The retinal and choroidal circulation in the eye changes with the diabetic retinopathy. Multiple vascular dysfunctions may be the primary DR markers [79]. Also, Nonproliferative Diabetic Retinopathy (NPDR) can be defined by loss of capillary protein, thickening morphology of the glomerular basement membrane, and loss of smooth muscle cells, which leads to micro aneurysm, which is an essential indicator of the prediction of DR’s progress [10, 11].

Additionally, in the foveal avascular zone (FAZ) and capillary outflow in florfenicol angiography (FA) studies, patients with primary DR have been observed. Recently, OCT angiography has shown the presence of microenvironment and capillary excision in patients with diabetic retinopathy. Although DR is generally known as a retina, histology, electron microscopy, and green and endogenous angiography (ICGA), studies have shown that diabetes with degeneration of cholangiocarcinoma (CC) and also the formation of laminar deposition the term “diabetic cryotherapy “ is used to refer to changes in the CC associated with diabetes [5, 12, 13]. Imaging of retinal and choroidal ducts in diabetic patients does not occur regularly, especially in patients referring to primary care centers, since color-based angiographic techniques, such as FA and ICGA, require an intravenous injection of external dyes [14].

In addition, considering its location and structure, CC visualization is challenging with both FA and ICGA. OCTA is a relatively new imaging technique that allows for indirect visualization of the retina and CC pathways. CATA is based on the contrast of the motor and, contrary to the FA and the ICGA, does not need to inject different colors [14, 15]. The OCTA B-scan is produced by acquiring OCT B-scans repeated in a fast series of a retina. If the texture is constant, repeat scans B are identical; however, if there is a flow of blood, moving erythrocytes create pixels in OCT B-scans that are converted to the amount of current flow rate. The volume of OCTA can be generated by acquiring redundant B-scans in several retina networks. Since the depth of the OCT depth has been resolved, the various vascular layers of the retina can be individually visualized, which is not possible with color-based angiography [2, 4, 16]. In addition, since it is undeniable, the OCTA can be repeated several times in subsequent reviews or even during the same visit. Finally, since OCTA is the standard OCT imaging format, structural information is obtained at the same time and is inherently recorded by angiography information. This feature provides a simultaneous representation of both structural and angiographic data [2, 4, 16].

Diabetes types I and II increase more not only in the elderly but also in the people suffering from obesity and also the young adolescents. Diabetes is a disease of multiple systems; thus, each part of the eye can be vulnerable. Although eye complications have been reported frequently by doctors for over two centuries, DR has played a significant role in its destructive impacts that result in the early loss of sight [15, 16]. DR is common today in the world, and prevention is difficult for people. The ophthalmologists typically detect the DR severity degree by various visual tests of the fundus through directly examining and evaluating the color photographs. A challenge in this sense is that this process is both time-consuming and costly. The diagnosis of DR and diagnosis of primary diseases remain somewhat personal, with statistics of agreement between trained specialists in studies previously recorded being different [2].

Additionally, 75% of patients with DR condition are currently living in underdeveloped or poor regions, where adequate doctors, medical health care facilities, and diagnostic infrastructures are not available. To control the rate of the increasing number of the DR patients which can lead to blindness all over the globe, many global screening solutions have been set up to deal with the spread of disease inside the eye and better maintain vision; however, such programs make extensive use of DR to diagnose and treat retinopathy and related problems on an individual basis in an efficient way [4, 15]. As a result, millions across the world are currently exposed to visual impairment without recognizing adequate eye care. For the aim of overcoming the current problematic conditions, automated solutions have been proposed for the diagnosis of retinal disease using a dyed fundus illustration. In fact, those methods that are based on only some obtained samples from different patients in different times in one medical clinic or screening center are not applicable to fundus images [2, 16]. This is due to the fact that various kinds of fundus cameras are used in different clinics for detecting eye dilatation. In addition, many of these algorithms are employed to manually extract the attributes related to the DR using some hand-crafted features, aimed at describing the prediction of anatomical structures in the fundus, such as optical discs, blood vessels, or macula. Though these hand-crafted representations might be run on the Foundation Individual Dataset, they are once again trying to accurately detect DR through fundus images that are tailor-made for different demographic purposes [15, 17]. The general-purpose features, comprising GLCM (Gray Level Cooccurrence Matrix) [18], GLRM (Gray Level Run Length Matrix) [19], and Histogram of Oriented Gradients (HOG) Histogram of Oriented Gradients, have been examined using nonspecific methods applicable to specifying DR properties; however, they have shown weaker and disproportionate properties, which cannot describe the nuances of retinopathy.

Regular screening of diabetic patients for DR is a highly technical and critical aspect of patient care. The care and timing of this care are very important for the cost of the treatment. If timely diagnosed, DR compensatory treatments are reachable, and this is a pivotal process for anyone who needs them. The DR classification involves weighting many features and the position of these features. This process is an arduous task for doctors and needs a lot of time. After training, systems can acquire much more rapid classifications, and this will enable them to help clinicians in a quick classification [17, 20, 21].

DR can be described by morphological lesions associated with abnormalities in the retinal blood flow. These lesions represent a regional distribution that encompasses risk factors in the initial phases of the disease and can predict disease progression. Moreover, this disease encounters two different threatening symptoms of Diabetic Maculopathy (DM) vision and Proliferative Diabetic Retinopathy (PDR) in the late stages of the disease. These symptoms can be identified through considering the difference in spreading and category of lesions [4, 20]. The progression of diabetic retinopathy is also associated with sectional changes in the retinal blood flow and the adjustment of the diameters of the retinal capillaries in the macular region and in the retinal environment. The literature consists of some research into the segmentation of images in medical research such as the fuzzy clustering method with residual driving and automatic fuzzy clustering methods [22, 23]. Algorithms that are without supervision do not involve training networks; they are called clustering methods, e.g., Fast FCM, and Robust FCM [2429].

On the other hand, methods with supervision (e.g., ANN and recurrent convolution neural network) train an optimum network for detection and segmentation of medical images [3034]. Unsupervised methods usually are faster than training methods and are typically used as a preprocess for supervised methods. Methods such as FCM [21, 35], KNN [21, 36], and SVM [37] are commonly used as automated methods for generating ground truth data for automated methods. The advantage of automated methods like FCM is the detection of locations with distinguished color [29, 38]. In other words, tumors should be light or dark and different from other pixels. Gadekall et al. [39] suggested a deep neural network model based on a Grey Wolf Optimization (GWO) technique to classify the extracted features of diabetic retinopathy dataset. Employing the GWO can be beneficial to find optimal parameters for training the DNN structure. Behera and Chakravarty [40] utilized a scale-invariant feature transform (SIFT) approach to extract some key features. Also, to capture the exudate regions on each retinal images, Speeded Up Robust Features (SURF) has been employed.

In this study, the growth region technique based on the combination of the FCM and Genetic Optimizer Algorithm (GA) strategies for the aim of diagnosing diabetes is proposed, considering the angiography images of the patients’ eyes. The main novelty of FCM and GA is the optimization of automated methods of clustering images. Several previous works are providing automated methods for the extraction of ground truth images. However, some are not flexible and need further improvement. The methods presented in this paper without using hand methods exploited a genetic algorithm to find the best initial seeds of FCM method in the growth region algorithm.

2. Extended Growth Region Method

2.1. Data Collection

The present paper was designed to assess the dimensions and shape of FAZ in patients with diabetic retinopathy in comparison with healthy controls, using OCT angiography. The database contains 224 images of a fluorescence-sized angiography, depths of 8 bits per pixel, and each pixel size of micrometers taken from the eye of 14 times over a different period. Among them, 12 are diabetic, and two are nondiabetic, with an average of 16 images per person (192 images from diabetic and 32 images from nondiabetic eyes). Indeed, 12 of the images are obtained from the right eye and 2 from the left eye. The Heidelberg Specialist apparatus performed imaging in fluorescing angiography from Urmia Hospital. These images are in the jpeg format. Figure 1 displays an instance of a database. The Heidelberg device is a prototype device with a bandwidth of 50 nm, a lateral resolution of 14 μm, and an axial resolution of 7 μm that obtain 85,000 A-scans per second. This device produced by Heidelberg Engineering (Heidelberg, Germany) and employs an amplitude decorrelation technique for applying to a volume scan on a or zone including an area with the size of  mm or  mm.

2.2. Image Processing

In recent years, image processing has been extensively used, especially with the advent of advanced techniques such as discriminatory information processing, e.g., digital cameras and scanners. On the other hand, the images resulted from these techniques are generally associated with different degrees of noise, and even sometimes, these techniques fail to fade boundary inside an image [33, 4143]. This problem finally decreases the resultant image resolution. In this context, image processing refers to all operations and techniques adopted by users for the purpose of decreasing the defects and increasing the image quality.

The region’s growth method is an attempt for splitting the image into discrete regions on the basis of the degree of homogeneity or similarity between two or more parts of the image in neighboring pixels; thus, at another level, it depends on the criteria applied to homogeneity analysis [44, 45]. The pixels in each area are gathered together using some specific criteria: illumination, color, etc. This growing-base strategy is a simple method in the category of area-based methods and is based on the testing the intensity of an initial pixel with all touching pixels and adds them to the first pixel to search again for finding other pixels that belong to the segment. In the image segmentation context, the histogram-based methods are only focused upon the distribution of image pixels in the gray level, whereas the local growth strategies consider that the neighboring pixels also possess adjacent gray levels [44, 46].

In the following, the way the area-based methods work is explained step by step [47, 48]: (1)The number of initial seeds is the beginning of the algorithm(2)With the use of these seeds, the regions start their growth, and the pixels that resemble the original pixels will be added to that area(3)Once the growth of area stops, the subsequent grain is taken into consideration, and the next area growth continues(4)The above-mentioned steps will be continued until all of the pixels that exist in the image belong to one area

The area’s growth method has the following steps (Figure 2).

2.3. Selecting Initial Seeds

For the algorithm to get started, the initial seeds need to be manually entered. In this state, the algorithm begins operating by choosing the user’s initial points. Several methods in the field have the capacity of taking the start locations without having any prior knowledge about it. For instance, a random step strategy can be employed to explore the first points [47, 49].

For the aim of choosing the initial points, the present study proposes a new model using the combination of fuzzy clustering (FCM) and genetic optimization (GA) approaches. As in a FCM method every pixel inside the image can belong to more than a cluster, it helps to detect the border of any objects more accurately. Initially, a clustering process is carried out on the input image by applying the fuzzy clustering technique. This clustering method can be implemented based on the defining the membership grade and cluster centers . So, these parameters can be selected based on a trial and error method or by and optimization approach [29, 50]. To address the problem of experimenting all possible values for the parameters to obtain the maximum segmentation accuracy, the GA responsible for minimizing the cost function (sum of squares of the error) and determining the best value for these parameters. The cost function is demonstrated in Equation (1) [51, 52]: where the relation can be expressed as follows:

Moreover, and are computed employing Equation (3), as follows:

2.4. Determine the Similarity of Regions

After the determination of the initial points during the above step, the criterion of similarity is chosen for the regions. The objective of using this criterion is to examine all pixels around the new attached pixels to decide whether they can be added to the working area or not. This procedure specifies the attribution of the novel pixel to the corresponding area [53, 54].

A similarity criterion, which is widely used in this field, is the standard deviation that can be applied to each evaluating segment, meaning the new pixel needs to be added in the segment area if it can pass the following condition:

where stands for the standard deviation, demonstrates the mean, and illustrates a weighting parameter used for defining how many pixels are different within the region. In general, in case , more likely all pixels (about 99.70 percent) of the evaluating locations (pixels) are chosen in the same segmented area. It should be mentioned that whatever the value of is smaller, the chance of finding more similar pixels in a vicinity is lower. In another word, by defining the low and high values for the parameter , the number of the segmented area inside the image will be higher and lower, respectively. In addition, another criterion that needs to be considered is the image mean level (256 possible intensity levels). For doing this, the mean of the area which has been segmented by this technique is calculated, and then the intensity value of the new evaluating pixel for assessing based on the Equation (6) is extracted. This means that Equation (6) must be true for adding the new pixel in the segmented area before.

2.5. Growth Region

When the initial seeds are determined for the algorithm commencement and also the similarity criterion is determined for pixels with areas, then, the area growth process gets started. The area growth, which starts from the initial seeds, is done through choosing the neighboring regions [53, 54].

2.5.1. Fuzzy C-Means (FCM)

Grouping similar data or pixels in the same group (segmented data or pixels) using machine learning techniques based on some criteria can be defined as clustering [38, 5559]. This diving data can be implemented by many approaches such as DBSCAN, fuzzy C-means, mean shift, and K-means algorithms. In the fuzzy C-means and K-means approaches, first, a number of groups/classes need to be defined, but it is not true when working with the mean shift and DBSCAN techniques (algorithms calculate the optimum number of the clusters) [43, 60].

The FCM method is based on finding the similarity between data points inside the image to define the groups using an unsupervised manner. Unlike some approaches such as K-means and C-means, in this method, each point is belonging to some clusters based on a weighting parameter. The objective function that needs to be minimized using FCM strategy is demonstrated in Equation (7) [38]: where stands for the th sample of , refers to the centre of the th cluster, stands for the membership score for the th sample that relates to the th divided group of pixels (clusters), depicts the number of clusters, and is a number for demonstrating how many pixels present in image .

2.5.2. Genetic Algorithm (GA)

Object detection or object recognition can be conducted by extracting key features of the image related to each possible object in the image. It means each pixel inside the image based on adjacent pixels can extract a lot of local or global features even by changing the image representation such as Local Directional Pattern (LDP) [61, 62] or Local Binary Pattern (LBP) [41, 63]. So, by investigating these features, we can find the structure and shape of each object in the image. This investigating process can be conducted by an optimization process to decrease the time of the consideration [6467]. The genetic optimization algorithm was designed on the basis of the evolution theory and the survival of the fittest or natural selection proposed by Charles Darwin. This popular algorithm has been extensively applied to optimization problems [64]. In the genetic algorithm, at each implementation stage, random processing is performed on a group of search spots. It means a lot of random initial points are evaluated for finding the best possible points to apply in the next step. This way, each point is assigned with a sequence of traits, and then, the sequences are exposed to genetic operators. Next, based on minimizing the cost function (or increasing the fitness function), these obtained sequences are divided to separate parts and then each part adds to another part to create a new chain of spots within the searching space until stopping criteria are reached. The task of dividing and then adding sequences is done by considering their participation probability [6870]. In the present research, the cost function is computed based on the difference between the input image and the image attained by the region growing technique. It gets started from the initial random point, as can be expressed in the following relation: where WM, BM, and GM demonstrate the white matter, black matter, and gray matter, respectively, and , , and represent the images acquired from segmentation process.

2.5.3. Performance Analysis

The Specificity, Precision, Accuracy, False positive rate (FPR), Sensitivity, Relative volume difference (RVD), Volume overlap error (VOE), and Dice similarity (DICE) are the eight evaluation metrics that were employed to assess the result of the proposed model (can be calculated using Equations (11)–(18)). Sensitivity or true positive rate refers to the percentage of the important objects that identified correctly. Specificity refers to the percentage of the unimportant objects (healthy tissue) identified as unimportant correctly. Accuracy (measure of statistical bias) and precision (measure of statistical variability) represent the closeness of the measurements to a predefined value and to each other, respectively [38, 41, 71, 72]. (i)True Positive (TP): the important object (retinopathy) is recognized perfectly.(ii)True Negative (TN): the healthy tissue is recognized perfectly.(iii)False Positive (FP): the important object is recognized with mistakes.(iv)False Negative (FN): the healthy tissue is recognized with mistakes.where and represent the segmentation and ground-truth results, respectively.

In a given test, specificity and sensitivity are dependent upon the nature of the test and the type of test used. This is worth mentioning that a test result is not interpretable using Specificity and Sensitivity only. Receiver Operating Characteristic curve (ROC curve) is a probability curve that refers to a graphical plot (TPR or true positive rate against the FPR or False positive rate; FPR is on the -axis and TPR is on -axis) for checking any classification model’s performance at various thresholds settings.

3. Experimental Results and Discussions

As the region growing algorithm is based on finding the homogeneity defined by calculating the pixel intensity statistics, the first step in the region growing is to identify a set of seed positions (initial set of small areas). The initial region starts as the exact location of these seeds that are based on some user criterion or can be selected based on the generating random number. The regions are then grown from these seed points to other points in the vicinity of pixels depending on a region membership criterion (pixel intensity). To determinate whether the new point is good enough to join the selected seed location (point) or not, the mean and standard deviation of the growth region need to be computed, as depicted in Equations (16) and (17).

Here, we use 8-connected neighborhood for our pixel’s adjacent relationship to grow from the seed points. By examining the value of the pixels in the vicinity of the seed locations, all tested pixels are categorized into (1) seed points and (2) background. It is an iterated process until there are no changes in two successive iterative stages. The number and location of these initial location are determined by GA. By examining the random values for both the location and number of the seed points obtained by GA, the algorithm calculates cost function for each of them. Then, those locations that reach the best result can be selected for the final segmentation using region growing approach. This process is expressed in the following relations:

Afterwards, for the neighboring points that have been inserted to this class, they will also be found in the same neighborhood, and then the update of the parameters is done. The searching process is continued until the detection of the first class is completely done, and no other pixel is appended.

The numbers of clusters in FCM are 6-10, and for the genetic algorithm, the population is fixed at 20, and the mutation rate is set to 0.2. Figure 3 demonstrates the results of the clustering process performed with the use of the proposed FCM-GA method. As clearly shown, the yellow areas (some similar parts of the image) are decreasing inside the loop until reaching the best possible solution. Regarding Figure 3(a), the retina target region is almost above the right part of the image with high contrast. After classification using the FCM method, the results of the membership function are depicted as can be seen in Figure 3. The final image could detect the precise situation of the retina target region. The confusion matrix for testing the dataset is demonstrated in Figure 4.

The TPR (Sensitivity), ACC (accuracy), TNR (Specificity), PPV (Precision), and FPR (False positive rate) values using some approaches are described in Table 1. As it is clearly shown, our approach gains the best scores among all evaluated methods. The Deep membrane and Improved U-NET obtain the significant value in ACC criteria. The worst result (except FPR) was reached by the Artificial Neural Network (, , ). Besides, the PPV values of the Improved U-NET and Deep membrane are considerably higher as compared to the Ant colony, Pixel-based Segmentation, Morphological Watershed, Artificial Neural Network, and FCM results. For the TPR values, there is a small difference between Deep membrane and PCNN models. The FPR and TPR values of the Deep membrane model is partly similar to the proposed algorithm; however, Deep membrane approach has a meaning unlike TNR. Also, the Pixel-based Segmentation method represents the worst result in terms of the FPR. The Ant colony, Pixel-based Segmentation, Morphological Watershed, Artificial Neural Network, and FCM methods obtain the worst results in term of TNR, whilst their TPR values are acceptable.


MethodACCTPRTNRPPVFPR

Deep membrane [4]92.4%97.3%69.2%93.8%84.4%
Improved U-NET [6]92.4%96.8%70.3%94.3%81.3%
PCNN model [2]91%97.8%66.7%92.7%87.5%
Ant colony algorithm [73]84.3%93.9%46.7%87.5%65.6%
Pixel-based Segmentation [1]84.8%93.4%47.6%88.5%62.5%
Artificial Neural Network [74]83.4%93.8%44.7%86.5%65.6%
Morphological Watershed [75]85.2%94.4%48.9%88%68.8%
Presented FCM80.3%93%38.5%83.3%62.5%
Presented FCM+GA95%97.9%80%96.4%85.4%

Table 2 illustrates the evaluation of our segmentation technique and the results from recently published models. The attained values of RVD measurement imply the amount of oversegmentation or undersegmentation. Therefore, a zero score stands for the best possible segmentation result, whilst a negative value represents the segmented result image is not as large as that of the corresponding actual annotated image. DICE equals one for a precise segmentation. The Dice scores of the Deep membrane and PCNN models are the same and are partly similar to those of the proposed algorithm; however, these two techniques have a significantly different RVD. Also, the Deep membrane and Pixel-based Segmentation algorithms show the undersegmentation results with -2.46 and -5.64 values, respectively. Besides, the RVD values of the Artificial Neural Network and Morphological Watershed models were considerably higher as compared to our outcomes. Our strategy and Ant colony algorithm have significantly different RVDs [19]. Considering the VOE, the lowest and highest values belong to our model and Artificial Neural Network, respectively. Moreover, there are minimum differences between VOEs in Ant colony, Pixel-based Segmentation, and Artificial Neural Network techniques.


MethodDiceRVD (%)VOE (%)

Deep membrane [4]92%-2.466.16
Improved U-NET [6]90%3.745.78
PCNN model [2]92%3.556.41
Ant colony algorithm [73]88%-5.647.94
Pixel-based Segmentation [1]88%-4.917.42
Artificial Neural Network [74]89%5.378.12
Morphological Watershed [75]87%4.967.76
Presented FCM86%4.077.39
Presented FCM+GA94%2.324.28

Therefore, the division of an ideal abnormal area into angiography images plays a vital role in medical imaging programs. Segmentation algorithms are applied to promoting the precision and function in case of images in medical contexts. The method proposed in this paper was applied to a number of images, and also, the actual image analyses were compared to each other in order to examine the results of the reading. This study conducted comparative research on various performance criteria. Generally, in the GA classification, FCM returned more accurate results. Extending this work to improve GA segmentation with some other algorithms, without changing the primary nature of medical images, is required.

As the obtained results indicate obviously, the proposed method performed successfully in reducing the VOE score. It indicates the fact that the image has been split with higher precision. In the present research, the initial points were segregated automatically with the use of the region growth method. In addition, a genetic algorithm was employed for the purpose of selecting the most proper initial points in an automatic way. The GA approach was utilized, and the best possible fitness function was defined for the aim of image segmentation. This way, the proper initial locations were determined to get the algorithm started. At the final step, the images were exposed to the proposed algorithm, and the obtained outcomes were compared to those of the growth method of the area wherein the manually chosen points were chosen. Findings confirmed the capability of our technique in decreasing the fragmentation errors.

In the present study, a total of 224 images of the retina were employed, which included 192 images from diabetic and 32 images from nondiabetic eyes. We used 70%, 20%, and 10% of total data for training, validation, and test, respectively. The proposed method showed the highest level of sensitivity among all. That indicates that GA can be used more effectively in selecting initial points.

4. Conclusions

In the present paper, a segmentation approach using the growth region technique is represented. This model is based on the combination of the FCM and GA methods for the purpose of diagnosing diabetes from the angiography images of the patients’ eyes. The algorithm started with the early locations inside the image, and the mean of these locations were assumed as the mean of the objective area, and the early value for standard deviation was assumed as zero. For detecting the extract target area, the retinopathy images were recommended with the employed dynamic image analysis on the basis of the genetic algorithm. The eight evaluation metrics including Specificity, Precision, Accuracy, False positive rate (FPR), Sensitivity, Relative volume difference (RVD), Volume overlap error (VOE), and Dice similarity (DICE) were employed to assess the result of the proposed structure. The fuzzy fit function in GA was compared to another technique, and the outcomes indicated that the proposed method was more suitable to the sensitivity and Dice measures, with the maximum sensitivity and dice index. Considering the VOE index, the lowest and highest values are related to the suggested model and Artificial Neural Network, respectively. Additionally, there are minimum differences between VOEs in the Artificial Neural Network, Pixel-based Segmentation, and Ant colony techniques.

The fuzzy fitness function c-means evaluate the segmentation of threshold-based physical fuzzy tools, the fuzzy C-means tool in GA. The method proposed in the GA classification returned more proper results in its general performance, extending this to improve GA segmentation with some other algorithms, without changing the nature of the main medical angiography images is required. As the results demonstrated, the method introduced in this study was successful in decreasing the value of the VOE and RVD indexes. It indicates that the image has been split with higher precision.

Furthermore, a genetic optimizer was employed in this study in order to apply the proper early locations for the segmentation task automatically. The suitable early points were identified to start the algorithm with the use of GA and through giving a definition for the proper fitness function for the image segmentation purposes. At the final step, the prepared images were exposed to the proposed algorithm, and the obtained outcomes were compared with the growth manner of the region where the points are chosen manually. The obtained results confirmed the high capacity of our proposed algorithm in terms of reducing fragmentation errors.

Data Availability

The dataset is available online: http://www.med.harvard.edu/AANLIB/.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

References

  1. M. Al-Antary, M. Hassouna, Y. Arafa, and R. Khalifah, “Automated identification of diabetic retinopathy using pixel-based segmentation approach,” in ACM International Conference Proceeding Series, hlm, pp. 16–20, Association for Computing Machinery, New York, NY, USA, 2019. View at: Google Scholar
  2. T. J. Jebaseeli, C. A. Deva Durai, and J. D. Peter, “Retinal blood vessel segmentation from diabetic retinopathy images using tandem PCNN model and deep learning based SVM,” Optik, vol. 199, p. 163328, 2019. View at: Publisher Site | Google Scholar
  3. Y. Moharamzad, A. H. Davarpanah, A. Yaghobi Joybari et al., “Diagnostic performance of apparent diffusion coefficient (ADC) for differentiating endometrial carcinoma from benign lesions: a systematic review and meta-analysis,” Abdominal Radiology, vol. 46, no. 3, pp. 1115–1128, 2021. View at: Publisher Site | Google Scholar
  4. J. Xue, S. Yan, J. Qu et al., “Deep membrane systems for multitask segmentation in diabetic retinopathy,” Knowledge-Based Systems, vol. 183, p. 104887, 2019. View at: Publisher Site | Google Scholar
  5. T. J. Jebaseeli, C. A. D. Durai, and J. D. Peter, “Segmentation of retinal blood vessels from ophthalmologic diabetic retinopathy images,” Computers and Electrical Engineering, vol. 73, pp. 245–258, 2019. View at: Publisher Site | Google Scholar
  6. Q. Li, S. Fan, and C. Chen, “An intelligent segmentation and diagnosis method for diabetic retinopathy based on improved U-NET network,” Journal of Medical Systems, vol. 43, no. 9, pp. 1–9, 2019. View at: Publisher Site | Google Scholar
  7. M. He, W. Wang, H. Yu et al., “Comparison of expression profiling of circular RNAs in vitreous humour between diabetic retinopathy and non-diabetes mellitus patients,” Acta Diabetologica, vol. 57, no. 4, pp. 479–489, 2020. View at: Publisher Site | Google Scholar
  8. A. Khan, I. N. Petropoulos, G. Ponirakis, and R. A. Malik, “Visual complications in diabetes mellitus: beyond retinopathy,” Diabetic Medicine, vol. 34, no. 4, pp. 478–484, 2017. View at: Publisher Site | Google Scholar
  9. K. A. Ponto, J. Koenig, T. Peto et al., “Prevalence of diabetic retinopathy in screening-detected diabetes mellitus: results from the Gutenberg Health Study (GHS),” Diabetologia, vol. 59, no. 9, pp. 1913–1919, 2016. View at: Publisher Site | Google Scholar
  10. S. P. Mortensen, K. M. Winding, U. W. Iepsen et al., “The effect of two exercise modalities on skeletal muscle capillary ultrastructure in individuals with type 2 diabetes,” Scandinavian Journal of Medicine & Science in Sports, vol. 29, no. 3, pp. 360–368, 2019. View at: Publisher Site | Google Scholar
  11. L. Perrone, T. S. Devi, K. I. Hosoya, T. Terasaki, and L. P. Singh, “Thioredoxin interacting protein (TXNIP) induces inflammation through chromatin modification in retinal capillary endothelial cells under diabetic conditions,” Journal of Cellular Physiology, vol. 221, no. 1, pp. 262–272, 2009. View at: Publisher Site | Google Scholar
  12. A. Mirshahi, F. Ghassemi, K. Fadakar, R. Mirshahi, F. Bazvand, and H. Riazi-Esfahani, “Effects of panretinal photocoagulation on retinal vasculature and foveal avascular zone in diabetic retinopathy using optical coherence tomography angiography: a pilot study,” Journal of Current Ophthalmology, vol. 31, no. 3, pp. 287–291, 2019. View at: Publisher Site | Google Scholar
  13. M. Niestrata-Ortiz, P. Fichna, W. Stankiewicz, and M. Stopa, “Enlargement of the foveal avascular zone detected by optical coherence tomography angiography in diabetic children without diabetic retinopathy,” Graefe’s Archive for Clinical and Experimental Ophthalmology, vol. 257, no. 4, pp. 689–697, 2019. View at: Publisher Site | Google Scholar
  14. C. Yu, S. Xie, S. Niu et al., “Hyper-reflective foci segmentation in SD-OCT retinal images with diabetic retinopathy using deep convolutional neural networks,” Medical Physics, vol. 46, no. 10, pp. 4502–4519, 2019. View at: Publisher Site | Google Scholar
  15. X. Li, L. Shen, M. Shen, F. Tan, and C. S. Qiu, “Deep learning based early stage diabetic retinopathy detection using optical coherence tomography,” Neurocomputing, vol. 369, pp. 134–144, 2019. View at: Publisher Site | Google Scholar
  16. P. Brata Chanda and S. K. Sarkar, “Automatic identification of blood vessels, exaudates and abnormalities in retinal images for diabetic retinopathy analysis,” SSRN Electronic Journal, vol. 1, pp. 1–9, 2019. View at: Publisher Site | Google Scholar
  17. L. Ye, W. Zhu, S. Feng, and X. Chen, “GANet: group attention network for diabetic retinopathy image segmentation,” in Medical Imaging 2020: Image Processing, pp. 1–13, Texas, USA, 2020. View at: Google Scholar
  18. J. Rahebi and F. Hardalaç, “Retinal blood vessel segmentation with neural network by using gray-level co-occurrence matrix-based features patient facing systems,” Journal of Medical Systems, vol. 38, no. 8, pp. 1–12, 2014. View at: Publisher Site | Google Scholar
  19. B. K. Wardani, N. E. Belinda, R. Rulaningtyas, and E. Purwanti, “Application of gray level run length matrices features extraction for diabetic retinopathy detection based on artificial neural network,” in The 2nd International Conference On Physical Instrumentation And Advanced Materials 2019, Indonesia, 2020. View at: Google Scholar
  20. M. M. Khansari, J. Zhang, Y. Qiao et al., “Automated deformation-based analysis of 3D optical coherence tomography in diabetic retinopathy,” IEEE Transactions on Medical Imaging, vol. 39, no. 1, pp. 236–245, 2020. View at: Publisher Site | Google Scholar
  21. D. Palani, K. Venkatalakshmi, A. Reshma Jabeen, and V. M. Arun Bharath Ram, “Effective detection of diabetic retinopathy from human retinal fundus images using modified FCM and IWPSO,” in 2019 IEEE International Conference on System, Computation, Automation and Networking, ICSCAN, Institute of Electrical and Electronics Engineers Inc, India, 2019. View at: Google Scholar
  22. S. Dorosti, M. Fathi, S. J. Ghoushchi, M. Khakifirooz, and M. Khazaeili, “Patient waiting time management through fuzzy based failure mode and effect analysis,” Journal of Intelligent & Fuzzy Systems, vol. 38, no. 2, pp. 2069–2080, 2020. View at: Publisher Site | Google Scholar
  23. H. Torabi Dashti, A. Masoudi-Nejad, and F. Zare, “Finding exact and solo LTR-retrotransposons in biological sequences using SVM,” Iranian Journal of Chemistry and Chemical Engineering (IJCCE), vol. 31, no. 2, pp. 111–116, 2012. View at: Google Scholar
  24. Y. Ganjkhanlou, A. Bayandori Moghaddam, S. Hosseini, T. Nazari, A. Gazmeh, and J. Badraghi, “Application of image analysis in the characterization of electrospun nanofibers,” Iranian Journal of Chemistry and Chemical Engineering (IJCCE), vol. 33, no. 2, pp. 37–45, 2014. View at: Google Scholar
  25. S. Jafarzadeh Ghoushchi, M. Khazaeili, A. Amini, and E. Osgooei, “Multi-criteria sustainable supplier selection using piecewise linear value function and fuzzy best-worst method,” Journal of Intelligent and Fuzzy Systems, vol. 37, no. 2, pp. 2309–2325, 2019. View at: Publisher Site | Google Scholar
  26. E. Kamari, A. A. Hajizadeh, and M. R. Kamali, “Experimental investigation and estimation of light hydrocarbons gas-liquid equilibrium ratio in gas condensate reservoirs through artificial neural networks,” Iranian journal of chemistry and chemical engineering, vol. 39, no. 6, pp. 163–172, 2020. View at: Google Scholar
  27. B. Kavitha and D. Sarala Thambavani, “Artificial neural network optimization of adsorption parameters for Cr(VI), Ni(II) and Cu(II) ions removal from aqueous solutions by riverbed sand,” Iranian journal of chemistry and chemical engineering, vol. 39, no. 5, pp. 203–223, 2020. View at: Google Scholar
  28. H. A. Purwanithami, C. Atika Sari, E. H. Rachmawanto, and D. R. I. M. Setiadi, “Hemorrhage diabetic retinopathy detection based on fundus image using neural network and FCM segmentation,” in 2020 International Seminar on Application for Technology of Information and Communication (iSemantic), pp. 45–49, Institute of Electrical and Electronics Engineers Inc. Indonesia, 2020. View at: Google Scholar
  29. W. Wiharto and E. Suryani, “The analysis effect of cluster numbers on fuzzy c-means algorithm for blood vessel segmentation of retinal fundus image,” in 2019 International Conference on Information and Communications Technology, ICOIACT, pp. 106–110, Institute of Electrical and Electronics Engineers Inc, Indonesia, 2019. View at: Google Scholar
  30. A. Azari, M. Shariaty-Niassar, and M. Alborzi, “Short-term and medium-term gas demand load forecasting by neural networks,” Iranian Journal of Chemistry and Chemical Engineering (IJCCE), vol. 31, no. 4, pp. 77–84, 2012. View at: Google Scholar
  31. Z. Hosseini-Dastgerdi and S. Jafarzadeh-Ghoushchi, “Investigation of asphaltene precipitation using response surface methodology combined with artificial neural network,” Journal of Chemical and Petroleum Engineering, vol. 2019, no. 2, pp. 153–167, 2019. View at: Google Scholar
  32. N. M. Ramli, M. A. Hussain, B. M. Jan, and B. Abdullah, “Online composition prediction of a debutanizer column using artificial neural network,” Iranian Journal of Chemistry and Chemical Engineering (IJCCE), vol. 36, no. 2, pp. 153–174, 2017. View at: Google Scholar
  33. R. Ranjbarzadeh, A. Bagherian Kasgari, S. Jafarzadeh Ghoushchi, S. Anari, M. Naseri, and M. Bendechache, “Brain tumor segmentation based on deep learning and an attention mechanism using MRI multi-modalities brain images,” Scientific Reports, vol. 11, no. 1, p. 10930, 2021. View at: Publisher Site | Google Scholar
  34. S. Salem and S. Jafarzadeh-Ghoushchi, “Estimation of optimal physico-chemical characteristics of nano-sized inorganic blue pigment by combined artificial neural network and response surface methodology,” Chemometrics and Intelligent Laboratory Systems, vol. 159, pp. 80–88, 2016. View at: Publisher Site | Google Scholar
  35. S. Anbazhagan, V. Thiruvengadam, and K. Kulanthai, “Adaptive neuro-fuzzy inference system and artificial neural network modeling for the adsorption of methylene blue by novel adsorbent in a fixed-bed column method,” Iranian journal of chemistry and chemical engineering, vol. 39, no. 6, pp. 75–93, 2020. View at: Google Scholar
  36. S. Bandyopadhyay, S. Choudhury, S. K. Latib, D. K. Kole, and C. Giri, “Gradation of diabetic retinopathy using KNN classifier by morphological segmentation of retinal vessels,” in Advances in Intelligent Systems and Computing, hlm, pp. 189–198, Springer, Verlag, 2018. View at: Google Scholar
  37. T. Vandarkuzhali and C. S. Ravichandran, “Detection of fovea region in retinal images using optimisation-based modified FCM and ARMD disease classification with SVM,” International Journal of Biomedical Engineering and Technology, vol. 32, no. 1, pp. 83–107, 2020. View at: Publisher Site | Google Scholar
  38. R. Ranjbarzadeh and S. B. Saadi, “Automated liver and tumor segmentation based on concave and convex points using fuzzy c-means and mean shift clustering,” Measurement, vol. 150, p. 107086, 2020. View at: Publisher Site | Google Scholar
  39. T. R. Gadekallu, N. Khare, S. Bhattacharya, S. Singh, P. K. R. Maddikunta, and G. Srivastava, “Deep neural networks to predict diabetic retinopathy,” Journal of Ambient Intelligence and Humanized Computing, vol. 1, pp. 1–14, 2020. View at: Publisher Site | Google Scholar
  40. M. K. Behera and S. Chakravarty, “Diabetic Retinopathy Image Classification Using Support Vector Machine,” in 2020 International Conference on Computer Science, Engineering and Applications, ICCSEA 2020, hlm, Institute of Electrical and Electronics Engineers Inc, India, 2020. View at: Google Scholar
  41. N. Karimi, R. Ranjbarzadeh Kondrood, and T. Alizadeh, “An intelligent system for quality measurement of golden bleached raisins using two comparative machine learning algorithms,” Measurement: Journal of the International Measurement Confederation, vol. 107, pp. 68–76, 2017. View at: Publisher Site | Google Scholar
  42. Y. Pourasad, R. Ranjbarzadeh, and A. Mardani, “A new algorithm for digital image encryption based on chaos theory,” Entropy, vol. 23, no. 3, p. 341, 2021. View at: Publisher Site | Google Scholar
  43. R. Ranjbarzadeh, S. Baseri Saadi, and A. Amirabadi, “LNPSS: SAR image despeckling based on local and non-local features using patch shape selection and edges linking,” Measurement, vol. 164, p. 107989, 2020. View at: Publisher Site | Google Scholar
  44. A. F. A. Fadzil, S. Ibrahim, and N. E. A. Khalid, “Blood vessels segmentation of retinal fundus image via wStack-based object-oriented region growing,” pp. 31–37, Springer. View at: Google Scholar
  45. A. Hamzenejad, S. J. Ghoushchi, V. Baradaran, and A. Mardani, “A robust algorithm for classification and diagnosis of brain disease using local linear approximation and generalized autoregressive conditional heteroscedasticity model,” Mathematics, vol. 8, no. 8, p. 1268, 2020. View at: Publisher Site | Google Scholar
  46. M. V. Maheswari and G. Murugeswari, “A survey on computer algorithms for retinal image preprocessing and vessel segmentation,” in 2020 Proceedings of the 5th International Conference on Inventive Computation Technologies, ICICT, hlm, pp. 403–408, Institute of Electrical and Electronics Engineers Inc, India, 2020. View at: Google Scholar
  47. R. Panda, N. B. Puhan, and G. Panda, “New binary Hausdorff symmetry measure based seeded region growing for retinal vessel segmentation,” Biocybernetics and Biomedical Engineering, vol. 36, no. 1, pp. 119–129, 2016. View at: Publisher Site | Google Scholar
  48. E. Rodrigues, A. Conci, and P. Liatsis, “ELEMENT: multi-modal retinal vessel segmentation based on a coupled region growing and machine learning approach,” IEEE Journal of Biomedical and Health Informatics, vol. 24, no. 12, pp. 3507–3519, 2020. View at: Publisher Site | Google Scholar
  49. N. D. Salih, M. D. Saleh, C. Eswaran, and J. Abdullah, “Fast optic disc segmentation using FFT-based template-matching and region-growing techniques,” Computer Methods in Biomechanics and Biomedical Engineering: Imaging and Visualization, vol. 6, no. 1, pp. 101–112, 2018. View at: Publisher Site | Google Scholar
  50. R. Ranjbarzadeh and S. Baseri Saadi, “Corrigendum to “Automated liver and tumor segmentation based on concave and convex points using fuzzy c-means and mean shift clustering” [Measurement 150 (2020) 107086],” Measurement, vol. 151, p. 107230, 2020. View at: Publisher Site | Google Scholar
  51. H. Huang, F. Meng, S. Zhou, F. Jiang, and G. Manogaran, “Brain image segmentation based on FCM clustering algorithm and rough set,” IEEE Access, vol. 7, pp. 12386–12396, 2019. View at: Publisher Site | Google Scholar
  52. K. Soppari and N. S. Chandra, “Development of improved whale optimization-based FCM clustering for image watermarking,” Computer Science Review, vol. 37, p. 100287, 2020. View at: Publisher Site | Google Scholar
  53. N. Jothiaruna, K. Joseph Abraham Sundar, and B. Karthikeyan, “A segmentation method for disease spot images incorporating chrominance in comprehensive color feature and region growing,” Computers and Electronics in Agriculture, vol. 165, p. 104934, 2019. View at: Publisher Site | Google Scholar
  54. D. Q. Zeebaree, H. Haron, A. M. Abdulazeez, and D. A. Zebari, “Machine learning and region growing for breast cancer segmentation,” in 2019 International Conference on Advanced Science and Engineering, ICOASE, pp. 88–93, Institute of Electrical and Electronics Engineers Inc. Iraq, 2019. View at: Google Scholar
  55. M. Bendechache and M. T. Kechadi, “Distributed clustering algorithm for spatial data mining,” in ICSDM 2015 - Proceedings 2015 2nd IEEE International Conference on Spatial Data Mining and Geographical Knowledge Services, hlm, pp. 60–65, Institute of Electrical and Electronics Engineers Inc, China, 2015. View at: Google Scholar
  56. M. Bendechache, M. T. Kechadi, and N. A. Le-Khac, “Efficient large scale clustering based on data partitioning,” in Proceedings -3rd IEEE International Conference on Data Science and Advanced Analytics, DSAA 2016, hlm, pp. 612–621, Institute of Electrical and Electronics Engineers Inc, Canada, 2016. View at: Google Scholar
  57. M. Bendechache, A. K. Tari, and M. T. Kechadi, “Parallel and distributed clustering framework for big spatial data mining,” International Journal of Parallel, Emergent and Distributed Systems, vol. 34, no. 6, pp. 671–689, 2019. View at: Publisher Site | Google Scholar
  58. R. Geetha Ramani and L. Balasubramanian, “Macula segmentation and fovea localization employing image processing and heuristic based clustering for automated retinal screening,” Computer Methods and Programs in Biomedicine, vol. 160, pp. 153–163, 2018. View at: Publisher Site | Google Scholar
  59. H. Xia, S. Deng, M. Li, and F. Jiang, “Robust retinal vessel segmentation via clustering-based patch mapping functions,” in 2016 IEEE International Conference on Bioinformatics and Biomedicine, BIBM, pp. 520–523, Institute of Electrical and Electronics Engineers Inc. China, 2017. View at: Google Scholar
  60. M. Bendechache, Study of Distributed Dynamic Clustering Framework for Spatial Data Mining, Theses and Dissertations of University College Dublin, Irland, 2019, http://oatd.org/oatd/record?record=handle%5C%3A10197%5C%2F10614.
  61. R. Srinivasa Perumal and P. V. S. S. R. Chandra Mouli, “Dimensionality reduced local directional pattern (DR-LDP) for face recognition,” Expert Systems with Applications, vol. 63, pp. 66–73, 2016. View at: Publisher Site | Google Scholar
  62. P. P. Sarangi, B. S. P. Mishra, and S. Dehuri, “Fusion of PHOG and LDP local descriptors for kernel-based ear biometric recognition,” Multimedia Tools and Applications, vol. 78, no. 8, pp. 9595–9623, 2019. View at: Publisher Site | Google Scholar
  63. T. Tuncer, S. Dogan, and F. Ozyurt, “An automated residual exemplar local binary pattern and iterative relief based COVID-19 detection method using chest X-ray image,” Chemometrics and Intelligent Laboratory Systems, vol. 203, p. 104054, 2020. View at: Publisher Site | Google Scholar
  64. H. Liu, R. Zhai, J. Fu, Y. Wang, and Y. Yang, “Optimization study of thermal-storage PV-CSP integrated system based on GA-PSO algorithm,” Solar Energy, vol. 184, pp. 391–409, 2019. View at: Publisher Site | Google Scholar
  65. H. Moayedi, M. Raftari, A. Sharifi, W. A. W. Jusoh, and A. S. A. Rashid, “Optimization of ANFIS with GA and PSO estimating α ratio in driven piles,” Engineering with Computers, vol. 36, no. 1, pp. 227–238, 2020. View at: Publisher Site | Google Scholar
  66. M. Shirmohammadi, S. J. Goushchi, and P. M. Keshtiban, “Optimization of 3D printing process parameters to minimize surface roughness with hybrid artificial neural network model and particle swarm algorithm,” Progress in Additive Manufacturing, vol. 6, no. 2, pp. 199–215, 2021. View at: Publisher Site | Google Scholar
  67. S. Zandevakili, M. R. Akhondi, R. Hosseini, and S. Mohammad, “Leaching optimization of Sarcheshmeh copper concentrate by application of Taguchi experimental design method,” Iranian journal of chemistry and chemical engineering, vol. 39, no. 6, pp. 229–236, 2020. View at: Google Scholar
  68. A. Khan, Z. u. Rehman, M. A. Jaffar et al., “Color image segmentation using genetic algorithm with aggregation-based clustering validity index (CVI),” Signal, Image and Video Processing, vol. 13, no. 5, pp. 833–841, 2019. View at: Publisher Site | Google Scholar
  69. L. Khrissi, N. El Akkad, H. Satori, and K. Satori, “Image segmentation based on K-means and genetic algorithms,” in Advances in Intelligent Systems and Computing, hlm, pp. 489–497, Springer, Singapore, 2020. View at: Google Scholar
  70. F. Torres, B. Escalante-Ramirez, J. Olveres, and P. L. Yen, “Lesion detection in breast ultrasound images using a machine learning approach and genetic optimization,” in Pattern Recognition and Image Analysis, pp. 289–301, Springer, Cham, 2019. View at: Google Scholar
  71. R. Gholipour Peyvandi and S. Z. Islami Rad, “Precise prediction of interface distribution of materials in multiphase separation facilities using a low-cost and simple technique: ANN,” Iranian journal of chemistry and chemical engineering, vol. 39, no. 5, pp. 285–291, 2020. View at: Google Scholar
  72. H. Mirshahvalad, R. Ghasemiasl, N. Raufi, and M. Malekzadeh Dirin, “A neural networks model for accurate prediction of the flash point of chemical compounds,” Iranian journal of chemistry and chemical engineering, vol. 39, no. 4, pp. 297–304, 2020. View at: Google Scholar
  73. T. SELÇUK and A. ALKAN, “Detection of microaneurysms using ant colony algorithm in the early diagnosis of diabetic retinopathy,” Medical Hypotheses, vol. 129, p. 109242, 2019. View at: Publisher Site | Google Scholar
  74. A. T. Nair and K. Muthuvel, “Blood vessel segmentation and diabetic retinopathy recognition: an intelligent approach,” Computer Methods in Biomechanics and Biomedical Engineering: Imaging and Visualization, vol. 8, no. 2, pp. 169–181, 2020. View at: Publisher Site | Google Scholar
  75. P. Siva Kalyani and G. Sasikala, “Morphological watershed approach for the analysis of diabetic nephropathy,” in Lecture Notes in Electrical Engineering, pp. 547–554, Springer, Singapore, 2021. View at: Publisher Site | Google Scholar

Copyright © 2021 Saeid Jafarzadeh Ghoushchi et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Related articles

No related content is available yet for this article.
 PDF Download Citation Citation
 Download other formatsMore
 Order printed copiesOrder
Views459
Downloads791
Citations

Related articles

No related content is available yet for this article.

Article of the Year Award: Outstanding research contributions of 2021, as selected by our Chief Editors. Read the winning articles.