Journal of Healthcare Engineering

Journal of Healthcare Engineering / 2021 / Article
Special Issue

Multiple Criteria Decision-Making Approaches for Healthcare Management Applications

View this Special Issue

Research Article | Open Access

Volume 2021 |Article ID 5549779 | https://doi.org/10.1155/2021/5549779

Jianfeng Cui, Xiaoyun Zhang, Feibing Xiong, Chin-Ling Chen, "Pathological Myopia Image Recognition Strategy Based on Data Augmentation and Model Fusion", Journal of Healthcare Engineering, vol. 2021, Article ID 5549779, 15 pages, 2021. https://doi.org/10.1155/2021/5549779

Pathological Myopia Image Recognition Strategy Based on Data Augmentation and Model Fusion

Academic Editor: Hao-Chun Lu
Received24 Feb 2021
Revised02 Apr 2021
Accepted27 Apr 2021
Published06 May 2021

Abstract

The automatic diagnosis of various retinal diseases based on fundus images is important in supporting clinical decision-making. Convolutional neural networks (CNNs) have achieved remarkable results in such tasks. However, their high expression ability possibly leads to overfitting. Therefore, data augmentation (DA) techniques have been proposed to prevent overfitting while enriching datasets. Recent CNN architectures with more parameters render traditional DA techniques insufficient. In this study, we proposed a new DA strategy based on multimodal fusion (DAMF) which could integrate the standard DA method, data disrupting method, data mixing method, and autoadjustment method to enhance the image data in the training dataset to create new training images. In addition, we fused the results of the classifier by voting on the basis of DAMF, which further improved the generalization ability of the model. The experimental results showed that the optimal DA mode could be matched to the image dataset through our DA strategy. We evaluated DAMF on the iChallenge-PM dataset. At last, we compared training results between 12 DAMF processed datasets and the original training dataset. Compared with the original dataset, the optimal DAMF achieved an accuracy increase of 2.85% on iChallenge-PM.

1. Introduction

Pathologic myopia (PM) is one of the major causes of visual impairment worldwide [13]. As myopia deepens, it is often accompanied by deforming changes in the posterior pole of the eye [3]. These changes are usually manifested as sclerotic atrophy, choroidal atrophy, and growth of the ocular axis, which may be associated with various complications of the eye, such as amblyopia, glaucoma, cataracts, vitreous clouding, and retinal detachment [1]. Complications of pathological myopia are considered to be the main reasons for visual impairment and blindness today, particularly in East Asia [4, 5]. Pathological myopia causes visual impairment due to various pathologies of the macula, peripheral retina, and optic nerve. Structural deformities of the eye, including posterior sclerite uveitis, may accelerate the progression of these diseases [1, 3, 69].

According to a summary of 145 studies regarding the global prevalence of myopia and PM, there are approximately 1950 million people with myopia (accounting for 28.3% of the global population) and 277 million people with PM (accounting for 4.0% of the global population), and these numbers are predicted to increase to 4758 million (accounting for 49.8% of the global population) for myopia and 938 million (accounting for 9.8% of the global population) for PM by 2050 [5].

The prevalence of high myopia and pathological myopia shows an increasing yearly trend due to the changes in environmental factors and lifestyle [1014]. In China, the number of ophthalmologists differs significantly from that in developed countries, but the demand for ophthalmologists in China is already huge; with the growth of social aging, the number of ophthalmologists has been rising in recent years [10, 12, 14]. In consequence, the visual impairment caused by pathological myopia complications will become increasingly serious in the coming decades. The people’s growing medical needs contrast sharply with the increasingly deficient medical resources in the current medical field.

On the one hand, the problem of “difficult and expensive access to medical care” still exists, which is mainly caused by the severe scarcity of talented physicians and the long training period for doctors. On the other hand, with the health problems gradually worsening, people are paying more and more attention to their health, which aggravates the demand for medical services. This is a social problem to which AI technology can offer the medical industry a solution [1518].

Over the past two decades, with the development of imaging techniques, such as optical coherence tomography, frequency-domain OCT, and 3D magnetic resonance imaging, the complications associated with high myopia have been well known [16, 1923]. For example, the optic nerve, macula, and neoplastic lesions can be magnified by OCT to extraordinary resolution for evaluation. In addition, myopic retractile macular lesions and domed macular lesions can be utilized in the same way. The advent of new therapies, including antineovascular drug therapies and vasectomy, has led to an improved prognosis for some of the complications associated with high myopia.

Medical artificial intelligence (AI) technologies have been well developed in recent years [20]. A case in point is the application of machine learning-based AI technology to ophthalmology [20, 21]. The diagnosis of many eye diseases relies heavily on the results of ophthalmic examinations, most of which are built on imaging studies. Eye images are delicate, complex, and informative, and diagnostic results are heavily dependent on the doctor’s knowledge and clinical experience, which makes diagnosis subjective and time-consuming [24, 25]. The development of medical AI has significantly improved the efficiency of ophthalmic disease diagnosis in clinical work and reduced the burden on ophthalmologists [2022].

CNNs, which are data-driven and can automatically extract relevant features, have secured better results in image recognition than traditional methods [26]. Therefore, it is considered to be a new choice to introduce CNN techniques into medical image processing. Lately, many studies have been conducted on this matter and applied CNNs to fundus image recognition, whose results generally surpass those based on traditional recognition ways [19, 2631].

Although the CNN-based fundus image recognition method outperforms traditional methods to a certain extent, there are still some problems. For example, the amount of medical image data is large, while the number of positive samples is small. Nevertheless, a common fact is that the training process of the model mainly depends on the dataset, namely, the training effect of the model can be well improved by the dataset after DA processing [32].

Furthermore, with the continuous development of deep learning in the field of images, CNNs have been increasingly complicated. Each model has its unique advantages, but it is not guaranteed that every aspect of the model will perform well. For instance, the model’s lack of expressive ability will lead to the weakness in the recognition of some rare lesion images [32]. To address this problem, researchers have proposed optimized neural network models from different perspectives and achieved effective results. However, there are a lot of difficulties in enhancing the existing models. For example, when the researchers optimize the models (such as widening and deepening models), they cannot predict the effectiveness of the models but just observe whether the optimization operation improves the performance of the original models through the training results. Besides, even if such optimization is effective, it may be computational and time-intensive or needs a long development cycle, so it cannot address the problem effectively [3335].

In order to deal with these issues, the main contributions of this paper are described as follows:(1)12 DAMFs will be designed based on the iChallenge-PM dataset. To our knowledge, these DA methods cover all the operations used in the current DA. The purpose is to increase data characteristics, suppress sample imbalance, and effectively improve the quality of datasets.(2)Based on AlexNet, VGG-16, GoogLeNet, and ResNet-50 models, different optimizers, loss functions, and learning rates are constructed. Then, the model with the highest accuracy will be adopted as the primary learner and trained based on 12 datasets. The generalization ability of the model will be enhanced through this method.(3)The abovementioned primary learner prediction will be used as a new input and put into the secondary learner, namely, the hard voting model, and then the fusion model is trained to form the final model.(4)The model optimized by the above operations achieves high accuracy without transfer learning. More importantly, by using the augmented dataset and the model fusion method, we effectively avoid overfitting and improve the generalization ability of the model when processing various data, which further improves the expressive ability of the model. As a result, the accuracy of the model in recognizing complex and rare case images will be effectively improved.

2. Literature Review

In this section, we mainly reviewed the related literature on automatic disease diagnosis from fundus photography and DA.

2.1. Automatic Disease Diagnosis from Fundus Photography

Most of the conventional ophthalmic diseases can be examined from fundus photography, including PM, age-related macular degeneration (AMD), diabetic retinopathy (DR), and glaucoma. Conventional diagnosis methods tremendously depend on doctors’ professional experience and knowledge, which results in a high misdiagnosis rate and a huge waste of medical data [36, 37]. The deep integration of ophthalmology and AI offers potential for revolutionizing current disease diagnosis patterns and generating a significant clinical impact. As for PM classification, Freire et al. employed Xception as the baseline architecture with ImageNet pretrain weights to diagnose PM from fundus images [38]. Zhang et al. used the feature selection of PM valuable information in the images to improve the training effect of the model [31]. They demonstrated that the new method was much efficient by using less than 25% of the initial candidate feature set. Not only in the field of image classification but DA also occupies a highly important position in the field of target detection. Sun et al. put forward two new DA modules consisting of channel-wise random Gamma correction and channel-wise random vessel augmentation [27]. They argued that their method could ameliorate the performance and robustness of a classic CNN architecture. However, most medical imaging samples are unbalanced. Although there may be a large number of samples, the types of samples are limited. Different from previous works, in this paper, we focused on the first stage of deep learning, which was DA. And by means of this strategy, the defect of data imbalance could be effectively solved and the overfitting in the training process could be suppressed.

2.2. Data Augmentation

DA highlights the characteristics of image data and prevents overfitting in the training effect [33]. In existing studies, researchers divide DA methods into the standard method, data disrupting method, data mixing method, DA method based on reinforcement learning, and fusion-based image augmentation method.(1)Standard Data Augmentation Method. AlexNet [39] preprocesses the data using random cropping and horizontal flipping, and it has been verified on CIFAR-10. Random cropping prevents CNN from overfitting specific features by changing the obvious features in the image. Facebook artificial intelligence research uses another color translation method, called color dithering, to improve the process of ResNet [40]. Color dithering randomly changes the brightness, contrast, and saturation of the image. These DA techniques play an important role in the training of a model. Fu et al. performed extensive experimentation on a large set of images with varying illuminations [41]. The performance is analyzed both quantitatively and qualitatively. However, as the number of parameters increases, the risk of overfitting also increases. Many studies have proposed more complex CNNs. Therefore, more powerful DA strategies are particularly important.(2)Data Disrupting Method. [42] Contrary to (1), a data disrupting method produces unnatural images by destroying images’ features. Dropout on the input layer is a DA technique that disturbs and masks the original information of given data by dropping pixels. Pixel dropping functions as injection of noise into an image. It makes the CNN robust to noisy images and contributes to generalization rather than enriching the dataset, randomly erasing a region in an image at every training step. It is an extension of dropout, where the masking of regions behaves like injected noise and makes CNNs robust to noisy images. Under this condition, CNNs need to learn other parts that are usually ignored.(3)Data Mixing Method. This is a special case of (2), where a mixup alpha blends two images to construct a new training image [43]. Mixup can train CNNs on convex combinations of pairs of training samples and their labels and enables CNNs to favor a simple linear behavior in-between training samples. This behavior makes the prediction confidence transit linearly from one class to another class, thus providing smoother estimation and margin maximization. Therefore, the mixup makes CNNs robust to adversarial examples and stabilizes the training of generative adversarial networks.(4)Data Augmentation Method Based on Reinforcement Learning [44]. Autoenhancement is a framework based on reinforcement learning to explore the optimal augmentation combination [45]. Hence, it is not a DA method but is an external framework. It achieves significant results during the CIFAR-10 classification and proves the contribution of reinforcement learning to DA research.(5)Fusion-Based Image Augmentation Method. Fusion is preferred to the direct application of traditional techniques since it involves an amalgamation of traditional techniques, rather than the application of a single technique. Fusion can be done in various ways, out of which multiscale fusion has proved to be one of the best. Parihar et al. presented a detailed analysis of image enhancement techniques based on multifusion, thereby giving an insight into the algorithm used in each method, along with its implementation framework [36, 42].

In this paper, we adopted the concept of (5) and integrated the image enhancement techniques of (1), (2), and (3), aiming to enlarge the features of images from the perspective of color, direction, shape, and so on.

3. Materials and Methods

In this session, we first introduced the dataset of iChallenge-PM [46] composed of 1200 annotated retinal images. In another part of this section, we introduced the primary learner, which included some classical convolutional neural networks and corresponding components. These frameworks have become general arrangements for image classification and have been extensively utilized in different computer vision tasks.

3.1. iChallenge-PM

The iChallenge-PM dataset contained 1200 annotated color fundus photos with non-PM (50%) and PM (50%) cases. Specifically speaking, the dataset contained 400 images in the training set, validation set, and test set, respectively, while the test set was not public. We reclassified the 800 public image data and divided the validation set into a new validation set and a test set according to the ratio of 50% : 50%. All images data were reshaped to a size of 224  224 before using DA. Under the partitioning of iChallenge-PM, the reference standard of PM presence was obtained from the health records, which was not based solely on fundus image but also took OCT, visual test, and other factors into consideration. For the training data, PM, HM, and normal labels were reflected in the image file names, with 0 denoting normal cases while 1 denoting abnormal cases. However, we did not utilize any human-annotated labels during network training. To evaluate the effectiveness of our method, we employed accuracy and loss rate as the evaluation indices. Accuracy was often the most important index for doctors and patients.

3.2. Classifiers and Components

(1)AlexNet. The AlexNet network proposed by Krizhevsky et al. was the first to use five convolutional layers and three fully connected expenses to achieve the classification of 1000 classes of images, thus becoming the seminal breakthrough in image classification based on deep learning. Compared with other traditional convolutional neural networks, AlexNet applied various methods to improve deep convolutional networks. For example, the rectified linear unit (ReLU) nonlinear activation function was used to speed up the training of the network, multi-GPU convolutional operations were implemented to address the limitations of insufficient graphic card resources at the time, and the DropOut random inactivation strategy was introduced to reduce overfitting at the full connection layer. Furthermore, strategies such as local response normalization, overlap pooling, as well as augmentation, were proposed by AlexNet to improve the classification and generalization capabilities of the model.(2)VGG. The VGG network was proposed by Simonyan et al., in which filters of 5  5 and 7  7 were replaced by filters of 3  3. It was a basic idea embodied by VGG that the receptive fields of multiple small convolutional layers in series could be in the same size as that of a large convolutional layer. For example, two 3  3 convolutional fields in sequences had the same field size as one 5  5 convolutional field, and their convolutional effects were the very same. However, multiple small convolutional layers concatenated together had fewer parameters and more nonlinear transformations. This was most effective in better learning the features. At the same time, VGG also increased the network structure to 16 or 19 layers. As the number of layers increased, the network enjoyed better feature representation and better model classification. Being simple and effective, VGG was still commonly used in the field of computer vision for image classification, detection, segmentation, super-resolution, and image styling.(3)GoogLeNet. Google made a significant contribution to the development of deep convolutional neural networks through the proposed inception family. The most significant contribution of Inception-V1 (GoogLeNet) was to propose the inception structure while deepening the depth of the convolutional neural network. The structure increased the width of the network by concatenating multiple convolutional blocks of different sizes, which allowed the convolutional blocks to acquire information from discrete receptive fields. In addition, the structure took full advantage of the 1  1 convolutional code to reduce a large number of network parameters, thereby improving the efficiency of computing resources. Inception-V2 proposed an excellent regularization method, namely, batch normalization, which made the data undergo a back normalization process before each convolution and was now the standard for deep convolutional networks. This approach was an excellent solution to the training problem of multilayered networks. The evolution of the inception structures is shown in Figure 1.(4)ResNet. Kaiming He et al. established a deep residual network, namely, ResNet [14], which increased the network depth to 152 layers while ensuring the network accuracy and then further increased the depth to 1000 layers. Theoretically, the deeper the network was, the higher the accuracy should be. However, the authors experimentally found that blindly increasing the depth would lead to the degradation of the network when the depth reached a certain level. The gradient explosion and gradient disappearance of deep networks failed to train the model correctly and led to poor network performance. Inspired by highway network, the authors proposed the residual structure by adding a jump connection between the input and output of the convolutional block, which enabled the input to be passed directly to the output. The residual structure was essentially designed to learn a constant mapping with the nonlinear layer portion of the stack learning another mapping. As shown in Figure 2, if the residual structure was zero, we could easily train a constant mapping. In short, if a network could achieve the desired result by simply setting parameters manually, it was not difficult to train the network to converge to that result so that the added residual structure would at least not degrade the overall performance of the network. ResNet’s residual module reduced the difficulty in training deep networks, solved the degradation problem well, and maximized the depth potential of convolutional networks. Eventually, ResNet outperformed human performance in terms of the ImageNet classification task for the first time.(5)Optimizer. SGD [47] experienced difficulty exploring gorges. For example, territories, where the surface bent considerably more steeply in one measurement than in another, were basically around neighborhood optima. In these situations, SGD swayed over the gorge’s slants and slowly advanced along the valley floor towards the local optimal direction, as indicated in Figure 3(a). The energy was a technique that quickened SGD and hose motions in a pertinent way, as shown in Figure 3(b). As suggested by formula (1), it achieved this by including a part γ of the update vector of the past time venture to the current update vector.The energy term was usually set to 0.9 or a comparative value. In general, when utilizing energy, we pushed a ball down a slope. The ball collected energy as it moved downhill, getting quicker and quicker in transit (until it arrived at its max speed, if there was air obstruction, for example,  < 1). Something very similar happened to our boundary refreshing: the momentum term increases for the latitude where the gradient points in the same direction, while for the dimension where the gradient changes direction, the momentum term will decrease and update. Accordingly, we increased quicker union and diminished swaying.(6)Optimizer. The logistic loss could be calculated by the following formula:After applying it to the above calculation, we got logistic loss formula as follows:In order to calculate stability and prevent overflowing, the loss function would be calculated using the following formula:

3.3. Voter Model

Voting was a combination strategy aimed at classification problems in ensemble learning. The basic idea was to select the class with the highest output among all machine learning algorithms. There were two types of output judging from a machine learning classification algorithm: one was the direct output of class labels and another was the output of class probabilities. Using the former for voting was called majority/hard voting while using the latter for classification was called soft voting. Hard voting selected the label with the most output of the algorithm. If the number of labels was equal, the selection was made in ascending order. Soft voting used the class probabilities output by each algorithm to select a class. If the weight was input, a weighted average of the class probabilities of each class would be obtained, and the class with a large value would be selected. In this paper, our experiment used the hard voting mechanism.

3.4. Data Augmentation Strategy

Binary coding was used to represent positive and negative samples in this paper. Existing common DA methods consisted of randomly flipping the image (horizontally or vertically), randomly adding noise, rotating the image, changing the brightness, contrast, and saturation of the image, randomly cropping the image, randomly scaling/stretching the image, and randomly changing the clarity of the image. All these methods belonged to (1), (2), and (3) of Section 2.2. Based on these DA methods, we performed 12 different DAMF combinations and made 12 new datasets, as listed in Table 1. Specifically, third-party DA libraries were used in the 11th DAMF and 12th DAMF, respectively. As shown in Figure 4, all the images in the original dataset represented images without corresponding augmentation. The enhanced images were displayed following each original image. Figure 4(b) shows the randomly rotated operation, with the rotation angle at 90/180/270/360 degrees. Figure 4(d) shows the consequence of randomly adding Gaussian white noise to the original operation. Figure 4(f) describes the operation after random adjustment of brightness, saturation, and contrast built on the original image. Figure 4(h) describes the random cropping and stretching based on the original image. Figure 4(j) displays the image after randomly adjusting the sharpness. Figure 4(l) displays the image after randomly adjusting the contrast, saturation, and brightness on the original image and adding random Gaussian white noise. Figure 4(n) shows a randomly rotated, cropped, and stretched image built on the original image. Figure 4(p) shows the effect of randomly superimposing the images after all operations on the original images. Figure 4(r) shows the effect of utilizing the third-party library imaging to mutate the image. Figure 4(t) shows the effect of randomly superimposing all the above special effects.


No.Training set nameDA methodQuantity

1PALM-Training800-overturningOriginal dataset + random flip (4 directions: up, down, left, and right)800
2PALM-Training800-noiseOriginal dataset + Gaussian white noise800
3PALM-Training800-colorOriginal dataset + randomly changing colors (brightness, contrast, saturation)800
4PALM-Training800-croppingOriginal dataset + random cropping800
5PALM-Training800-deformingOriginal dataset + random scaling, stretching (stretched into a square by the length or width of the images)800
6PALM-Training800-dimmingOriginal dataset + change clarity800
7PALM-Training1600-overturning-noise-colorRandomly stack method 3 or 4 (serial number) on the basis of PALM-Training800-overturning1600
8PALM-Training1600-overturning-cropping-deformingRandomly stack method 5 or 6 (serial number) on the basis of PALM-Training800-overturning1600
9PALM-Training1600-overturning-dimmingRandomly stack method 7 (serial number) on the basis of PALM-Training800-overturning1600
10PALM-Training3200-overturning-noise-color-cropping-deforming-dimmingRandomly superimpose method 5 or 6 or 7 (serial number) on the basis of PALM-Training800-overturning-noise-color3200
11PALM-Training800-imgaug1Original dataset + random cropping with 0–50 pixels around, 50% probability horizontal flip, Gaussian blur (sigma = 0 to 3.0)800
12PALM-Training1600-overturning-dimming-imgaug2PALM-Training800-overturning-dimming dataset + multiple mixed random overlay1600

3.5. Primary Learner Model

AlexNet, GooLeNet, VGG-16, and ResNet-50 were used as the primary learners in this paper. The experiment set the learning rate to 0.001 and utilized the optimizer and loss function described above. Each model was formed for 30 epochs, with each epoch covering all images in the training set.

The highest accuracy of each model was selected after comprehensive training, and the corresponding model parameters were saved. Then, the convolution layer of all primary learners would be frozen, which meant that the data could only be transmitted forward instead of backward after entering the primary learners.

3.6. Staking Model Integration Strategy and Hard Voting Model

Staking, as a hierarchical model integration framework, was one of the main strategies widely used in model integration. Taking two layers as an example, the first layer consisted of multiple base learners, and the original training set was the input of the primary learners. The output of the primary learners was treated as the secondary learners’ input, which was the training set of secondary learners. The secondary learners continued training on the above training set to obtain the complete staking model.

As shown in Algorithm 1, processes 1–3 constructed the trained primary learners. Processes 5–9 were the prediction results of the training set using the trained primary learners, and this prediction was used as the training set for the secondary learners. Process 11 used the prediction results of the primary learners to train the secondary learners to get the fusion model.

Model input:
Training set D = {(x1, y1), (x2, y2), …, (xm, ym)}
   Primary learner £1, £2, …, £T
   Secondary learner £
Process:
 1: for t = 1, 2, …, T do
 2:   ht  =  £T (D)
 3: end for
 4: D = Ø;
 5: for t = 1, 2, …, m do
 6:   for t = 1, 2, …, T do
 7:    Zit = ht (xi);
 8:   end for
 9:   Dj  =  DjU ((zi1, zi2, …, ziT), yi)
 10:   hj  =  (Dj)
 11: end for
  Output: H (x) = hj (h1 (x), h2 (x), h3 (x), …, hT (x))

The design facilitated the extension of the model. In other words, the hard voting model could be replaced with other secondary learners based on different datasets. Figure 5 displays the framework of the fusion model in this paper. First, the training set normalized the data through the input layer, namely, processed it into a format of the same size (224 ∗ 224). Next, the images were input into each classifier, respectively, in the primary learner. The classifier performed 30 times epoch supervised training on the image according to the label of all inputs and then used the training result as the input of the secondary learning, and the final classification result was obtained after voting.

4. Results

4.1. Lab Environment

Hardware environment is as follows: CPU 4 cores, RAM 32 GB, GPU v100, video memory 16 GB, and disk 100 GB.

Environment configuration is as follows: Python version python3.7 and framework version PaddlePaddle 1.8.0.

4.2. Evaluation Indices

The primary reference record was the accuracy of the model forecast. In this paper, recall rate, specificity, and sensitivity were not used as evaluation indices. The classification performance was mainly evaluated by the classification accuracy, which was defined as follows:where TP, TN, FP, and FN denoted the true positive, the true negative, the false positive, and the false negative, respectively.

The loss function in the model was measured by the root mean square error, namely, a risk metric corresponding to the expected value of the squared (quadratic) error or loss. If was the predicted value of the i-th sample and was the corresponding true value, then the mean squared error (MSE) estimated over was defined as follows:

4.3. Primary Learner Training Process and Results

As shown in Figure 6, firstly, VGG-16 was used as a dataset filter, and training was conducted on all datasets. Each dataset was trained for 30 epochs. Each epoch would traverse all the datasets once to form the corresponding trained models on different datasets. These 13 datasets were adopted to make predictions on the test set, and the final results are listed in Table 2. According to Table 2, the overall accuracy of the enhanced dataset was higher than that of the original dataset. To go into detail, the average accuracy of PALM-Training1600-overturning-dimming-imgaug2, PALM-Training3200-overturning-noise-color-cropping-deforming-dimming, PALM-Training1600-overturning-cropping-deforming, and PALM-Training800-color exceeded 95%. Therefore, these 4 datasets were selected to be the candidate datasets. GoogLeNet, AlexNet, and ResNet-50 were also trained on these 4 datasets. Each model was trained on each dataset for 30 epochs, and then the training model was tested on the test set.


No.DatasetAccuracyLoss

1PALM-Training1600-overturning-dimming-imgaug20.958583360.18674079
2PALM-Training3200-overturning-noise-color-cropping-deforming-dimming0.955500010.27185006
3PALM-Training1600-overturning-cropping-deforming0.952666680.16523545
4PALM-Training800-color0.950333360.17019135
5PALM-Training800-dimming0.948750020.17919912
6PALM-Training800-cropping0.946250.18303553
7PALM-Training1600-overturning-dimming0.945250030.23124305
8PALM-Training1600-overturning-noise-color0.940083330.21350351
9PALM-Training800-overturning0.938583350.20894363
10PALM-Training800-deforming0.937083340.20814224
11PALM-Training800-noise0.936083350.26124661
12PALM-Training800-imgaug10.933916670.19876853
13PALM-Training4000.930166670.19310093

Follow-up training was made on the candidate datasets, and the results are shown in Table 3. We used the same parameters during the process of training. It could be seen from Table 3 that the optimal DAMF datasets also varied because of the differences in the expression ability of different models. To be specific, GoogLeNet and ResNet-50 were both trained on the PALM-Training3200-overturning-noise-color-cropping-deforming-dimming dataset. AlexNet and VGG-16 had the highest scores on the PALM-Training1600-overturning-dimming-imgaug2 dataset. The above four models with the highest accuracy were used as primary learners.


Primary learnerPALM-Training800-colorPALM-Training1600-overturning-cropping-deformingPALM-Training3200-overturning-noise-color-cropping-deforming-dimmingPALM-Training1600-overturning-dimming-imgaug2
AccuracyLoss rateAccuracyLoss rateAccuracyLoss rateAccuracyLoss rate

AlexNet0.9460830.1623230.9508330.1606980.9546670.1970960.9575830.157773
GoogLeNet0.9090.2034610.93950.1697590.9624170.136177270.9479170.170418
ResNet-500.93950.2195710.9535830.1518360.9559170.1567370.951250.166358
VGG-160.95030.170190.952670.165230.95550.271850.958580.18674

In this paper, the accuracy rate was the average of the accuracy rate results of 30 epochs of training. The highest accuracy of the model training set reached 100% (Figure 6). Table 2 still shows the comparison between applying DAMF and not applying DAMF, in which the 13th dataset is the original dataset. We can see that the results of all the datasets processed by DAMF were better than the results of the 13th dataset, and the best DAMF corresponding to the first group of results is 2.84% higher than the 13th group, reached 95.85%. This improvement is obvious. Similarly, we have observed AlexNet, GoogLeNet, and ResNet-50. From the results for training, the accuracy rates of these were, respectively, 95.76%, 96.24%, and 95.60%, which are highly similar to the best training results of VGG-16, i.e., all exceed 95.00%. The above results could show that DAMF is universal.

Figure 7 displays the loss trend during the training process of VGG-16. Taking the loss rate of the training original dataset as a baseline for comparison, we could find that although some DAMF loss rates were higher than 0.19, the highest loss rate was 0.27, which means the loss rate was acceptable. Among them, the loss rate corresponding to the best DAMF is 0.19, which was the same as the original dataset, which means that the enhanced data did not cause additional loss.

4.4. Fusion Model Training Process and Results

Figure 8 displays the logic diagram of the fusion model, with the most accurate models (AlexNet, GoogLeNet, ResNet-50, and VGG-16) used as the primary learners. The predictions of all primary learners were used as the training dataset for secondary learners. Meanwhile, the original dataset label was treated as the label of the new dataset to build the training set of the secondary model. After the primary model, the hard voting model was inserted as a classifier to form the framework of the secondary learners. In the secondary model, the four prediction results (AlexNet result, GoogLeNet result, ResNet-50 result, and VGG-16 result) in each sample were counted as the final prediction results. After 30 epochs of training, the model was saved and validated on a test set. The final accuracy of the fusion model reached 97.25% (average accuracy), with a maximum accuracy of 98.00%.

5. Discussion

First, 12 DAMF strategies were implemented on the iChallenge-PM dataset, resulting in the formation of 13 datasets, including the original one. Then, the experiment used VGG-16 as a dataset picker to train each of these 13 datasets for 30 epochs, each epoch covering all the data once. The accuracy of the model after training on each dataset was obtained on the validation set. As universally agreed, the four datasets with the best accuracy were selected as the preselected datasets to be used in the training of all remaining models. At the end of the training, the model with the highest prediction accuracy on the validation set was chosen as the primary learner.

Through this experiment, the advantages of DA were evident, and the data augmented datasets generally obtained higher accuracy than that of the original dataset. The average accuracy of the VGG-16 models trained by the four preselected datasets mentioned above was 95.85%, which was 2.84% higher than that of the original dataset.

This study adopted the strategy of model integration. The experiment retrained the output of the first-level model. The average accuracy of the first-level model was 95.86%, and the prediction accuracy of the fusion model was once again improved by 1.39%. Particularly, the greater significance of the fusion model was that the shortcomings of each primary learner were balanced. Therefore, the generalization performance and expression of the model were effectively improved.

We could observe the performance of DAMF in the primary learners again. Figure 9 displays the training process of AlexNet, GoogLeNet, and ResNet-50 on the four optimal DAMF datasets. The details were expressed as follows: the solid line part in each subfigure referred to the accuracy, the dotted line part referred to the loss rate, the abscissa represented the training epoch, and the ordinate represented the percentage (%). It was worth noting that the accuracy of each model could reach about 95% at the end of training without model fusion, which meant that DAMF played a highly favorable role.

We compared all DAMFs during the training process of VGG-16 by making them learn from all 13 datasets. Figure 10 exhibits the training effect of VGG-16 on the passing accuracy and loss rate of each dataset. It was believed that DAMF should not be as complicated as possible. The best result appeared in PALM-Training1600-overturning-dimming-imgaug1. When DAMF got complicated, the effect would decrease instead. Excessively complex processing of images might destroy valuable features in the image. This meant that DA was a process rather than formula and that we needed to locate the DAMF dataset that best fit each dataset in the dynamic process.

We compared the classification results under different strategies on a fixed dataset (iChallenge-PM). Fully considering the differences of research studies, part of the research focused on the optimization of the network structure, while others focused on the DA. The overall idea of the research is to start from the optimization of the network model and DA direction, with the ultimate goal of model accuracy. Table 4 displays the accuracy of different studies on the iChallenge-PM dataset in recent years. The results showed that although our accuracy rate is not the highest, it is also encouraging. In particular, our calculation cost was low. All training took 19 hours and 56 minutes, and no expensive calculation methods such as transfer learning were used.


Accuracy (%)Methods

Siying Dai [28]81.82Optimize network structure + DA
InstDis [48]95.32Optimize network structure + DA
Contrastive [49]96.94Optimize network structure + DA
Invariant [50]97.30Optimize network structure + DA
Xiaomeng Li [51]98.65Optimize network structure + DA
Ours97.25Optimize network structure + DA

6. Conclusion

The in-depth analysis of the discussed image DA techniques based on fusion has taken PM images as the research object and convincingly displayed their wide variety of applications. DAMF has the advantage of effectively improving the accuracy of model training, and the optimal enhanced set of different datasets can be matched through this strategy. By analyzing the image DA method proposed in this paper, DAMF proves to be better and more effective than other methods. The experiment results have shown that DA still has an optimal complexity in the combination of DAMF. Otherwise, too much complexity may destroy the original features. DAMF can effectively find the best combination of DA which well retains the characteristics of the input images and provides better contrast through 11 contrast combinations. It has also been observed that DAMF can train the model to satisfactory results without using transfer learning or other methods. Arguably, DAMF can be used as an effective DA method during the training of CNNs in the field of fundus image processing.

Data Availability

The data used to support the findings of this study are included within the article.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This work was supported by the Xiamen Science and Technology Program (3502Z20183047) and China Education Research Network (NGII20170620).

References

  1. C. W. Wong, B. J. Fenner, and G. C. M. Cheung, “Epidemiology of myopic and vitreomaculopathies,” in A. Chang, W. F. Mieler, M. Ohji (eds), pp. 285–295, Macular Surgery, Springer, Singapore, 2020. View at: Publisher Site | Google Scholar
  2. K. Ohno-Matsui, “Pathologic myopia,” Asia-Pacific Journal of Ophthalmology, vol. 5, no. 6, pp. 415–423, 2016. View at: Publisher Site | Google Scholar
  3. T. Y. Wong, A. Ferreira, R. Hughes, G. Carter, and P. Mitchell, “Epidemiology and disease burden of pathologic myopia and myopic choroidal neovascularization: an evidence-based systematic review,” American Journal of Ophthalmology, vol. 157, no. 1, pp. 9–25, 2014. View at: Publisher Site | Google Scholar
  4. S.-K. Jung, J. H. Lee, H. Kakizaki, and D. Jee, “Prevalence of myopia and its association with body stature and educational level in 19-year-old male conscripts in Seoul, South Korea,” Investigative Opthalmology & Visual Science, vol. 53, no. 9, p. 5579, 2012. View at: Publisher Site | Google Scholar
  5. B. A. Holden, T. R. Fricke, D. A. Wilson et al., “Global prevalence of myopia and high myopia and temporal trends from 2000 through 2050,” Ophthalmology, vol. 123, no. 5, pp. 1036–1042, 2016. View at: Publisher Site | Google Scholar
  6. Y.-L. Wong, C. Sabanayagam, Y. Ding et al., “Prevalence, risk factors, and impact of myopic macular degeneration on visual impairment and functioning among adults in Singapore,” Investigative Opthalmology & Visual Science, vol. 59, no. 11, p. 4603, 2018. View at: Publisher Site | Google Scholar
  7. C. W. Wong, N. Brennan, and M. Ang, “Introduction and overview on myopia: a clinical perspective,” in M. Ang, T. Wong (eds) Updates on Myopia, pp. 1–26, Springer, Singapore, 2020. View at: Publisher Site | Google Scholar
  8. S. Vitale, M. F. Cotch, and R. D. Sperduto, “Prevalence of visual impairment in the United States,” JAMA, vol. 295, no. 18, p. 2158, 2006. View at: Publisher Site | Google Scholar
  9. I. G. Morgan, K. Ohno-Matsui, and S.-M. Saw, “Myopia,” The Lancet, vol. 379, no. 9827, pp. 1739–1748, 2012. View at: Publisher Site | Google Scholar
  10. W.-M. Hsu, C.-Y. Cheng, J.-H. Liu, S.-Y. Tsai, and P. Chou, “Prevalence and causes of visual impairment in an elderly Chinese population in Taiwan the authors have no proprietary interest in any aspect of the study,” Ophthalmology, vol. 111, no. 1, pp. 62–69, 2004. View at: Publisher Site | Google Scholar
  11. A. Iwase, M. Araie, A. Tomidokoro, T. Yamamoto, H. Shimizu, and Y. Kitazawa, “Prevalence and causes of low vision and blindness in a Japanese adult population,” Ophthalmology, vol. 113, no. 8, pp. 1354–1362, 2006. View at: Publisher Site | Google Scholar
  12. L. Xu et al., “Causes of blindness and visual impairment in urban and rural areas in beijing: the beijing eye study,” Ophthalmology, vol. 113, no. 7, pp. 1134.e1–1134.e11, 2006. View at: Publisher Site | Google Scholar
  13. M. Yamada, Y. Hiratsuka, C. B. Roberts et al., “Prevalence of visual impairment in the adult Japanese population by cause and severity and future projections,” Ophthalmic Epidemiology, vol. 17, no. 1, pp. 50–57, 2010. View at: Publisher Site | Google Scholar
  14. Q. S. You, L. Xu, H. Yang, Y. X. Wang, and J. B. Jonas, “Five-year incidence of visual impairment and blindness in adult Chinese,” Ophthalmology, vol. 118, no. 6, pp. 1069–1075, 2011. View at: Publisher Site | Google Scholar
  15. J. He, S. L. Baxter, J. Xu, J. Xu, X. Zhou, and K. Zhang, “The practical implementation of artificial intelligence technologies in medicine,” Nature Medicine, vol. 25, no. 1, pp. 30–36, 2019. View at: Publisher Site | Google Scholar
  16. W. Lu, Y. Tong, Y. Yu, Y. Xing, C. Chen, and Y. Shen, “Applications of artificial intelligence in ophthalmology: general overview,” Journal of Ophthalmology, vol. 2018, Article ID 5278196, 15 pages, 2018. View at: Publisher Site | Google Scholar
  17. M. I. Razzak, S. Naz, and A. Zaib, “Deep learning for medical image processing: overview, challenges and the future,” in Lecture Notes in Computational Vision and Biomechanics, pp. 323–350, Springer International Publishing, Berlin, Germany, 2018. View at: Publisher Site | Google Scholar
  18. K.-H. Yu, A. L. Beam, and I. S. Kohane, “Artificial intelligence in healthcare,” Nature Biomedical Engineering, vol. 2, no. 10, pp. 719–731, 2018. View at: Publisher Site | Google Scholar
  19. A. V. Varadarajan, R. Poplin, K. Blumer et al., “Deep learning for predicting refractive error from retinal fundus images,” Investigative Opthalmology & Visual Science, vol. 59, no. 7, p. 2861, 2018. View at: Publisher Site | Google Scholar
  20. D. S. W. Ting, L. R. Pasquale, L. Peng et al., “Artificial intelligence and deep learning in ophthalmology,” British Journal of Ophthalmology, vol. 103, no. 2, pp. 167–175, 2019. View at: Publisher Site | Google Scholar
  21. S. Kuwayama, “Automated detection of macular diseases by optical coherence tomography and artificial intelligence machine learning of optical coherence tomography images,” Journal of Ophthalmology, vol. 2019, Article ID 6319581, 7 pages, 2019. View at: Publisher Site | Google Scholar
  22. D. T. Hogarty, D. A. Mackey, and A. W. Hewitt, “Current state and future prospects of artificial intelligence in ophthalmology: a review,” Clinical & Experimental Ophthalmology, vol. 47, no. 1, pp. 128–139, 2019. View at: Publisher Site | Google Scholar
  23. K. Suzuki, “Overview of deep learning in medical imaging,” Radiological Physics and Technology, vol. 10, no. 3, pp. 257–273, 2017. View at: Publisher Site | Google Scholar
  24. Z. Li, S. Keel, C. Liu et al., “An automated grading system for detection of vision-threatening referable diabetic retinopathy on the basis of color fundus photographs,” Diabetes Care, vol. 41, no. 12, pp. 2509–2516, 2018. View at: Publisher Site | Google Scholar
  25. J. De Fauw, J. R. Ledsam, B. Romera-Paredes et al., “Clinically applicable deep learning for diagnosis and referral in retinal disease,” Nature Medicine, vol. 24, no. 9, pp. 1342–1350, 2018. View at: Publisher Site | Google Scholar
  26. K. Shankar, Y. Zhang, Y. Liu, L. Wu, and C.-H. Chen, “Hyperparameter tuning deep learning for diabetic retinopathy fundus image classification,” IEEE Access, vol. 8, pp. 118164–118173, 2020. View at: Publisher Site | Google Scholar
  27. X. Sun, X. Cao, Y. Yang, L. Wang, and Y. Xu, “Robust retinal vessel segmentation from a data augmentation perspective,” 2020, https://arxiv.org/abs/2007.15883. View at: Google Scholar
  28. S. a. C. Dai, Leiting, T. Lei, C. Zhou, and Y. Wen, “Automatic detection of pathological myopia and high myopia on fundus images,” in Proceedings of the IEEE International Conference on Multimedia and Expo (ICME), pp. 1–6, London, UK, 2020. View at: Google Scholar
  29. J. Cheng, Z. Li, Z. Gu, H. Fu, D. W. K. Wong, and J. Liu, “Structure-preserving guided retinal image filtering and its application for optic disk analysis,” IEEE Transactions on Medical Imaging, vol. 37, no. 11, pp. 2536–2546, 2018. View at: Publisher Site | Google Scholar
  30. L. Wang, “Comparative analysis of image classification methods for automatic diagnosis of ophthalmic images,” Scientific Reports, vol. 7, no. 1, p. 41545, 2017. View at: Publisher Site | Google Scholar
  31. Z. Zhang, J. Cheng, J. Liu, Y. C. M. Sheri, C. C. Kong, and S. S. Mei, “Pathological myopia detection from selective fundus image features,” in Proceedings of the 7th IEEE Conference on Industrial Electronics and Applications (ICIEA), pp. 1742–1745, Singapore, 2012. View at: Google Scholar
  32. J. Jiang, “Automatic diagnosis of imbalanced ophthalmic images using a cost-sensitive deep convolutional neural network,” BioMedical Engineering OnLine, vol. 16, no. 1, 2017. View at: Publisher Site | Google Scholar
  33. H. Ačkar, A. A. Allmisreb, and M. A. Saleh, “A review on image enhancement techniques,” International Journal of Engineering and Applied Computer Science (IJEACS), vol. 2, no. 7, 2017. View at: Google Scholar
  34. R. Bernardes, P. Serranho, and C. Lobo, “Digital ocular fundus imaging: a review,” Ophthalmologica, vol. 226, no. 4, pp. 161–181, 2011. View at: Publisher Site | Google Scholar
  35. N. V. Chawla, A. Lazarevic, L. O. Hall, and K. W. Bowyer, “SMOTEBoost: improving prediction of the minority class in boosting,” Knowledge Discovery in Databases: PKDD 2003, Springer, Berlin, Germany, 2003. View at: Publisher Site | Google Scholar
  36. A. S. Parihar, K. Singh, H. Rohilla, G. Asnani, and H. Kour, “A comprehensive analysis of fusion-based image enhancement techniques,” in Proceedings of the 4th International Conference on Intelligent Computing and Control Systems (ICICCS), IEEE, Melur, India, 2020. View at: Google Scholar
  37. A. Maheshwari and Y. Gupta, “Empirical aspect of big data to enhance medical images using HIPI,” in Proceedings of the 2018 Second International Conference on Intelligent Computing and Control Systems (ICICCS), IEEE, Melur, India, 2018. View at: Google Scholar
  38. F. c. c. o. Chollet, “Xception: deep learning with depthwise separable convolutions,” 2017, https://arxiv.org/abs/1610.02357. View at: Google Scholar
  39. A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet classification with deep convolutional neural networks,” Communications of the ACM, vol. 60, no. 6, pp. 84–90, 2017. View at: Publisher Site | Google Scholar
  40. K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” 2015, https://arxiv.org/abs/1512.03385. View at: Google Scholar
  41. X. Fu, D. Zeng, Y. Huang, Y. Liao, X. Ding, and J. Paisley, “A fusion-based enhancing method for weakly illuminated images,” Signal Processing, vol. 129, pp. 82–96, 2016. View at: Publisher Site | Google Scholar
  42. R. Takahashi, T. Matsubara, and K. Uehara, “Data augmentation using random image cropping and patching for deep CNNs,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 30, no. 9, pp. 2917–2931, 2020. View at: Publisher Site | Google Scholar
  43. H. Zhang, M. Cisse, Y. N. Dauphin, and D. Lopez-Paz, “Mixup: beyond empirical risk minimization,” 2018, https://arxiv.org/abs/1710.09412. View at: Google Scholar
  44. B. Zoph and Q. V. Le, “Neural architecture search with reinforcement learning,” 2017, https://arxiv.org/abs/1611.01578. View at: Google Scholar
  45. E. D. Cubuk, B. Zoph, D. Mane, V. Vasudevan, and Q. V. Le, “AutoAugment: learning augmentation policies from data,” 2019, https://arxiv.org/abs/1805.09501. View at: Google Scholar
  46. H. Fu, J. Cheng, Y. Xu, D. Wong, J. Liu, and X. Cao, “Joint optic disc and cup segmentation based on multi-label deep network and polar transformation,” IEEE Transactions on Medical Imaging, vol. 37, no. 7, pp. 1597–1605, 2018. View at: Publisher Site | Google Scholar
  47. B. Normalization, “Accelerating deep network training by reducing internal covariate shift,” 2015, https://arxiv.org/abs/1502.03167. View at: Google Scholar
  48. T. Chen, S. Kornblith, M. Norouzi, and G. Hinton, “A simple framework for contrastive learning of visual representations,” in Proceedings of the 37th International Conference on Machine Learning, Vienna, Austria, 2020, http://proceedings.mlr.press/v119/chen20j.html. View at: Google Scholar
  49. Q. V. Le, “Building high-level features using large scale unsupervised learning,” in Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, IEEE, Vancouver, Canada, 2013. View at: Google Scholar
  50. H. Lee, R. Grosse, R. Ranganath, and A. Y. Ng, “Convolutional deep belief networks for scalable unsupervised learning of hierarchical representations,” in Proceedings of the 26th Annual International Conference on Machine Learning, Montreal, Canada, 2009. View at: Publisher Site | Google Scholar
  51. Z. Wu, Y. Xiong, S. X. Yu, and D. Lin, “Unsupervised feature learning via non-parametric instance discrimination,” in Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, IEEE, Salt Lake City, UT, USA, 2018. View at: Google Scholar

Copyright © 2021 Jianfeng Cui et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Related articles

No related content is available yet for this article.
 PDF Download Citation Citation
 Download other formatsMore
 Order printed copiesOrder
Views354
Downloads352
Citations

Related articles

No related content is available yet for this article.

Article of the Year Award: Outstanding research contributions of 2021, as selected by our Chief Editors. Read the winning articles.