Computational Intelligence and Neuroscience

Computational Intelligence and Neuroscience / 2021 / Article

Research Article | Open Access

Volume 2021 |Article ID 5540186 | https://doi.org/10.1155/2021/5540186

Wenjun Li, Siyi Cheng, Kai Qian, Keqiang Yue, Hao Liu, "Automatic Recognition and Classification System of Thyroid Nodules in CT Images Based on CNN", Computational Intelligence and Neuroscience, vol. 2021, Article ID 5540186, 11 pages, 2021. https://doi.org/10.1155/2021/5540186

Automatic Recognition and Classification System of Thyroid Nodules in CT Images Based on CNN

Academic Editor: Elpida Keravnou
Received14 Jan 2021
Accepted15 May 2021
Published27 May 2021

Abstract

Thyroid nodule lesions are one of the most common lesions of the thyroid; the incidence rate has been the highest in the past thirty years. X-ray computed tomography (CT) plays an increasingly important role in the diagnosis of thyroid diseases. Nonetheless, as a result of the artifact and high complexity of thyroid CT image, the traditional machine learning method cannot be applied to CT image processing. In this paper, an end-to-end thyroid nodule automatic recognition and classification system is designed based on CNN. An improved Eff-Unet segmentation network is used to segment thyroid nodules as ROI. The image processing algorithm optimizes the ROI region and divides the nodules. A low-level and high-level feature fusion classification network CNN-F is proposed to classify the benign and malignant nodules. After each module is connected in series with the algorithm, the automatic classification of each nodule can be realized. Experimental results demonstrate that the proposed end-to-end thyroid nodule automatic recognition and classification system has excellent performance in diagnosing thyroid diseases. In the test set, the segmentation IOU reaches 0.855, and the classification output accuracy reaches 85.92%.

1. Introduction

Thyroid nodule lesions are the most common lesions of the thyroid, and the overall human prevalence can reach 19% to 46% [1]. Lesions can be divided into benign and malignant. Compared with benign nodules, malignant nodules have an irreversible impact on human health. The necessary condition for clinically correct treatment of malignant nodules is to accurately detect benign and malignant thyroid nodules [2].

In order to better solve this problem, many scholars have done a lot of research using computer-aided diagnosis technology. Nugroho used adaptive median filter and bilateral filtering for preprocessing the ultrasound images, eight geometric features were used to support vector machine (SVM) for classification, and the accuracy of thyroid nodule lesions is 92.30% [3]. Wang proposed a deep learning method to diagnose thyroid nodules using multiple ultrasound images in an examination as input. An attention-based feature aggregation network is proposed to automatically integrate the features extracted from multiple images in one examination, utilizing different views of the nodules to improve the performance of recognizing malignant nodules in the ultrasound images. In the self-built large database, the classification of benign and malignant can reach 87.32% [4]. Iakovidis et al. proposed encoding the ultrasonic features by antinoise representation and used the fusion of fuzzy local binary pattern and fuzzy gray histogram features to classify the ultrasonic features by polynomial kernel support vector machine. According to the area under the ROC curve, the classification accuracy reached 97.5% [5].

The above researches are based on ultrasound images and have achieved good results, but compared with CT images, ultrasound images have some problems, such as low definition, less useful information, and so on [6], and there is a big difference between the two images, which cannot achieve the simple moving of the method.

Now, in the study of thyroid nodules based on CT, Peng used the first-order texture features in nonenhanced CT images and support vector machine analysis to classify the nodules, and the accuracy reached 88% [7]. Liu used traditional machine learning methods, the recognition performance of thyroid nodules was evaluated according to 17 first-order statistics and gray cooccurrence matrix features. The sensitivity, specificity, positive predictive value, and negative predictive value of the final algorithm were 0.8673, 0.9105, 0.9130, 0.8269, 0.8235, and 0.9146, respectively [8]. Although the above researches have reached a high level in the study of thyroid classification, they require the prerequisite for the nodule information to be additionally labeled, and it is impossible to achieve high-efficiency data processing. On this basis, Zhao et al. designed an improved Unet architecture, Dense-Unet, to achieve the region of interest (ROI) segmentation of thyroid nodules and used multidimensional input fusion CNN model to achieve the classification of benign and malignant nodules. However, limited by the algorithm design, the algorithm is only suitable for single-nodule images, which cannot meet the situation of single picture and multiple nodule [9].

In order to solve the above problems, this study collected and established a thyroid nodule enhanced CT image database, aiming to realize the integration of thyroid nodule recognition and classification diagnosis. In CT images, the distinction between benign and malignant thyroid mainly lies in whether the shape and edges are divergent, and there is no obvious difference in color. As a result, a single multiclass segmentation network cannot achieve high accuracy and the classification effect of benign and malignant is poor. Therefore, in this research, a CNN-based semantic segmentation network plus an image classification network is designed, and image processing algorithms are used to make the two connected in a reasonable manner, so that the segmentation network focuses on the segmentation of the thyroid nodules and the background. The classification network focuses on extracting image features of thyroid nodules to achieve benign and malignant classification. In addition, for multinodule CT images, nodules are segmented to achieve separate classification of each nodule.

The main contributions of this work are as follows: (1)Based on the traditional Unet, an improved network called Eff-Unet is proposed for the segmentation of contrast-enhanced thyroid CT images, and an improved loss function is proposed to improve the semantic segmentation performance.(2)A new thyroid nodule classification method is designed by fusing two different CNN networks.(3)The image processing algorithm is designed to adapt the output image of segmentation network to the input of classification network and realize the intelligent segmentation of multinodule CT image. The end-to-end diagnosis and classification of each nodule is realized without any additional marking tasks.

2. Materials and Methods

The algorithm flowchart proposed in this paper is shown in Figure 1. It can be divided into 4 parts: image preprocessing part, segmentation network semantic segmentation part, image processing algorithm part, and classification network prediction and output visualization part. Input the CT image after image preprocessing into the system, realize the nodule splitting of multinodule pictures through image processing algorithms, and smoothly transition the output pictures of the segmentation network to the input of the classification network to realize end-to-end automatic recognition and integration diagnose thyroid nodules.

2.1. Database Characteristics

In this experiment, the CT image database of the First People’s Hospital of Pinghu City and the First People’s Hospital of Jiaxing City was used to construct the database. A total of 248 patients’ CT images from 2013 to 2020 were obtained. All images were taken by a 16-slice spiral CT scanner (Siemens Emotion 16, one of the most popular CT scanners in the world). The basic settings were CT slice thickness of 4 nm, window width and window level of 500 and 60, and convolution kernel parameter of b50s. All the images were enhanced CT images after injection of water-soluble iodine contrast agent and scanning, which can increase the density difference between the lesion tissue and adjacent normal tissue, so as to improve the display rate of lesions [10]. All the included lesions were confirmed by biopsy or surgical pathology, and the lesion scope was clear. Because there were multiple nodules, CT images with malignant nodules were classified as malignant images in this experiment. After being selected by professional doctors, a total of 832 CT images were obtained. The detailed data of CT images data collected are shown in Table 1. The detailed data of thyroid nodes data collected are shown in Table 2.


Training and validation setTest setTotal

Benign images50797604
Malignant images19830228
Total705127832


Training and validation setTest setTotal

Benign nodes543103646
Malignant nodes24334277
Total786137923

2.2. Image Preprocessing

The format of CT exported image is JPEG, and the size is 512 ∗ 512 pixels. In order to avoid learning error noise, such as bones and blood vessels, the central part of CT image is intercepted uniformly, and the size of saved core area is 256 ∗ 256 pixels. The image labeling work is assisted by two professional doctors, and the LabelMe labeling tool is used to generate the mask image. In the original mask image, the pixels of the generated background area are 0, the pixels of the benign thyroid area are 1, and the pixels of the malignant thyroid area are 2. The total number of images is relatively small, which cannot meet the condition of deep learning algorithm training with a large number of data. Therefore, data enhancement is used to generate new images to improve the performance and robustness of the model. Through the data enhancement tool of Keras framework, setting appropriate rotation angle limit and offset amplitude limit, and flipping horizontally to realize various transformations of the original image, a new image is generated and input into the algorithm to trick the model into thinking that it is receiving new data, thereby making the model more powerful in processing new types of images [11].

2.3. Image Segmentation Model

In the automatic recognition of thyroid nodules, the Unet [12] network model is used for image semantic segmentation, and the deep learning algorithm is used to automatically divide the location, shape, and size of the nodules. The organ structure of thyroid CT image is fixed and the semantic information is not particularly rich. The fusion of different scale features is an important means to improve the segmentation performance. Low-level features have higher resolution and contain more location and detail information. However, due to less convolution, they have lower semantics and more noise. High-level features have stronger semantic information, but the resolution is very low, and the perception of details is poor. Therefore, both high-level semantic information and low-level features are very important. Consequently, the jump connection structure and U-shaped structure network of Unet can combine the high-resolution local information and the lower-resolution and larger-area information at the same time, making the segmentation of thyroid nodules more accurate [13].

In addition, the EfficientNet [14] network structure is selected as its backbone [15]. EfficientNet uses the depth, width, and resolution adjustment technology. Compared with other mature networks, EfficientNet has the huge advantage of number of parameters and computation [16]. It can extract image features quickly and efficiently, which lays the foundation for the realization of accurate segmentation of Unet network. The Eff-Unet designed in this experiment is shown in Figure 2. As shown in Figure 2, its basic structure is similar to the Unet network, but the downsampling encoder part rewrites its original Vgg-like structure and chooses the EfficientNet with excellent performance.

For the segmentation network, the main function is to automatically label the nodule part, and there is no need to distinguish between benign and malignant nodules. Therefore, the mask image needs additional processing before it is input into the network. When there is a malignant area with a pixel value of 2, it is automatically converted to a pixel value of 1. The output image and input image of the model are the same size. After specific image processing, it is used as the subsequent classification network input.

2.4. Image Processing Algorithm

A specific image processing algorithm is designed to realize the smooth connection of the two networks for making the output image of the segmentation network better adapt to the classification network. At the same time, we also achieve the segmentation of each nodule in the multinodule image and complete the image classification of the nodule part.

Aiming at the situation that the pixel value of the output image of the segmentation network is a floating-point value, in order to facilitate the classification network learning, according to several attempts to verify, a threshold of 0.3 is designed to binarize the background and nodule of the output image.

The output of the segmentation network may have a small range of noise color blocks caused by some misjudgment. The length width ratio and the total area limit are designed to eliminate the interference without excluding the small-area nodules.

A single CT image will have multiple nodules. In order to avoid the degradation of classification effect caused by the hybridity of multiple nodules and single nodules, OpenCV is used to analyze the minimum connected domain of all mask images, find out all single nodules, and lay them in the black background of 256 ∗ 256 pixels according to the position of the original image, so as to facilitate the subsequent mapping to the original image for visualization. In addition, according to the pixel value of the corresponding mask image in the ROI region of the nodule, when the average pixel value tends to 1, the cut nodule is marked as benign nodule, and when the average pixel value tends to 2, it is marked as malignant nodule.

2.5. Image Classification Model

The classification of benign and malignant thyroid nodules is based on CNN image classification [17]. The nodules of CT images are gray-black pixel blocks with subtle color differences, so the classification is mainly based on the shape and edge differences of benign and malignant nodules [18]. Based on this feature, a low-level and high-level feature fusion classification network CNN-F is designed. Its basic structure is shown in Figure 3. The shallow network is used to extract low-level image features, and the deep network is used to extract complex high-level image features. After fusion training, we can learn the performance of thyroid nodules at different characteristic levels.

The shallow network CNN-1 is built independently, the main structure is shown in Figure 4. It is mainly composed of a combination of 3 parts of convolutional structure and pooling structure. The image features are extracted through the convolutional layer, and the resulting feature maps are compressed through the maximum pooling layer. On the one hand, the feature maps are reduced, and the network calculation complexity is simplified; on the one hand, feature compression is performed, the main features are extracted, and one dropout layer is added to avoid overfitting. Then, through the Flatten layer, the multidimensional input is made one-dimensional, and the transition from the convolutional layer to the fully connected layer is realized. Add two layers of fully connected layers and one layer of dropout, plus the nonlinear mapping of the activation function, to extract and integrate useful information to achieve the role of a classifier.

The deep network structure CNN-2 uses the InceptionV3 [19] network as the basic structure, and the CNN-2 structure is shown in Figure 5. The Inception network structure is a “basic neuron” structure constructed by the GoogLeNet team to build a sparse, high-performance network structure. Convolution kernels of different sizes are used to enable the existence of different sizes of receptive fields. Finally, the splicing is achieved to achieve the fusion of features of different scales, which can not only maintain the sparsity of the network structure, but also utilize the high computational performance of the dense matrix. The InceptionV3 network introduces the idea of Factorization into small convolutions, which splits a larger two-dimensional convolution into two smaller one-dimensional convolutions. On the one hand, it saves a large number of parameters and reduces computing time and overfitting, while increasing a layer of nonlinear extended model expression capabilities [20]. The first module uses two 3 × 3 convolutions on a traditional basis instead of 5 × 5 convolutions. The second module reduces the feature map and adds filters and structure, and the third module uses the convolutional pooling parallel structure. After calling the basic concept V3 (excluding the top-level classifier), the global average pooling layer [21] is added, which directly eliminates the black box feature in the full-connection layer, and gives each channel the actual internal meaning, greatly reducing the network parameters and avoiding overfitting. After that, a full-connection layer of 128 nodes is added to integrate the local feature map to get the global feature information.

After removing the output layer and using concatenate to splice the two models, 256 nodes are merged through the fully connected layer. Finally, the output layer uses the sigmoid activation function to predict the category and can also get approximate probability prediction, which can use probability to assist with classification decision [22].

2.6. Loss Function

In this experiment, binary dice loss is used to improve the loss of segmentation model. Loss is divided into Dice Loss and Binary Cross Entropy. Among them, the Dice coefficient is a measurement function used to measure the similarity of the set, which is actually used to calculate the pixels between samples [23]. The formula of Dice coefficient is as follows: where X represents the marked ground truth and Y represents the output image of the segmentation network. The formula of Dice Loss is defined as

Using Dice Loss as the loss function can achieve the real goal of segmentation more appropriately, optimize the evaluation criteria directly, and better deal with the imbalance of data. However, Dice Loss will have an adverse impact on the back propagation and easily make the training unstable [24]. Therefore, on the basis of Dice Loss, Binary Cross Entropy is added [25]. The calculation formula is as follows: where N is the output size, is the label, and is the predicted probability of the label being positive for all N labels.

3. Results and Discussion

3.1. Experiment Setup

The experiment was conducted on a personal computer with NVIDIA GEFORCE RTX-2080Ti GPU and Intel® Core™ i7-8700K Processor, 3.7 GHz. The experimental operating system is 64-bit Windows 10. The environment for deep learning is Python 3.6 and Keras 2.3.1 with TensorFlow GPU 1.14 as the back end. In the model training, the loss function used in segmentation network is Dice Loss + Binary Cross Entropy, the optimizer is Adam, and the performance evaluation index is IOU. In the classification network, CNN-1, CNN-2, and CNN-F used Binary Cross Entropy as the loss function, Adam as the optimizer, and ACC as the performance evaluation index. The initial learning rate of both methods was 0.001. The ReduceLROnPlateau callback function was used to dynamically adjust the learning rate. The attenuation coefficient was set to 0.5, and the cooldown was set to 10 epochs. A total of 200 epochs were trained. In addition, the parameter setting details of comparison methods are also consistent with the above, such as Vgg-16, DenseNet, and so on.

3.2. Performance Evaluation

The performance of the network model after concatenation can be evaluated by determining statistical values (recall, specificity, precision, and accuracy) and the F1-score. The recall reflects the positive proportion of correct recognition, the specificity reflects the negative proportion of correct recognition, the precision is the repeatability, or reproducibility of the measurement, and the accuracy is the proximity of measurement results to the true value. F1-score is an evaluation index which takes both precision and recall into account. The calculation formulas are shown as follows, where TP represents true positive, TN represents true negative, FP represents false positive, and FN represents false negative.

3.3. Semantic Segmentation Performances

In this experiment, in order to better segment the thyroid nodules and background parts, EfficientNetB4 was selected as the Unet model of the backbone network. In order to test the performance of the model, on the basis of Unet, we tried the common vgg backbone network and resnet18 backbone network. In addition, we also choose Linknet [26] and FPN [27] as the segmentation network and find that the performance of segmentation is worse than that of UNET in the case of EfficientNetB4 backbone network. The training curve diagram of the training set of various network structures is shown in Figure 6. The IOU scores of each method in the test set are shown in Table 3, in which Eff-Unet gets the highest score.


MethodIOU

Eff_FPN0.768
Eff_Linknet0.836
Res18-Unet0.833
Vgg-Unet0.842
Eff-Unet0.856

3.4. Classification Performances

In this experiment, the multiclass segmentation network of Unet structure was first tried. It aims to achieve three classifications of background, benign nodules, and malignant nodules at the segmentation level at one time. However, due to the nature of the image, the segmentation and classification effects are very poor. Therefore, under the condition of ensuring the performance of the segmentation network, in order to evaluate the CNN-F we designed, we have also implemented a variety of networks to extract nodule features for classification based on the morphology and edge features of the nodules. Among them, Vgg16 [28] and DenseNet [29] networks have been used by Jeong Hoon Lee in the classification of metastatic lymph nodes on CT images and have achieved excellent results [30]. The diagnostic performance of various network structures in the test data set is shown in Table 4. No matter in the comparison between Vgg16 and CNN-1, which is a shallow model, or in the comparison between DenseNet and CNN-2, which is a deep model, the single network model proposed in this experiment has advantages in performance. Moreover, the CNN-F obtained by fusing the shallow network and the deep network can clearly see that the performance is significantly improved compared to a single network model.


MethodAccuracyRecallPrecisionSpecificityF1-scoreROC

Unet (3)53.3345.7188.880.0060.350.6394
Unet + Vgg1671.1175.4987.5063.3381.050.7413
Unet + DenseNet79.2683.8188.8963.3386.280.7704
Unet + CNN-180.7486.6788.3560.0086.510.7835
Unet + CNN-282.2284.7691.7573.3388.120.7968
Unet + CNN-F85.9291.4390.5766.6791.000.8253

4. Discussion

Timely diagnosis of benign and malignant thyroid nodules plays a key role in the treatment quality and prognosis of patients. Because there are complex tissues and organs in the CT images, unreasonable feature extraction and image preprocessing will lead to classification deviation, so machine learning algorithm cannot effectively solve such problems. While deep learning can improve the accuracy of classification or prediction by building a multihidden layer machine learning model and training with massive sample data [31]. This paper used a cascaded CNN method for intelligent recognition and classification of thyroid nodules with high accuracy without any annotation processing on CT images. When addressing the problem of imbalance in the number of benign and malignant nodules, class_weight was used, and the benign and malignant weight was set to 1 : 3 to balance the loss calculation. Test time enhancement [32] was used to further enhance the accuracy of the final result, predicting the original image, horizontal mirror image, vertical mirror image, and rotating image, and the final average value was taken as the final output. After effective segmentation of thyroid nodules by Eff-Unet, CNN networks with different depths were trained by fusion. Image features of different dimensions were extracted by the trained convolution filter and they were combined and normalized to achieve the classification of thyroid nodules. Figure 7 shows the effect of some images of the test set in the system, in which the first column is the original image, the second column is manually labeled mask, the third column splits the nodule part of the network output, and the fourth column is the visualization result of the classification network output mapped to the original image. Blue represents benign nodules and red represents malignant nodules. The single nodule image can achieve clear classification, and the system can achieve accurate classification of each nodule for multiple nodule images classification.

5. Conclusion

In this work, we designed a set of end-to-end automatic recognition and classification system for thyroid nodules, which can realize the benign and malignant classification of nodules without any marking on CT images and can achieve accurate diagnosis of a single nodule, meeting the need of multinodule CT images. After comparing with the actual diagnosis report, compared with the common multiclass segmentation, we designed the segmentation network plus the classification network model, which greatly improves the accuracy of segmentation and classification. Finally, the total score of the algorithm is that the IOU score of the segmentation network is 0.856, and the overall classification performance has accuracy of 85.92%, recall of 91.43%, precision of 90.57%, specificity of 66.67%, ROC of 0.8253, and F1-score of 91.0% in this work. The overall performance of the system is shown in Table 5. The results show that the deep learning algorithm based on CNN can achieve accurate automatic diagnosis of thyroid nodules, provide doctors with reliable auxiliary report opinions, reduce the difficulty of doctors’ work, and reduce the misdiagnosis rate. However, the current data is limited to enhanced CT images, and the database capacity is not large enough to prove the generalization of the model. Therefore, in the future research, we may focus on the relationship between ordinary CT and enhanced CT and further improve the model, trying to perfect the structure of the hybrid model to improve the robustness of the model as much as possible.


IOUAccuracyRecallPrecisionSpecificityF1-scoreROC

0.8560.8590.9140.9060.6670.9100.825

Data Availability

The image data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare that they have no conflicts of interest regarding the publication of this paper.

Acknowledgments

This project was supported by grants from Zhejiang Key Research and Development Project (2019C03088).

References

  1. S. R. Bomeli, S. O. LeBeau, and R. L. Ferris, “Evaluation of a thyroid nodule,” Otolaryngologic Clinics of North America, vol. 43, no. 2, pp. 229–238, 2010. View at: Publisher Site | Google Scholar
  2. L. Davies and H. G. Welch, “Current thyroid cancer trends in the United States,” JAMA Otolaryngology-Head & Neck Surgery, vol. 140, no. 4, pp. 317–322, 2014. View at: Publisher Site | Google Scholar
  3. H. A. Nugroho, “Classification of thyroid nodules based on analysis of margin characteristic,” in 2017 International Conference on Computer, Control, Informatics and its Applications (IC3INA), IEEE, Jakarta, Indonesia, October 2017. View at: Google Scholar
  4. L. Wang, “Automatic diagnosis for thyroid nodules in ultrasound images by deep neural networks,” Medical Image Analysis, vol. 61, Article ID 101665, 2020. View at: Publisher Site | Google Scholar
  5. D. K. Iakovidis, E. G. Keramidas, and D. Maroulis, “Fusion of fuzzy statistical distributions for classification of thyroid ultrasound patterns,” Artificial Intelligence in Medicine, vol. 50, no. 1, pp. 33–41, 2010. View at: Publisher Site | Google Scholar
  6. D. N. Levin, “Retrospective geometric correlation of MR, CT, and PET images,” Radiology, vol. 169, pp. 817–823, 1988. View at: Publisher Site | Google Scholar
  7. W. Peng, “Thyroid nodule recognition in computed tomography using first order statistics,” Biomedical Engineering Online, vol. 16, no. 1, p. 67, 2017. View at: Publisher Site | Google Scholar
  8. C. Liu, S. Chen, Y. Yang et al., “The value of the computer-aided diagnosis system for thyroid lesions based on computed tomography images,” Quantitative Imaging in Medicine and Surgery, vol. 9, no. 4, pp. 642–653, 2019. View at: Publisher Site | Google Scholar
  9. Z. Zhao, “Cascade and fusion of multitask convolutional neural networks for detection of thyroid nodules in contrast-enhanced CT,” Computational Intelligence and Neuroscience, vol. 2019, 2019. View at: Publisher Site | Google Scholar
  10. E. M. Caoili, M. Korobkin, I. R. Francis et al., “Adrenal masses: characterization with combined unenhanced and delayed enhanced CT,” Radiology, vol. 222, no. 3, pp. 629–633, 2002. View at: Publisher Site | Google Scholar
  11. P. D. Faris, W. A. Ghali, R. Brant, C. M. Norris, P. D. Galbraith, and M. L. Knudtson, “Multiple imputation versus data enhancement for dealing with missing data in observational health care outcome analyses,” Journal of Clinical Epidemiology, vol. 55, no. 2, pp. 184–191, 2002. View at: Publisher Site | Google Scholar
  12. O. Ronneberger, P. Fischer, and T. Brox, “U-net: convolutional networks for biomedical image segmentation,” in Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer, Cham, Switzerland, October 2015. View at: Google Scholar
  13. O. Oktay, “Attention u-net: learning where to look for the pancreas,” 2018, https://arxiv.org/abs/1804.03999. View at: Google Scholar
  14. M. Tan and Q. V. Le, “Efficientnet: rethinking model scaling for convolutional neural networks,” 2019, https://arxiv.org/abs/1905.11946. View at: Google Scholar
  15. Z. Li and Detnet, “Design backbone for object detection,” in Proceedings of the European conference on computer vision, ECCV), Munich, Germany, September 2018. View at: Google Scholar
  16. L. T. Duong, “Automated fruit recognition using EfficientNet and MixNet,” Computers and Electronics in Agriculture, vol. 171, Article ID 105326, 2020. View at: Publisher Site | Google Scholar
  17. S. Hershey, “CNN architectures for large-scale audio classification,” in Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), IEEE, New Orleans, LA, USA, March 2017. View at: Google Scholar
  18. S. Ishigaki, K. Shimamoto, H. Satake et al., “Multi-slice CT of thyroid nodules: comparison with ultrasonography,” Radiation Medicine, vol. 22, no. 5, pp. 346–353, 2004. View at: Google Scholar
  19. C. Szegedy, “Rethinking the inception architecture for computer vision,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, June 2016. View at: Google Scholar
  20. Y. Mednikov, “Transfer representation learning using Inception-V3 for the detection of masses in mammography,” in Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), IEEE, Honolulu, HI, USA, July 2018. View at: Google Scholar
  21. T.-Y. Hsiao, Y.-C. Chang, H.-H. Chou, and C.-T. Chiu, “Filter-based deep-compression with global average pooling for convolutional networks,” Journal of Systems Architecture, vol. 95, pp. 9–18, 2019. View at: Publisher Site | Google Scholar
  22. X. Yin, “A flexible sigmoid function of determinate growth,” Annals of Botany, vol. 91, no. 3, pp. 361–371, 2003. View at: Publisher Site | Google Scholar
  23. N. Tustison and J. C. Gee, “Introducing Dice, Jaccard, and other label overlap measures to ITK,” Insight Journal, vol. 2, 2009. View at: Google Scholar
  24. V. Thada and V. Jaglan, “Comparison of jaccard, dice, cosine similarity coefficient to find best fitness value for web retrieved documents using genetic algorithm,” International Journal of Innovations in Engineering and Technology, vol. 2, no. 4, pp. 202–205, 2013. View at: Google Scholar
  25. M. Yi-de and Q. Zhi-Bai, “Automated image segmentation using improved PCNN model based on cross-entropy,” in Proceedings of 2004 International Symposium on Intelligent Multimedia, Video and Speech Processing, IEEE, Hong Kong, China, October 2004. View at: Google Scholar
  26. L. Zhou, C. Zhang, and M. Wu, “D-LinkNet: linknet with pretrained encoder and dilated convolution for high resolution satellite imagery road extraction,” in Proceedings of the CVPR Workshops, Salt Lake City, UT, USA, June 2018. View at: Google Scholar
  27. G. Ghiasi, T.-Yi Lin, and V. L. Quoc, “NAS-FPN: learning scalable feature pyramid architecture for object detection,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, June 2019. View at: Google Scholar
  28. H. Qassim, A. Verma, and D. Feinzimer, “Compressed residual-VGG16 CNN model for big data places image recognition,” in 2018 IEEE 8th Annual Computing and Communication Workshop and Conference (CCWC), IEEE, Las Vegas, NV, USA, January 2018. View at: Google Scholar
  29. F. Iandola, “Densenet: implementing efficient convnet descriptor pyramids,” 2014, https://arxiv.org/abs/1404.1869. View at: Google Scholar
  30. J. H. Lee, E. J. Ha, and Ju Han. Kim, “Application of deep learning to the diagnosis of cervical lymph node metastasis from thyroid cancer with CT,” European Radiology, vol. 29, no. 10, pp. 5452–5457, 2019. View at: Publisher Site | Google Scholar
  31. Y. Zhou and Cia-net, “Robust nuclei instance segmentation with contour-aware information aggregation,” in Proceedings of the International Conference on Information Processing in Medical Imaging, Springer, Cham, Switzerland, June 2019. View at: Google Scholar
  32. A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet classification with deep convolutional neural networks,” Communications of the ACM, vol. 60, no. 6, pp. 84–90, 2017. View at: Publisher Site | Google Scholar

Copyright © 2021 Wenjun Li et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Related articles

No related content is available yet for this article.
 PDF Download Citation Citation
 Download other formatsMore
 Order printed copiesOrder
Views700
Downloads753
Citations

Related articles

No related content is available yet for this article.

Article of the Year Award: Outstanding research contributions of 2021, as selected by our Chief Editors. Read the winning articles.