Computational and Mathematical Methods in Medicine

Computational and Mathematical Methods in Medicine / 2020 / Article

Research Article | Open Access

Volume 2020 |Article ID 6029258 | https://doi.org/10.1155/2020/6029258

Xiaodong Huang, Hui Zhang, Li Zhuo, Xiaoguang Li, Jing Zhang, "TISNet-Enhanced Fully Convolutional Network with Encoder-Decoder Structure for Tongue Image Segmentation in Traditional Chinese Medicine", Computational and Mathematical Methods in Medicine, vol. 2020, Article ID 6029258, 13 pages, 2020. https://doi.org/10.1155/2020/6029258

TISNet-Enhanced Fully Convolutional Network with Encoder-Decoder Structure for Tongue Image Segmentation in Traditional Chinese Medicine

Academic Editor: Seiya Imoto
Received08 Jan 2020
Revised07 May 2020
Accepted15 Jun 2020
Published07 Aug 2020

Abstract

Extracting the tongue body accurately from a digital tongue image is a challenge for automated tongue diagnoses, as the blurred edge of the tongue body, interference of pathological details, and the huge difference in the size and shape of the tongue. In this study, an automated tongue image segmentation method using enhanced fully convolutional network with encoder-decoder structure was presented. In the frame of the proposed network, the deep residual network was adopted as an encoder to obtain dense feature maps, and a Receptive Field Block was assembled behind the encoder. Receptive Field Block can capture adequate global contextual prior because of its structure of the multibranch convolution layers with varying kernels. Moreover, the Feature Pyramid Network was used as a decoder to fuse multiscale feature maps for gathering sufficient positional information to recover the clear contour of the tongue body. The quantitative evaluation of the segmentation results of 300 tongue images from the SIPL-tongue dataset showed that the average Hausdorff Distance, average Symmetric Mean Absolute Surface Distance, average Dice Similarity Coefficient, average precision, average sensitivity, and average specificity were 11.2963, 3.4737, 97.26%, 95.66%, 98.97%, and 98.68%, respectively. The proposed method achieved the best performance compared with the other four deep-learning-based segmentation methods (including SegNet, FCN, PSPNet, and DeepLab v3+). There were also similar results on the HIT-tongue dataset. The experimental results demonstrated that the proposed method can achieve accurate tongue image segmentation and meet the practical requirements of automated tongue diagnoses.

1. Introduction

In the field of complementary medicine, tongue diagnosis is the most active approach compared to other diagnostic methods such as palpation and pulse diagnosis [1]. Because of its simplicity, immediacy, and effectiveness, tongue diagnosis is widely used in Traditional Chinese Medicine (TCM), Japanese traditional herbal medicine, and Traditional Korean Medicine (TKM) [2].

Tongue diagnosis means that the doctors judge the status of a patient’s internal organs by the visual information such as the color, form, substance, and coating of the tongue. The diagnostic results heavily rely on the doctor’s experiences. Moreover, the inspecting circumstances, such as illumination, also affects the judgment of doctors. So, it is difficult for tongue diagnosis to obtain objective and standardized results. In the past decades, researchers have developed various types of computer-aided tongue diagnosis systems (CATDS) based on computer vision and machine learning to address these problems in many countries [36]. The schematic diagram of a typical CATDS is shown in Figure 1 [7]. The system is mainly composed of four ordered modules: tongue image acquisition, image preprocessing, feature analysis, and diagnosis. Plenty of research have already been carried out on each module.

Generally, tongue images acquired by CATDSs include some nontongue regions, such as the face, teeth, and lips. However, the results of tongue diagnosis are only related to the tongue body according to the theory of traditional oriental medicine. So, it is an essential step to segment the tongue body from a tongue image in CATDSs. Furthermore, high accuracy is hugely demanded in tongue image segmentation because the accurate results of the tongue feature analysis mainly depend on the result of segmentation. This close relationship should be emphasized because pathological information may exist on any part of the tongue body, for example, teeth marks, which is one of the important tongue features with rich disease-related information, usually occur at the edge of the tongue body. For automated tongue image segmentation, there are still some factors that make it challenging to extract tongue body accurately and robustly. The specific cases are shown in Figure 2. (i)Different shooting angle and circumstance and wide variations of patients in the shape and size of the tongue body(ii)The variation in tongue color, texture, and shape due to the pathology of the tongue(iii)A blurred edge of the tongue body caused by a similar color between the tongue body and the lips.

During the past decades, researchers have developed many methods to segment tongue body from digital tongue images effectively. Generally, these methods can be divided into two categories: methods based on traditional image processing technology and methods based on deep learning.

In the first category, some methods employed the techniques of the adaptive threshold, region growing and merging, and edge detection to segment the tongue body from tongue images [810]. However, these segmentation methods just took the low-level visual information (such as color, edge, and texture) into account and often failed to extract the tongue body from the surroundings completely. They ignored high-level information such as shape attribute, which is an indispensable factor for this segmentation task. Therefore, the other methods which introduce high-level information were proposed to achieve better performance. The active contour model (ACM) [11] was the most widely used method based on high-level information in the field of tongue image segmentation in the first category, which is also known as Snakes. According to the initialization methods and the strategy of curve evolution, there were many variants of the ACM-based algorithms. Concerning the initialization, Pang et al. [12] applied the bielliptical deformable template to obtain the initial evolving contour. Zuo et al. [13] used a polar edge detector to detect the initial boundary of the tongue body. Furthermore, Wu et al. [14] applied the watershed transformation to obtain the initial contour of the snake. Regarding the curve evolution, Yu et al. [15] extracted the tongue body by the addition of a color gradient to the gradient vector flow (GVF) snake. Shi et al. [16] applied double geodesic flow to extract tongue body based on the prior information of tongue shape and location. Later, Shi et al. [17] continued the work with the color control-geometric and gradient flow snake algorithm-enhanced curve velocity. However, the traditional methods still could not meet the demands of accuracy and robustness simultaneously, which are essential for the following automated analysis of the tongue features.

Since 2012, the convolutional neural network (CNN) has made significant progress in various fields of computer vision. Some researchers also tried to apply CNNs to tongue image segmentation [1820]. In these studies, the methods based on deep learning could get better results than traditional methods. However, it was still difficult for these methods to get accurate segmentation results because of their simple network architecture, when they deal with similar cases shown in Figure 2.

After Shelhamer et al. [21] introduced the fully convolutional neural network to the field of image segmentation, various new models emerged and achieved remarkable results in the field of image segmentation. There were still several factors that affect the accuracy of segmentation results. The first one is related to the fact that the effective receptive field on high-level layers of CNN is much smaller than the theoretical one, especially on high-level layers [22], which may make many networks only “see” a small part of the entire image and be unable to integrate the sufficient contextual prior. Some methods employed spatial pyramid pooling to capture multiscale context to address the problem. DeepLab v3+ [23] used Atrous Spatial Pyramid Pooling (ASPP) to capture contextual information. In [24], ASPP was replaced by the Vortex Pooling Module, which can aggregate features around the target position more efficiently by assigning different attention. Pyramid scene parsing net [25] introduced the pyramid pooling module to contain information with different scales. All of the above works demonstrated outstanding performance. The second factor is caused by the inherent limitation of CNN. In the task of image classification, deeper networks with multiple down-sampling by max-pooling or convolution operation () have proven to be most successful. However, for image segmentation, higher layer feature maps usually contain more semantic meanings and yet lose more positional information inevitably, which may cause rough object boundaries. Meanwhile, clear and sharp object boundaries are critical in some applications. The researchers have proceeded in two directions to deal with this problem [25]. The first one was the structure prediction. In researches [2628], the conditional random field (CRF) was adopted to improve the segmentation result. The other direction was to fuse feature maps from different intermediate layers in CNNs to predict the object boundaries better. The skip connection or encoder-decoder structure was the primary form of multiscale feature map fusion. DeepLab v3+ and RefineNet [29], which use the encoder-decoder structure, showed effectiveness in the field of image segmentation.

Inspired by these works, an automated tongue image segmentation network named TISNet was presented, which was based on the fully convolutional network with encoder-decoder structure. The ResNet101 [30] was adopted as an encoder to obtain dense feature maps, and the Receptive Field Block (RFB) [31] was assembled behind the encoder to integrate adequate global contextual prior. Moreover, a Feature Pyramid Network (FPN) [32] was used as a decoder to fuse multiscale feature maps to recover the clear and sharp boundary of the tongue body.

2. Methods

As mentioned above, for a well-performance segmentation network, it is significant to have a large receptive field for integrating adequate global contextual prior and maintain more positional information at the same time. In the TISNet, Receptive Field Block was introduced to integrate adequate global contextual prior, and Feature Pyramid Network was adopted to gather sufficient positional information to cover the contour of the tongue body. Next, the structures of RFB, FPN, and the architecture of TISNet were described in detail.

2.1. Receptive Field Block

Currently, in several state-of-art segmentation networks, a particular module is designed to obtain global contextual information for more accurate segmentation results [2325, 33, 34]. However, these modules commonly set the receptive field of the same size on the feature map with a regular sampling grid, which may cause some loss in the robustness and discernibility of the feature. For example, the Atrous Spatial Pyramid Pooling (ASPP) module used in DeepLab v3+ is employed to capture the multiscale context. The ASPP module, which has several parallel branches of convolutional layers with different atrous rates, is applied on the top feature maps. Though the ASPP module can change the sampling distance from the center of feature maps, resulting features captured by it still have a uniform resolution and tend to be less distinctive than daisy-shaped ones. There are the same problems in Pyramid Pooling Module [25] and Vortex Pooling Module [34].

In the field of neuroscience, researchers find that the size of population Receptive Field (pRF) is a function of eccentricity in the retinotopic maps in the human visual system, and the size of pRF increases with eccentricity in spite of varying in the maps. This mechanism means that the region closer to the center of retinotopic maps is more important than others in distinguishing object [35].

Inspired by the mechanism of pRF in the human visual system, a module named Receptive Field Block was introduced to strengthen global contextual prior information learned from CNN in this study, which is helpful for accurate image segmentation. Figure 3 illustrates the details of the internal structure of RFB. The RFB includes a multibranch convolutional block and a shortcut connection. The convolutional block consists four branches, and each branch has a convolutional layer with a particular kernel size and a corresponding atrous convolutional layer with the same value in dilation rate. The functional relation between the kernel size and dilation rate is similar to that of the size and eccentricity of pRFs in the human visual system. The ordinary convolutional layer generates the multisize pRF, and the atrous convolutional layer simulates the human visual system to build the relationship between the size and eccentricity of pRF. Finally, all feature maps from the different branches are merged into one convolutional layer and pixel-wise added with the shortcut connection branch.

2.2. Feature Pyramid Network

Before pixel prediction has arisen, segmentation models generally adopt CNN models as a backbone to extract the feature maps. However, these CNN models transferred from the tasks of image classification usually lose more positional information due to CNN invariance, which causes that segmentation network cannot obtain accurate object boundaries. Therefore, for better performance, some of the new methods of image segmentation attend to use the different layers in a CNN [21, 36] or concatenate the feature maps from multiple layers before final mask predictions [37, 38]. Also, some methods associate the high-level semantic information with low-level feature maps though lateral/skip connections to recover the positional information [39, 40].

In this paper, the Feature Pyramid Network (FPN) was introduced to gather more positional information for obtaining fine object boundaries. Giving a single-scale image of arbitrary size, the FPN can generate a set of feature maps with proportionally sized at multilevel. As shown in Figure 4, the structure of FPN consists three components: a bottom-up pathway, a top-down pathway, and lateral connections. The details of FPN would be explained in the next section.

2.3. Tongue Image Segmentation Network

As shown in Figure 5, the TISNet consists four parts: backbone module, RFB module, FPN module, and final prediction module. The ResNet101 is adopted as a backbone to extract the feature maps, and the feature maps output by each stage’s last residual block are used. The feature maps of the last residual blocks of conv2, conv3, conv4, and conv5 are denoted as C2, C3, C4, and C5, which correspond to the strides of 4, 8, 16, and 32 pixels to the input image. The conv1 is not included because of the large requirement of memory.

Behind the backbone, the RFB module is applied to integrate adequate global contextual prior. The structure of the RFB module has been described in detail in Section 3.1. By four branches, the RFB module can cover the whole, half, and small portions of the image. The feature maps from all the branches are fused and then fed into the FPN. The details of the FPN is shown in the box of FPN. It fuses high-level semantic information with the middle- and low-level features by lateral connections in the top-down path. The coarser-resolution feature maps are up-sampled by a factor of 2 and then merged with the corresponding feature maps from the backbone by element-wise addition. Before merging, the feature maps from backbone go though a convolutional layer to reduce channel dimensions. Then, a convolution operation is applied to the merged map to reduce the aliasing effect of up-sampling before producing the final feature maps of this level. This process iterates until the highest resolution feature maps are generated. The set of final feature maps are denoted as P2, P3, P4, and P5, which correspond to C2, C3, C4, and C5, and have the same spatial sizes. Also, the number of channels of all final feature maps in each level is set to 256 for simplifying the computational complexity.

In the final prediction module, a strategy of fusion of multiscale feature maps is applied to recover the contour details instead of only using the feature map with the highest resolution (P2). The final set of feature maps P3, P4, and P5 is first bilinearly up-sampled to the same size of P2, and then all of the four feature maps are concatenated. After the concatenation, a convolution is appended on the feature maps, followed by another simple bilinear up-sampled by a factor of 4 to obtain the final mask.

3. Experiments and Results

In this section, the performance of TISNet was evaluated on two tongue image datasets: HIT-tongue dataset and SIPL-tongue dataset. All experiments were conducted on the public platform PyTorch 0.4.1 on the Windows system. The system configuration is Intel (R) i5-7500 CPU @ 3.4GHz with 16 G memory and NVIDIA Titan XP graphics card.

3.1. Implementation Details
3.1.1. Datasets

(1)HIT-Tongue Dataset. This dataset is an open tongue image dataset from the Harbin Institute of Technology. The dataset contains 300 RGB tongue images, and the corresponding manual masks as the ground truth. All the tongue images are acquired in a semienclosed environment under stable lighting conditions. The image size is .

(2) SIPL-Tongue Dataset. This dataset is constructed exclusively for automated tongue image analysis by ourselves. All the images are acquired by a self-designed acquisition device in a closed environment and carefully selected and labelled by the experts in Chinese Traditional Medicine from Beijing Xuanwu Hospital. In the SIPL-tongue dataset, there are 700 RGB tongue images and the corresponding well-labelled masks for tongue image segmentation. The image size is .

3.1.2. Data Augmentation

The research suggests that the number of training samples is critical to the performance of CNN. For a small-scale dataset, artificial data augmentation is a common method to generate enough training samples. Because of the limited scale of the tongue image dataset, the strategies, including random Horizontal Flip, random resize between 0.5 and 2, random rotation between -10 and 10 degrees, and random Gaussian blur, were adopted for data augmentation in this study.

3.1.3. Training Parameters

In this study, the weights of pretrained Resnet101 were directly employed in the backbone module of the TISNet. Then TISNet was trained on the PASCAL VOC 2012 semantic segmentation benchmark [41] and was fine-tuned, respectively, on two tongue image datasets. Table 1 shows the parameters used in the training step.


HyperparametersInterpretation

NormalizationMean centering and standard deviation normalization of the intensities were performed
CroppingAll the images were center-cropped to a pixel size
OptimizationSGD optimizer with the base learning rate 0.01
Learning rate schedulingThe poly learning rate policy was used, where the current learning rate equals to the base one multiplying
Batch size4
Epoch number size50
Momentum0.9
Weight decay0.0005

3.2. Evaluation Criteria

In this study, there are six metrics used to evaluate the performance of TISNet. Table 2 gives a detailed description of the six metrics.


Metric nameAbbr.RangeInterpretationCategory

Dice similarity coefficientDSC0–1Similarity between masksOverlap
Hausdorff distanceHD>0Longest Euclidean distance between mask contours (absolute error)Distance
Symmetric mean absoluteMSD>0Mean Euclidean distance between mask contours (mean error)Distance
Surface distance precisionPPV0–1Low values mean that the method tends to over segmentStatistical
SensitivityTPR0–1Low values mean that the method tends to under segmentStatistical
SpecificityTNR0–1Quality of segmented backgroundStatistical

3.3. Experimental Results

To validate the effectiveness of the proposed method, a comparison of performance was made between our method and the other four deep-learning-based methods (including FCN, SegNet, PSPNet, and DeepLab v3+) on the HIT-tongue dataset and SIPL-tongue dataset. For a fair comparison, these four models have the same training hyperparameters as our model and follow the same training procedure during experiments. Among these methods, FCN is the first work that trains a fully convolutional network end-to-end for semantic segmentation and exceeds the state-of-the-art at that time. Many follow-up excellent works are based on FCN. SegNet is a model with the encoder-decoder structure for pixel-wise image segmentation and obtains a good balance between accuracy and requirement of memory. PSPNet embeds a pyramid pooling module into the scene parsing network to aggregate region-based context and achieves state-of-the-art performance on various datasets by exploiting the capability of global context information. DeepLab v3+ combines the advantages of spatial pyramid pooling module and encoder-decoder structure and obtains the best results on the PASCAL VOC 2012 semantic segmentation benchmark [41] and Cityscapes datasets [42].

3.3.1. Experiments on HIT-Tongue Dataset

A comparison of the performance of five methods was conducted on the HIT-tongue dataset, and Table 3 shows the quantitative results. It can be seen that TISNet gets the best results in terms of DSC, HD, MSD, accuracy, and specificity. Also, Figure 6 exhibits the segmentation results of six typical cases from the HIT-tongue dataset. Rows 1 to 4 are the four cases in which the tongue body is with different size and shape. Rows 5 and 6 give two cases in which the color of the tongue body is similar to the lip, and the interference from the teeth is also visible. The segmentation results show that all deep models can locate the tongue body regardless of the difference in size and shape. Although FCN, SegNet, and PSPNet can obtain the smooth contour, the segmentation results are not accurate, usually accompanied by oversegmentation or undersegmentation, especially FCN. Instead, DeepLab v3+ and TISNet can retain more details and present clear and sharp contours.


MethodDSCHD (pixel)MSD (pixel)PrecisionSensitivitySpecificity

SegNet
FCN
PSPNet
DeepLab v3+
Ours

3.3.2. Experiments on SIPL-Tongue Dataset

Table 4 lists the quantitative results obtained by different methods. It can be seen that TISNet can get the best results on 5 out of 6 metrics and can also perform well in terms of sensitivity. TISNet has the lowest value in DSC, HD, and MSD along with small standard deviations, and this indicates that TISNet can obtain a more accurate contour of the tongue body. It also is confirmed by qualitative results shown in Figure 7, where the segmentation results of six cases from the SIPL-tongue dataset are illustrated. Rows 1 and 2 are two cases in which the tongue body is with different size and shape. Rows 3 and 4 show two cases in which there are severe fissures, tooth marks, and coating on the surface of the pathological tongue. Rows 5 and 6 are two cases in which the tongue body has blurred contour between the tongue body and the lip along with the interference from the teeth. The tongue images from the SIPL-tongue dataset are more diverse and complicated compared with the HIT-tongue dataset, which can test the performance of the model more effectively. From Figure 7, FCN and SegNet all fail to extract the tongue body in case 1, and there are oversegmentation or undersegmentation in the other five cases. Because of more complicated and powerful models, PSPNet and DeepLab v3+ get better results than FCN and SegNet, but they still cannot segment the tongue body accurately, and especially, there are obvious oversegmentation in case 5 and case 6. It means that PSPNet and DeepLab v3+ cannot deal with the problem of the blurred edge well. Instead, TISNet solves the problem better and gets the finest segmentation results, which principally benefits from the elaborately designed network structure.


MethodDSCHD (pixel)MSD (pixel)PrecisionSensitivitySpecificity

SegNet
FCN
PSPNet
DeepLab v3+
Ours

3.3.3. Ablation Study

In this study, RFB, FPN, and multiscale fusion strategy are employed to improve the performance of the segmentation network. The ablation study is conducted on the SIPL-tongue dataset to validate the effectiveness of these modules and strategies. Table 5 shows the results of the ablation study. It should be noted that the “+” means adding a new module or strategy based on the last row instead of the baseline. The first row in Table 5 contains the results of the baseline. The baseline uses the Resnet101 pretrained on ImageNet as a backbone. A convolution layer with channel dimension 2 is embedded on top of the backbone to classify the pixels, and then the coarse output is up-sampled to produce the final segmentation results. Because the size of the final outputs of the backbone is 1/32 of the input image, the scale of details in the final prediction is limited, which leads to the worst results. The second row in Table 5 shows the results when adding the RFB before the final prediction occurs. As we can see, the performance is overall improved, and especially, the HD is decreased from 17.0623 to 13.8587. Then, the FPN is embedded in the previous structure, which further improves the performance in all six metrics. Finally, after the multiscale fusion strategy is added, the best results are yielded and listed in the fourth row in Table 5. It is concluded that these modules and strategies reinforce the performance of the network for tongue image segmentation.


MethodDSCH D(pixel)MSD (pixel)PrecisionSensitivitySpecificity

Baseline
+RFB
+FPN
+MS

+: adding a new module or strategy based on the last row instead of the baseline. RFB: embedding RFB block into the segmentation network. FPN: employing the FPN structure. MS: fusing multiscale feature maps before final pixel prediction.

4. Discussion and Conclusion

In this study, a new tongue image segmentation method named TISNet was proposed, and it achieved the best performance on two tongue image datasets compared with the other four deep-learning-based methods. The qualitative and quantitative evaluation showed that the proposed method met the practical requirements of automated tongue image segmentation and could be embedded in computer-aided tongue diagnosis systems.

According to the experimental results, it can be concluded that the deep-learning-based methods are robust to the variation in the size and shape of the tongue body, as well as pathological details on the surface of the tongue. For the method of tongue image segmentation, the key point is to deal with the problem of the blurred edge of the tongue body to obtain a sharp and clear contour. An elaborately designed model is critical.

In this study, three factors make TISNet achieve the best performance. Firstly, the ResNet101 model is adopted as a backbone to extract the dense feature maps, which is also used in PSPNet and DeepLab v3+. The ResNet101 is more potent in feature extraction and representation than VGG16 [43] used as the backbone in FCN and SegNet. It is one of the reasons that TISNet, DeepLab v3+, and PSPNet can perform better than FCN and SegNet in total. Secondly, RFB is introduced behind the backbone to integrate adequate global contextual prior. Moreover, because the structure of the RFB simulates the pRF in the human vision system, the feature extracted by RFB is more distinctive than the ones obtained by PPM in PSPNet and by ASPP in DeepLab v3+. Finally, FPN is applied to produce a fine mask. FFN is a typical encoder-decoder structure and can gather sufficient positional information by fusing multiscale feature maps, and this positional information is much valuable for recovering the accurate contour of the tongue body.

In the future, other than applying the proposed method to a more number of clinical data, our research will focus on making the model light-weight and efficient to run it in mobile devices.

Data Availability

The HIT-tongue dataset used in this research can be taken from https://github.com/BioHit/TongeImageDataset. The SIPL-tongue dataset used in this research has not been allowed to be public now because this dataset comes from an ongoing collaborative project and is being used for other research at the same time.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

The authors thank Dr. Wenqiang Chen and Dr. Wenjing Chen from Beijing Xuanwu Hospital for their help in capturing the tongue images and guiding in Traditional Chinese Medicine. This study is supported by National Natural Science Foundation of China (61871006).

References

  1. Y. Q. Liu, Y. X. Wang, N. N. Shi, X. J. Han, and A. P. Lu, “Current situation of international organization for standardization/technical committee 249 international standards of traditional Chinese medicine,” Chinese Journal of Integrative Medicine, vol. 23, no. 5, pp. 376–380, 2017. View at: Publisher Site | Google Scholar
  2. M. H. Tania, K. Lwin, and M. A. Hossain, “Advances in automated tongue diagnosis techniques,” Integrative Medicine Research, vol. 8, no. 1, pp. 42–56, 2019. View at: Publisher Site | Google Scholar
  3. C. C. Chiu, “A novel approach based on computerized image analysis for traditional Chinese medical diagnosis of the tongue,” Computer Methods and Programs in Biomedicine, vol. 61, no. 2, pp. 77–89, 2000. View at: Publisher Site | Google Scholar
  4. T. Oji, T. Namiki, T. Nakaguchi et al., “Study of factors involved in tongue color diagnosis by Kampo medical practitioners using the Farnsworth-Munsell 100 hue test and tongue color images,” Evidence-based Complementary and Alternative Medicine, vol. 2014, Article ID 783102, 9 pages, 2014. View at: Publisher Site | Google Scholar
  5. Y. G. Wang, Y. Zhou, J. Yang, and Q. Xu, “An image analysis system for tongue diagnosis in traditional chinese medicine,” in Computational and Information Science. CIS 2004, J. Zhang, J. H. He, and Y. Fu, Eds., vol. 3314 of Lecture Notes in Computer Science, pp. 1181–1186, Springer, Berlin, Heidelberg, 2004. View at: Publisher Site | Google Scholar
  6. J. Kim, G.-J. Han, B.-H. Choi et al., “Development of differential criteria on tongue coating thickness in tongue diagnosis,” Complementary Therapies in Medicine, vol. 20, no. 5, pp. 316–322, 2012. View at: Publisher Site | Google Scholar
  7. D. Zhang, H. Z. Zhang, and B. Zhang, Tongue Image Analysis, Springer, Singapore, 2017.
  8. J. F. Ning, D. Zhang, C. Wu, and F. Yue, “Automatic tongue image segmentation based on gradient vector flow and region merging,” Neural Computing and Applications, vol. 21, no. 8, pp. 1819–1826, 2012. View at: Publisher Site | Google Scholar
  9. C. Li, D. Wang, Y. Liu, X. Gao, and H. Shang, “A novel automatic tongue image segmentation algorithm: Color enhancement method based on Lab color space,” in IEEE International Conference on Bioinformatics and Biomedicine, pp. 990–993, Washington, DC, USA, November 2015. View at: Publisher Site | Google Scholar
  10. K. Wu and D. Zhang, “Robust tongue segmentation by fusing region-based and edge-based approaches,” Expert Systems with Applications, vol. 42, no. 21, pp. 8027–8038, 2015. View at: Publisher Site | Google Scholar
  11. M. Kass, A. Witkin, and D. Terzopoulos, “Snakes: active contour models,” International Journal of Computer Vision, vol. 1, no. 4, pp. 321–331, 1988. View at: Publisher Site | Google Scholar
  12. B. Pang, D. Zhang, and K. Wang, “The bi-elliptical deformable contour and its application to automated tongue segmentation in Chinese medicine,” IEEE Transactions on Medical Imaging, vol. 24, no. 8, pp. 946–956, 2005. View at: Publisher Site | Google Scholar
  13. W. Zuo, K. Wang, D. Zhang, and H. Zhang, “Combination of polar edge detection and active contour model for automated tongue segmentation,” in Third International Conference on Image and Graphics (ICIG'04), pp. 270–273, Hong Kong, 2004. View at: Publisher Site | Google Scholar
  14. J. Wu, Y. Zhang, and J. Bai, “Tongue area extraction in tongue diagnosis of traditional chinese medicine,” in 2005 IEEE Engineering in Medicine and Biology 27th Annual Conference, pp. 4955–4957, Shanghai, China, 2005. View at: Publisher Site | Google Scholar
  15. S. Yu, J. Yang, Y. Wang, and Y. Zhang, “Color active contour models based tongue segmentation in traditional chinese medicine,” in 2007 1st International Conference on Bioinformatics and Biomedical Engineering, pp. 1065–1068, Wuhan, China, 2007. View at: Publisher Site | Google Scholar
  16. M. Shi, G. Z. Li, F. Li, and X. Chao, “A novel tongue segmentation approach utilizing double geodesic flow,” in 2012 7th International Conference on Computer Science & Education (ICCSE), pp. 21–25, Melbourne, VIC, Australia, 2012. View at: Publisher Site | Google Scholar
  17. M. J. Shi, G. Li, and F. Li, “C2g2fsnake: automatic tongue image segmentation utilizing prior knowledge,” Science China Information Sciences, vol. 56, no. 9, pp. 1–14, 2013. View at: Publisher Site | Google Scholar
  18. L. Jiang, B. Xu, X. Ban, T. Ping, and B. Ma, “A tongue image segmentation method based on enhanced hsv convolutional neural network,” in Cooperative Design, Visualization, and Engineering. CDVE 2017, Y. Luo, Ed., vol. 10451 of Lecture Notes in Computer Science, pp. 252–260, Springer, Cham, 2017. View at: Publisher Site | Google Scholar
  19. P. Qu, Z. Hui, Z. Li, Z. Jing, and G. Chen, “Automatic tongue image segmentation for traditional chinese medicine using deep neural network,” in Intelligent Computing Theories and Application. ICIC 2017, D. S. Huang, V. Bevilacqua, P. Premaratne, and P. Gupta, Eds., vol. 10361 of Lecture Notes in Computer Science, pp. 247–259, Springer, Cham, 2017. View at: Publisher Site | Google Scholar
  20. B. Lin, J. Xle, C. Li, and Y. Qu, “Deeptongue: tongue segmentation via resnet,” in 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1035–1039, Calgary, AB, Canada, 2018. View at: Publisher Site | Google Scholar
  21. J. Long, E. Shelhamer, and T. Darrell, “Fully convolutional cetworks for semantic segmentation,” 2014, https://arxiv.org/abs/1411.4038. View at: Google Scholar
  22. W. Luo, Y. Li, R. Urtasun, and R. Zemel, “Understanding the effective receptive field in deep convolutional neural networks,” 2014, https://arxiv.org/pdf/1701.04128.pdf. View at: Google Scholar
  23. L. Chen, Y. Zhu, G. Papandreou, F. Schroff, and H. Adam, “Encoder-decoder with atrous separable convolution for semantic image segmentation,” in Computer Vision – ECCV 2018. ECCV 2018, V. Ferrari, M. Hebert, C. Sminchisescu, and Y. Weiss, Eds., vol. 11211 of Lecture Notes in Computer Science, pp. 833–851, Springer, Cham, 2018. View at: Publisher Site | Google Scholar
  24. C. Szegedy, W. Liu, Y. Jia et al., “Going deeper with convolutions,” in 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1–9, Boston, MA, USA, 2015. View at: Publisher Site | Google Scholar
  25. H. Zhao, J. Shi, X. Qi, X. Wang, and J. Jia, “Pyramid scene parsing network,” in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6230–6239, Honolulu, HI, USA, 2018. View at: Publisher Site | Google Scholar
  26. B. Zhou, A. Khosla, A. Lapedriza, A. Oliva, and A. Torralba, “Object detectors emerge in deep scene CNNs,” 2014, https://arxiv.org/abs/1412.6856. View at: Google Scholar
  27. A. Arnab, S. Jayasumana, S. Zheng, and P. H. S. Torr, “Higher order conditional random fields in deep neural networks,” in Computer Vision – ECCV 2016. ECCV 2016, B. Leibe, J. Matas, N. Sebe, and M. Welling, Eds., vol. 9906 of Lecture Notes in Computer Science, pp. 524–540, Springer, Cham, 2016. View at: Publisher Site | Google Scholar
  28. Z. W. Liu, X. X. Li, L. Ping, C. G. Chen, and X. O. Tang, “Semantic image segmentation via deep parsing network,” in 2015 IEEE International Conference on Computer Vision (ICCV), pp. 1377–1385, Santiago, Chile, 2015. View at: Publisher Site | Google Scholar
  29. G. S. Lin, A. Milan, C. H. Shen, and I. D. Reid, “Refinenet: multi-path refinement networks for high-resolution semantic segmentation,” in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5168–5177, Honolulu, HI, USA, 2017. View at: Publisher Site | Google Scholar
  30. K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770–778, Las Vegas, NV, USA, 2016. View at: Publisher Site | Google Scholar
  31. S. T. Liu, D. Huang, and Y. H. Wang, “Receptive field block net for accurate and fast object detection,” in Computer Vision – ECCV 2018. ECCV 2018, V. Ferrari, M. Hebert, C. Sminchisescu, and Y. Weiss, Eds., vol. 11215 of Lecture Notes in Computer Science, pp. 404–419, Springer, Cham, 2018. View at: Publisher Site | Google Scholar
  32. T. Y. Lin, P. Dollár, R. Girshick, K. M. He, B. Hariharan, and S. Belongie, “Feature pyramid networks for object detection,” in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 936–944, Honolulu, HI, USA, 2017. View at: Publisher Site | Google Scholar
  33. L. Chen, G. Papandreou, F. Schroff, and H. Adam, “Rethinking atrous convolution for semantic image segmentation,” 2017, https://arxiv.org/abs/1706.05587. View at: Google Scholar
  34. C. W. Xie, H. Y. Zhou, and J. X. Wu, “Vortex pooling: improving context representation in semantic segmentation,” 2018, https://arxiv.org/abs/1804.06242. View at: Google Scholar
  35. B. A. Wandell and J. Winawer, “Computational neuroimaging and population receptive fields,” Trends in Cognitive Sciences, vol. 19, no. 6, pp. 349–357, 2015. View at: Publisher Site | Google Scholar
  36. B. Hariharan, P. Arbeláez, R. Girshick, and J. Malik, “Hypercolumns for object segmentation and fine-grained localization,” in 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR),, pp. 447–456, Boston, MA, USA, 2015. View at: Publisher Site | Google Scholar
  37. T. Kong, A. Yao, Y. Chen, and F. Sun, “Hypernet: towards accurate region proposal generation and joint object detection,” in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 845–853, Las Vegas, NV, USA, 2016. View at: Publisher Site | Google Scholar
  38. W. Liu, A. Rabinovich, and A. C. Berg, “ParseNet: looking wider to see better,” 2015, https://arxiv.org/abs/1506.04579. View at: Google Scholar
  39. O. Ronneberger, P. Fischer, and T. Brox, “U-Net: convolutional networks for biomedical image segmentation,” 2015, https://arxiv.org/abs/1505.04597. View at: Google Scholar
  40. V. Badrinarayanan, A. Kendall, and R. Cipolla, “SegNet: a deep convolutional encoder-decoder architecture for image segmentation,” 2015, https://arxiv.org/abs/1511.00561. View at: Google Scholar
  41. M. Everingham, L. Van Gool, C. Williams, J. Winn, and A. Zisserman, “The Pascal visual object classes (VOC) challenge,” International Journal of Computer Vision, vol. 88, no. 2, pp. 303–338, 2010. View at: Publisher Site | Google Scholar
  42. M. Cordts, M. Omran, S. Ramos et al., “The cityscapes dataset for semantic urban scene understanding,” in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3213–3223, Las Vegas, NV, USA, 2016. View at: Publisher Site | Google Scholar
  43. K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” 2014, https://arxiv.org/abs/1409.1556. View at: Google Scholar

Copyright © 2020 Xiaodong Huang et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


More related articles

 PDF Download Citation Citation
 Download other formatsMore
 Order printed copiesOrder
Views457
Downloads346
Citations

Related articles

Article of the Year Award: Outstanding research contributions of 2020, as selected by our Chief Editors. Read the winning articles.