Wireless Communications and Mobile Computing

Wireless Communications and Mobile Computing / 2020 / Article
Special Issue

Mobile Intelligence Assisted by Data Analytics and Cognitive Computing 2020

View this Special Issue

Research Article | Open Access

Volume 2020 |Article ID 9595687 | https://doi.org/10.1155/2020/9595687

Zhiqiang Tian, Jingyi Song, Chenyang Zhang, Xiaohui Tian, Zhong Shi, Xiaofu Yu, "A Multiscale-Based Adjustable Convolutional Neural Network for Multiple Organ Segmentation", Wireless Communications and Mobile Computing, vol. 2020, Article ID 9595687, 13 pages, 2020. https://doi.org/10.1155/2020/9595687

A Multiscale-Based Adjustable Convolutional Neural Network for Multiple Organ Segmentation

Academic Editor: Yin Zhang
Received07 Feb 2020
Revised16 Jun 2020
Accepted21 Jul 2020
Published03 Aug 2020

Abstract

Accurate segmentation ofs organs-at-risk (OARs) in computed tomography (CT) is the key to planning treatment in radiation therapy (RT). Manually delineating OARs over hundreds of images of a typical CT scan can be time-consuming and error-prone. Deep convolutional neural networks with specific structures like U-Net have been proven effective for medical image segmentation. In this work, we propose an end-to-end deep neural network for multiorgan segmentation with higher accuracy and lower complexity. Compared with several state-of-the-art methods, the proposed accuracy-complexity adjustment module (ACAM) can increase segmentation accuracy and reduce the model complexity and memory usage simultaneously. An attention-based multiscale aggregation module (MAM) is also proposed for further improvement. Experiment results on chest CT datasets show that the proposed network achieves competitive Dice similarity coefficient results with fewer float-point operations (FLOPs) for multiple organs, which outperforms several state-of-the-art methods.

1. Introduction

Radiation therapy is the main clinical method of treating various cancers [30]. It can be seen as a trade-off between sending maximum dose to the target-volume (TV) and minimum dose to the OARs [1, 31]. Accurate segmentation of OARs contributes to the protection of normal organs during treatment [32]. Manually delineating OARs slice-by-slice in CT scans requires expertise and lots of time. Adjacent soft tissue with low contrast and noise blurs the boundaries of organs, which may also lead to delineation errors. Therefore, automatic segmentation has become a research hotspot [33]. Such a framework could help radiologists segment OARs more accurately with much less time.

Deep convolutional models have shown state-of-the-art performance in image segmentation especially after fully convolutional networks (FCN) [2] were proposed. In FCN, fully convolutional layers are substituted for fully connected layers in a classic classification network to reserve spatial information, which makes FCN a reasonable choice for dense prediction. Based on FCN, Badrinarayanan et al. [3] introduced a concept of encoders and decoders and upsampled the low-level inputs in turn to the original resolution, which achieves impressive performance in scene understanding applications. U-Net [4] adopted the similar encode and decode paths and used several skip connections between them to combine the low-level and high-level features. Thanks to this carefully designed structure, it achieves considerable success in medical image segmentation tasks. In addition, great efforts have been subsequently made to improve the performance of U-Net [5, 6].

Targets in images usually have various sizes. It is crucial to let a network “see” objects of different scales. What kind of multiscale information to use and how to integrate information are the purpose of this work. PSPNet [7] merged convolutional features from different region-based context aggregations. DeepLabs [811] similarly introduced a spatial pyramid pooling with different dilated kernels. An attention mechanism was first proposed to model long-range dependencies in machine translation and has been successfully transferred to computer vision tasks. It allocates more weights to the potentially interesting regions and suppresses the less informative ones. DANet [12] built a dual attention module to simultaneously model channel-wise and spatial-wise semantics. Oktay et al. [5] proposed an attention gate in the skip connection layer of the U-Net to teach the decoder where to “look.” Selective Kernel [13] (SK) took advantage of the attention to guide the fusion of multiscale information. It embeds multiple lightweight SE [14] attentions to dynamically select the receptive field size of each neuron in convolutional layers, which is also consistent with the neuroscience cognition that the visual cortical neuron adaptively adjusts the size of its receptive field according to stimulus. In such case, the learned attention maps play the role of a stimulus.

Aside from accuracy, computation complexity is another factor to measure model performance. A widely used evaluation metric of computation complexity is the number of float-point operations, also known as FLOPs [15]. Substantial efforts have been made to reduce FLOPs of convolutional networks. MobileNets [16, 17] adopted depth-wise separable convolution, which splits a standard convolution into a depth-wise and a point-wise convolution. It can greatly reduce the computation cost. ShuffleNets [18, 19] further integrated group convolution and reduced channel redundancy in point-wise connectivity. These methods focus on sparse connection between channels. In contrast, Octave Convolution [20] (OctConv) focused on the spatial dimension and argued that potential redundancy may exist. Each location independently stores its own features, ignoring the possibility that there could be similar information between adjacent locations. OctConv presented a multifrequency processing algorithm, which splits feature maps into high and low frequency to reduce spatial redundancy. Through sharing information with adjacent neurons, the spatial resolution of the low-frequency parts can be reduced, which helps to save computation and memory cost.

A highly generalized CNN should take the computing power and memory consumption into consideration or otherwise may fail to run on the limited computation conditions, especially the training stage [21]. Several strategies have been proposed to alleviate this issue. A most straightforward method is to reduce batch size during training. Heavy models like V-Net [22] and No-New-Net [23] adopted a small batch size of two. However, reducing batch size would make the network hard to converge. In particular, when the batch size is reduced to one, the Batch Normalization [24] that is important for stable gradient propagation becomes invalid. Another possible way to reduce computation is to compress the width and depth of the network and image resolution, which is under the risk of accuracy decrease. Researches have shown that the increase of any of these three elements can improve model performance (the width [25], the depth [26], the image size [27], and all three aspects [28]). In this paper, we propose an end-to-end network with symmetrical encode and decode paths to automatically segment multiple organs. The model complexity of the proposed network can be flexibly adjusted to meet the computation requirements without accuracy sacrifice. Neither the batch size nor the model capacity needs to be changed. Concretely, we introduce an accuracy-complexity adjustment module (ACAM) inspired by Chen et al. [20] throughout the encoder and decoder. A hyperparameter is adopted to control the compression of spatial redundancy of feature maps. To further enrich feature representations and capture information of organs in different sizes, we add a nonlinear multiscale aggregation module (MAM) after the encoder. The branches of the module have different rates of dilations, which is inspired by Li et al. [13].

Our contributions are summarized as follows: (i)We introduce an accuracy-complexity adjustment module throughout the encoder and decoder to increase the segmentation accuracy and reduce the model complexity and memory usage simultaneously(ii)We present an attention-based multiscale aggregation module after the encoder to enrich feature representation for further boosting the segmentation accuracy(iii)The proposed network achieves competitive results, which outperform several state-of-the-art methods on chest CT segmentation datasets

The rest of this paper is organized as follows. In Section 2, the proposed method is described in detail. In Section 3, experimental results of our method and the state-of-the-art methods are presented. Finally, we state the conclusion in Section 4.

2. Materials and Methods

Our network is a U-shaped architecture, which contains two symmetrical encode and decode paths [4]. Skip connections are used to transfer the features from the encoder directly to the decoder. ACAMs are applied all-through the encoder and decoder. MAM is added after the encoder to further enrich feature representation and boost segmentation performance. Details of the network architecture can be seen in Figure 1. The numbers of channels of all layers can be seen in Table 1.


Layer nameInput channelsOutput channels
High frequencyLow frequencyHigh frequencyLow frequency

InConv16464
DownConv16464128128
DownConv2128128256256
DownConv3256256512512
DownConv4512512512512

MAM512512512512

UpConv110241024256256
UpConv2512512128128
UpConv32562566464
UpConv41281286464
OutConv6464num of classes

2.1. Accuracy-Complexity Adjustment Module

Since images can be divided into high frequency with details and low frequency with global layout, Chen et al. [20] infer that feature maps can also be divided into high and low frequency with different spatial contexts. Following this idea, we present an ACAM and apply it in every convolutional layer to get high- and low-frequency feature maps. The width and the height of the low frequency are reduced to half size, and the number of channels is controlled by a parameter to meet the computation resource.

To be specific, an input feature map , where denotes the spatial resolution and the number of channels, can be factorized into along the channel dimension via . and represent high and low frequency, respectively, which is shown in Figure 2. These two frequency feature maps extract information independently through intrafrequency update and communicate with each other through interfrequency exchange. An output feature map is represented as , which can be defined as follows: where and denote intraupdates, while and denote interexchanges. The intra- and interfrequency communications help to strengthen information propagation between channels, which brings potential improvements.

To compute the output feature map , a standard convolution kernel with weight of size is split into four parts via shown in Figure 3. For the intrapart, a vanilla convolution is the only requirement, while for the interpart, resolution matching (i.e., downsampling or upsampling) should be performed before or after convolution. Here, we use average pooling for downsampling and nearest interpolation for upsampling, which is shown in the bottom left corner of Figure 1. Then, Equation (1) can be rewritten as where denotes a convolution function with weight parameter , denotes upsampling by a factor of , and denotes downsampling with kernel size and strike .

In ACAM, two operation strategies are used to improve the performance. First, the upsampling is performed after convolution in , and downsampling is performed before convolution in . Such an order makes convolutions operate on smaller feature maps, which can save computations. Second, using the same size of kernels on the low frequency is equivalent to enlarging the receptive field size in the original pixel space, which introduces additional multiscale information.

The setting of can be divided into three stages. In the first convolutional layer, we set and , in which case is disabled. This stage is usually used at the beginning of the network, where an original CT image is split into high and low frequency. The high frequency is obtained by a convolution function, and the low frequency is obtained through a sequential downsampling and convolution operation. In the last convolutional layer, we set and , in which case is disabled. This stage is usually used at the end of the network, mapping the feature map to the prediction mask. The low frequency is restored to high frequency through a sequential convolution and an upsampling operation, which is added to the high-frequency part to get the final result. In all the middle convolutional layers, we keep in ACAMs. The intermediate feature maps are obtained by Equation (3).

There are three kinds of operations in the ACAM, which are convolution, pooling, and interpolation. Compared with the convolution, the FLOPs of the pooling and interpolation are negligible. Therefore, we only compute the FLOPs of the convolution. For each convolution operation in ACAM, namely, , , , and , the theoretical FLOPs can be obtained by

The final FLOPs include four subconvolutions, which are merged as follows:

Therefore, by controlling the rate of low-frequency , ACAM can save computations compared with a vanilla convolution.

2.2. Multiscale Aggregation Module

Targets in medical images usually have different sizes (e.g., spinal cord and lungs in this segmentation task). The size of a specific organ also varies in different slices. Multiscale information should be taken into consideration during training. Therefore, the MAM is added after the encoder to enrich the feature representation.

Formally, let be an input feature map, where denotes the spatial resolution and the number of channels. The MAM with branches is proposed for to integrate multiscale information. Taking as an example, the module structure is shown in Figure 4. The basic sized convolution kernels corresponding to different branches have different dilated rates. The MAM generates feature maps in parallel, namely, . Remember that the purpose of MAM is to adaptively adjust the receptive field sizes of neurons in the next layer with multiscale information from these branches. Therefore, an element-wise fusion operation needs to be done first to integrate all the learned feature maps:

Then, the channel-wise statistics of , expressed as a tensor , can be built through a global average pooling to embed spatial context of each channel:

Then, is passed through a bottleneck made by a fully connected layer with weight and transformed into a compressed tensor , where is

The ratio relates to the degree of compression, which helps to limit model complexity and assist feature generalization. denotes a safe margin, which is used to prevent too much loss of channel-wise information caused by overcompression.

In order to separately calculate the attentions of branches, is split into independent tensors through fully connected layers with weight , decoding back to the original size. Then, an activation function across channels is used to build the corresponding attention maps , which is a softmax function in this case. These attention maps emphasize the important channels and ignore the less important ones in branches, acting as gates to constraint the information propagation. The final output is obtained through a summation over channel-wise products of these attention weights and feature maps:

There are four operations in one MAM, which are convolutions, global average poolings, fully connected layers, and softmax. To analyze the model complexity, we compute FLOPs of each operation individually. There are fully connected layers, which are all performed between one-dimensional tensors. Their total FLOPs are at most , depending on the ratio parameter , which is negligible compared with convolutions. We also ignore the FLOPs of softmax operation, which is performed times on one-dimensional tensors. Element-wise additions and multiplies are conducted and times, respectively. The sum of their FLOPs is , also much smaller than that of convolutions. Therefore, the main additional complexity is brought by the convolution performed early in each branch. To reduce the computation cost of the MAM, we put it in a bottleneck layer and compress the number of channels through point-wise convolutions.

3. Results and Discussion

3.1. Dataset and Metrics

An in-house dataset was used to evaluate the proposed method, which is provided by the Institute of Cancer and Basic Medicine (ICBM), Chinese Academy of Sciences, and Cancer Hospital of the University of Chinese Academy of Sciences. Manually labelled masks of the left lung, right lung, heart, and spinal cord of 36 patients were defined by two experienced radiation oncologists. The CT scans have pixels in resolution. The pixel spacing varies from 0.95 mm to 1.37 mm, and all thicknesses are 5 mm. We randomly shuffled the dataset and split it into three subsets: training set, validation set, and test set. The training set accounts for 70% of the entire data set. The validation set accounts for 10%, and the test set accounts for 20%. No preprocessing was performed before training. A public dataset called SegTHOR is also used to test our method, which can be found in CodaLab. There are four organs in this dataset, esophagus, heart, trachea, and aorta. The CT scans have pixels in resolution. The pixel spacing varies from 0.90 mm to 1.37 mm. The thicknesses vary from 2 mm to 3.7 mm. Image data and labels of 40 patients can be downloaded to train the network. Image data of 20 patients can be tested offline and uploaded onto the website to obtain the test results (https://competitions.codalab.org/competitions/21145). In the following sections, except for special declarations, we default to using the in-house dataset for experiments.

The proposed method was implemented by using a deep learning framework called Pytorch and trained on a single NVIDIA GTX 1080 Ti GPU with 11 GB of memory. The code of our complete method can be found online https://github.com/Jennsoo/UNet-Accuracy-Complexity.git. The network was trained with a stochastic gradient descend optimizer with an initial learning rate 0.03, momentum 0.9, and weight decay 0.0005. Poly learning rate policy was adopted to further decrease the learning rate by a factor of . We trained our model using crossentropy loss with 100 epochs in total. Weights were initialized according to MSRA [29], a zero-mean Gaussian distribution with variance , in which is the number of input elements. The Dice similarity coefficient (DSC) is our evaluation metric, which is defined as follows:where is the ground truth and is the prediction.

3.2. Guidance for Adjusting Accuracy and Complexity

The hyperparameter in ACAM is the main factor to balance the accuracy and model complexity. To evaluate how affects the DSC-FLOP trade-off in the segmentation task, we conduct an experiment with on our in-house dataset. The corresponding mean DSC, FLOPs, and memory are listed in Table 2. In particular, relates to the baseline, which means no low frequency is separated out. We observe that ACAMs notably improve the segmentation DSC, while having less computation complexity and memory cost. Such boost can be attributed to the multifrequency processing in ACAMs, which brings an extra multiscale complement. We plot the mean DSC, FLOPs, and memory in Figure 5. From the figure, we can see that the DSC first significantly increases and then slowly declines, while FLOPs and memory decrease monotonically. The network reaches its best DSC at , about 1% DSC improvement, while decreasing 18% complexity and 9% memory cost. For higher , the DSC slowly goes down from the peak but is still better than the baseline. An intuitive explanation is that the spatial compression of low-frequency feature maps does not lead to serious information loss. This confirms the argumentation that feature maps in convolutional layers do have spatial redundancy. Through compressing the feature maps, computation and memory cost can be effectively saved without sacrificing accuracy. This experiment provides guidance for selection. The DSC of is almost the same with the DSC of , but the FLOPs and memory of take more advantages. Therefore, we set for all the ACAMs.


DSC (%)FLOPs (G)Memory (106)

Baseline ()94.34123.931439
0.12595.43102.09 (82%)1313 (91%)
0.2595.4183.26 (67%)1188 (83%)
0.595.3254.28 (44%)938 (65%)
0.7595.0936.84 (30%)687 (47%)
0.87594.8832.45 (26%)562 (39%)
1.093.5230.98 (25%)365 (25%)

3.3. The Influence of the Number of Branches in Multiscale Aggregation Module

Here, we conduct an experiment on the in-house dataset to explore the best choice of branches of the MAM. Concretely, we use kernels with different dilated rates to generate feature maps with different receptive field sizes. The basic kernel size is with dilated rate . For each new branch, the dilated rate of the kernel is multiplied by a power of two. Note that a standard MAM requires at least two branches. In this paper, three designs of branches are presented, and its corresponding results are recorded in Table 3. It is worth mentioning that we have tried to add the MAM in every convolutional layer as a parallel module, but the results were not satisfactory. We plot the mean DSC, FLOPs, and memory in Figure 6. No preprocessing was performed before training.


#Branches (dilation)DSC (%)FLOPs (G)Memory (106)

Baseline94.34123.931439
2 (1, 2)94.54125.41 (+1.19%)1454 (+1.04%)
3 (1, 2, 4)94.71126.02 (+1.69%)1457 (+1.25%)
4 (1, 2, 4, 8)95.02126.62 (+2.17%)1460 (+1.46%)

We observe that DSC is positively correlated with FLOPs and memory. The MAM with more branches performs better. A possible explanation is that more branches bring more multiscale information, which helps to capture a richer global context. Another observation is that the MAM is lightweight with negligible extra model complexity. Even if we use four branches, it will not consume much extra computation and memory resources. In all the following experiments, we use dilation of (1, 2, 4, 8) for the MAM.

3.4. Ablation Experiment

We investigate the performance of aggregating ACAMs and MAM in the proposed method. We list all the DSCs of every single organ of the in-house dataset in Table 4 and plot it in Figure 7. Aside from the conclusion in Sections 3.2 and 3.3 that both ACAMs and MAM can boost the performance of the baseline, we can get two more observations when looking into the details. First, the results obtained by jointly applying ACAMs and MAM are better than using them alone, which indicates that there is a certain complementary relationship between the two modules. Figure 8 shows qualitative results of four images, validating this observation. Second, the DSCs of small organs get more improvement, especially the spinal cord. Our method improves the DSCs of the lungs and heart by about 0.8%, while increasing the DSC of the spinal cord up to 3%.


L-lungR-lungHeartSp-CoMean

Baseline96.8696.8894.4089.2094.34
Baseline+MAM97.5397.2794.9790.3195.02
Baseline+ACAMs97.6397.4395.2091.3695.41
Baseline+ACAMs+MAM97.7697.5295.2792.2495.70

4. Comparison Experiment

Table 5 and Figure 9 present the comparison results on the in-house dataset with several state-of-the-art methods, which are U-Net [4], PSPNet [7], DeepLab v3+ [11], DANet [12], Attention U-Net [5], and Nested U-Net [6]. The implementation of these methods can be found online (U-Net https://github.com/milesial/Pytorch-UNet, PSPNet https://github.com/hszhao/PSPNet, DeepLab v3+ https://github.com/jfzhang95/pytorch-deeplab-xception, DANet https://github.com/junfu1115/DANet, and Attention U-Net and Nested U-Net https://github.com/bigmb/Unet-Segmentation-Pytorch-Nest-of-Unets. Although DANet achieves state-of-the-art performance on the Cityscape dataset, it is not applicable to our dataset. From the table, we can see that the DSCs of the proposed method on four organs are the best among state-of-the-art methods. This can be attributed to the combination of the ACAMs and MAM, which strengthens the feature extraction capabilities. We compare our method on each individual organ with other methods, which performs the best on it. Our network additionally increases the DSC of the left lung by 0.51% over DeepLab v3+, the right lung by 0.42% over Attention U-Net, the heart by 0.76% over Nested U-Net, and the spinal cord by 1.15% over Attention U-Net.


L-lungR-lungHeartSp-CoMean

U-Net [4]96.8696.8894.4089.2094.34
PSPNet [7]96.8396.7094.1283.7092.84
DeepLab v3+ [11]97.2597.0194.1089.0794.36
DANet [12]93.2493.7485.6859.7683.11
Attention U-Net [5]97.2497.1094.3691.0994.95
Nested U-Net [6]96.6996.7694.5189.9194.47
Ours97.7697.5295.2792.2495.70

Figure 10 shows qualitative results of four methods, which are DeepLab v3+, Attention U-Net, Nested U-Net, and our network. In the first image, the left lung looks very different from the common mirrored “C” shape. DeepLab v3+, Attention U-Net, and Nested U-Net are confused that they fail to recognize it, while our method gets a more accurate segmentation. In the second image, there is a thinner structure at the top of the left lung. Attention U-Net and Nested U-Net both ignore this structure and disrupt the continuity of the left lung during segmentation. DeepLab v3+ and our method perform well in such case. This may be due to multiscale modules applied in both methods, which is ASPP in the DeepLab v3+ and MAM in our method. In the third image, there is a noticeable small but deep depression in the left lung. Attention U-Net totally omits it and fills in this depression. DeepLab v3+ and Nested U-Net seem to observe this structure and “dig a hole” on the left lung. Our method performs the best and delineates the shape well. This further indicates that our method not only improves the segmentation of small targets but also is effective for the subtle structures. In addition, we randomly select a few cases and display the visualization results in Figure 11. In all these cases, our method has achieved satisfactory results, but there are still some defects on complex boundaries.

We observed that DANet [12] does not fit our dataset, even though it has achieved advanced results on other datasets. One possible reason is that the targets to segment in our cases have different sizes. The spinal cord is extremely small compared to other organs and background. Both the PSPNet and DeepLab v3+ in our comparison experiment introduce pyramid pooling modules. The U-Net-based methods also introduce multiscale components due to the skip connections between different levels. However, DANet only considers the long-range dependencies between pixels but ignores the target scale information, which may become a potential factor for small object segmentation errors.

In the SegTHOR dataset, our method ranks in 4 out of 19 methods (until 6/15/2020). The esophagus, trachea, and aorta are small organs, while the heart is a relatively big organ. Except for the esophagus, our method generally achieves competitive results, which shows the effectiveness and generalization of our method. All the methods on the ranking perform the worst on the segmentation of the esophagus. We analyze the dataset and find that the esophagus is more special compared with the others, making it more challenging to segment. The esophagus is small in size and variable in shape. The boundary of the esophagus is blurred due to its low contrast displayed in CT scans. Our method is not designed for such problems and is not robust enough to deal with such situations.

Model Complexity Analysis. Since the model complexity plays an important role in the design of our network, we compare the FLOPs and memory of our method with that of other state-of-the-art methods in Table 6. It is obvious that our method consumes the least computation and memory resources. The memory usage of our method is reduced nearly by 50%. The FLOPs of our method are reduced by an average of 67% except Nested U-Net. Nested U-Net has much more FLOPs due to its denser connections between convolutional layers. The reason is that the ACAMs embedded in our method compress the spatial redundancy of the feature maps, which can significantly reduce the computation cost.


FLOPs (G)Memory (106)

U-Net [4]123.931439
PSPNet [7]262.132525
DANet [12]275.422579
Attention U-Net [5]266.252402
Nested U-Net [6]552.012715
Ours89.771221

5. Conclusions

In this paper, we propose an end-to-end method to segment multiple organs in CT scans. Benefiting from the compression of spatial redundancy in the applied accuracy-complexity adjustment module, the model complexity can be reduced, while achieving higher accuracy. We also present the experimental results to provide guidance for the hyperparameter selection. A nonlinear multiscale aggregation module is added after the encoder to further enrich feature representation. The combination of these two modules in the proposed method achieves higher DSC and lower complexity than several state-of-the-art methods. Such an idea provides a direction of improving performance of high-complexity models like 3D U-Net, which will be our future work.

Data Availability

The in-house dataset used to support the findings of this study was supplied by the Chinese Academy of Sciences under license and so cannot be made freely available. Requests for access to these data should be made to zhiqiangtian@xjtu.edu.cn.

Conflicts of Interest

The authors declare that there is no conflict of interest regarding the publication of this paper.

Acknowledgments

This work was supported in part by NSFC under grant No. 61876148. This work was also supported in part by the Fundamental Research Funds for the Central Universities No. XJJ2018254 and China Postdoctoral Science Foundation No. 2018M631164.

References

  1. R. Trullo, C. Petitjean, B. Dubray, and S. Ruan, “Multiorgan segmentation using distance-aware adversarial networks,” Journal of Medical Imaging, vol. 6, no. 1, article 014001, 2019. View at: Publisher Site | Google Scholar
  2. J. Long, E. Shelhamer, and T. Darrell, “Fully convolutional networks for semantic segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 3431–3440, Boston, MA, USA, 2015. View at: Publisher Site | Google Scholar
  3. V. Badrinarayanan, A. Kendall, and R. Cipolla, “Segnet: a deep convolutional encoder-decoder architecture for image segmentation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 39, no. 12, pp. 2481–2495, 2017. View at: Publisher Site | Google Scholar
  4. O. Ronneberger, P. Fischer, and T. Brox, “U-net: convolutional networks for biomedical image segmentation,” in International Conference on Medical image computing and computer-assisted intervention, pp. 234–241, Munich, Germany, 2015. View at: Publisher Site | Google Scholar
  5. O. Oktay, J. Schlemper, L. Le Folgoc et al., “Attention u-net: learning where to look for the pancreas,” 2018, https://arxiv.org/abs/1804.03999. View at: Google Scholar
  6. Z. Zhou, M. M. R. Siddiquee, N. Tajbakhsh, and J. Liang, “Unet++: a nested u-net architecture for medical image segmentation,” in Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, pp. 3–11, Springer, 2018. View at: Google Scholar
  7. H. Zhao, J. Shi, X. Qi, X. Wang, and J. Jia, “Pyramid scene parsing network,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 2881–2890, Honolulu, HI, USA, 2017. View at: Publisher Site | Google Scholar
  8. L. C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, and A. L. Yuille, “Semantic image segmentation with deep convolutional nets and fully connected CRFs,” 2014, https://arxiv.org/abs/1412.7062. View at: Google Scholar
  9. L.-C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, and A. L. Yuille, “Deeplab: semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 40, no. 4, pp. 834–848, 2017. View at: Publisher Site | Google Scholar
  10. L.-C. Chen, G. Papandreou, F. Schroff, and H. Adam, “Rethinking atrous convolution for semantic image segmentation,” 2017, https://arxiv.org/abs/1706.05587. View at: Google Scholar
  11. L.-C. Chen, Y. Zhu, G. Papandreou, F. Schroff, and H. Adam, “Encoder-decoder with atrous separable convolution for semantic image segmentation,” in Proceedings of the European conference on computer vision (ECCV), pp. 801–818, Munich, Germany, 2018. View at: Publisher Site | Google Scholar
  12. J. Fu, J. Liu, H. Tian et al., “Dual attention network for scene segmentation,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3146–3154, Long Beach, CA, USA, 2019. View at: Publisher Site | Google Scholar
  13. X. Li, W. Wang, X. Hu, and J. Yang, “Selective kernel networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 510–519, Long Beach, CA, USA, 2019. View at: Google Scholar
  14. J. Hu, L. Shen, and G. Sun, “Squeeze-and-excitation networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 7132–7141, Salt Lake City, UT, USA, 2018. View at: Publisher Site | Google Scholar
  15. P. Molchanov, S. Tyree, T. Karras, T. Aila, and J. Kautz, “Pruning convolutional neural networks for resource efficient inference,” 2016, https://arxiv.org/abs/1611.06440. View at: Google Scholar
  16. A. G. Howard, M. Zhu, B. Chen et al., “Mobilenets: efficient convolutional neural networks for mobile vision applications,” 2017, https://arxiv.org/abs/1704.04861. View at: Google Scholar
  17. M. Sandler, A. Howard, M. Zhu, A. Zhmoginov, and L.-C. Chen, “Mobilenetv2: inverted residuals and linear bottlenecks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4510–4520, Salt Lake City, UT, USA, 2018. View at: Publisher Site | Google Scholar
  18. X. Zhang, X. Zhou, M. Lin, and J. Sun, “Shufflenet: an extremely efficient convolutional neural network for mobile devices,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6848–6856, Salt Lake City, UT, USA, 2018. View at: Publisher Site | Google Scholar
  19. N. Ma, X. Zhang, H.-T. Zheng, and J. Sun, “Shufflenet v2: practical guidelines for efficient cnn architecture design,” in Proceedings of the European Conference on Computer Vision (ECCV), pp. 116–131, Munich, Germany, 2018. View at: Publisher Site | Google Scholar
  20. Y. Chen, H. Fan, B. Xu et al., “Drop an octave: reducing spatial redundancy in convolutional neural networks with octave convolution,” 2019, https://arxiv.org/abs/1904.05049. View at: Google Scholar
  21. R. Brügger, C. F. Baumgartner, and E. Konukoglu, “A partially reversible u-net for memory-efficient volumetric image segmentation,” 2019, https://arxiv.org/abs/1906.06148. View at: Google Scholar
  22. F. Milletari, N. Navab, and S.-A. Ahmadi, “V-net: fully convolutional neural networks for volumetric medical image segmentation,” in 2016 Fourth International Conference on 3D Vision (3DV), pp. 565–571, Stanford, CA, USA, 2016. View at: Publisher Site | Google Scholar
  23. F. Isensee, P. Kickingereder, W. Wick, M. Bendszus, and K. H. Maier-Hein, “No new-net,” in International MICCAI Brainlesion Workshop, pp. 234–244, Granada, Spain, 2018. View at: Publisher Site | Google Scholar
  24. S. Ioffe and C. Szegedy, “Batch normalization: accelerating deep network training by reducing internal covariate shift,” 2015, https://arxiv.org/abs/1502.03167. View at: Google Scholar
  25. S. Zagoruyko and N. Komodakis, “Wide residual networks,” 2016, https://arxiv.org/abs/1605.07146. View at: Google Scholar
  26. K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778, Las Vegas, NV, USA, 2016. View at: Publisher Site | Google Scholar
  27. Y. Huang, Y. Cheng, A. Bapna et al., “Gpipe: efficient training of giant neural networks using pipeline parallelism,” in Advances in Neural Information Processing Systems, pp. 103–112, Curran Associates, Inc., 2019. View at: Google Scholar
  28. M. Tan and Q. V. Le, “Efficientnet: rethinking model scaling for convolutional neural networks,” 2019, https://arxiv.org/abs/1905.11946. View at: Google Scholar
  29. K. He, X. Zhang, S. Ren, and J. Sun, “Delving deep into rectifiers: surpassing human-level performance on imagenet classification,” in Proceedings of the IEEE international conference on computer vision, pp. 1026–1034, Santiago, Chile, 2015. View at: Publisher Site | Google Scholar
  30. W. Owadally and J. Staffurth, “Principles of cancer treatment by radiotherapy,” Surgery (Oxford), vol. 33, no. 3, pp. 127–130, 2015. View at: Publisher Site | Google Scholar
  31. A. L. Grosu, L. D. Sprague, and M. Molls, “Definition of target volume and organs at risk. Biological Target Volume,” in New Technologies in Radiation Oncology, Springer, Berlin Heidelberg, 2006. View at: Google Scholar
  32. S. Scoccianti, B. Detti, D. Gadda et al., “Organs at risk in the brain and their dose-constraints in adults and in children: a radiation oncologist's guide for delineation in everyday practice,” Radiotherapy and Oncology, vol. 114, no. 2, pp. 230–238, 2015. View at: Publisher Site | Google Scholar
  33. L. D. van Harten, J. M. H. Noothout, J. J. C. Verhoeff, J. M. Wolterink, and I. Iˇsgum, “Automatic segmentation of organs at risk in thoracic CT scans by combining 2D and 3D convolutional neural networks,” in Proceedings of the 2019 Challenge on Segmentation of THoracic Organs at Risk in CT Images (SegTHOR2019) co-located with the 16th International Symposium on Biomedical Imaging (ISBI), Venice, Italy, 2019. View at: Google Scholar

Copyright © 2020 Zhiqiang Tian et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


More related articles

 PDF Download Citation Citation
 Download other formatsMore
 Order printed copiesOrder
Views451
Downloads286
Citations

Related articles

Article of the Year Award: Outstanding research contributions of 2020, as selected by our Chief Editors. Read the winning articles.