Journal of Healthcare Engineering

Journal of Healthcare Engineering / 2019 / Article

Research Article | Open Access

Volume 2019 |Article ID 5191630 | 10 pages |

Towards Fine Whole-Slide Skeletal Muscle Image Segmentation through Deep Hierarchically Connected Networks

Academic Editor: Norio Iriguchi
Received19 Nov 2018
Accepted14 Mar 2019
Published27 Jun 2019


Automatic skeletal muscle image segmentation (MIS) is crucial in the diagnosis of muscle-related diseases. However, accurate methods often suffer from expensive computations, which are not scalable to large-scale, whole-slide muscle images. In this paper, we present a fast and accurate method to enable the more clinically meaningful whole-slide MIS. Leveraging on recently popular convolutional neural network (CNN), we train our network in an end-to-end manner so as to directly perform pixelwise classification. Our deep network is comprised of the encoder and decoder modules. The encoder module captures rich and hierarchical representations through a series of convolutional and max-pooling layers. Then, the multiple decoders utilize multilevel representations to perform multiscale predictions. The multiscale predictions are then combined together to generate a more robust dense segmentation as the network output. The decoder modules have independent loss function, which are jointly trained with a weighted loss function to address fine-grained pixelwise prediction. We also propose a two-stage transfer learning strategy to effectively train such deep network. Sufficient experiments on a challenging muscle image dataset demonstrate the significantly improved efficiency and accuracy of our method compared with recent state of the arts.

1. Introduction

Skeletal muscle accounts for approximately body mass. As the largest body tissue, skeletal muscle has been extensively recognized as the biomedical health biomarker related to many diseases such as cancer cachexia, heart failure, and chronic obstructive pulmonary disease (COPD) [13]. In recent years, the growing attentions, in the muscle biology community, have been paid to the analysis of histological images of skeletal muscle to assist the diagnosis of relevant diseases [1].

The quantification of morphological characteristics of muscle fibres plays an important role in the assistance of disease diagnosis and clinical studies. Critical morphological characteristics, including cross-section area, fiber type and shape, and the minimum Feret diameter, are closely related to the functionality and health of muscle [4, 5]. To accurately quantify these morphological characteristics of muscle fibres, an accurate skeletal muscle image segmentation (MIS) system is the prerequisite.

Currently, the segmentation for muscle fibres in routine practice still highly relies on experts’ manual labors or semiautomatic process [6], which are not only expensive but also contain large interobserver variations. The increasing demand of a fast and accurate automatic MIS attracts many attentions recently. Various approaches have been proposed to address this task [4, 79].

The difference between MIS and standard histological image cell segmentation is attributed to the specific morphology of skeletal muscle. Skeletal muscle is composed of long, multinucleated cells (fibres) tightly grouped into fascicles, interspersed with other mononucleated cell types and surrounded by connective tissues and fat. This tightly grouped anatomical structure, coupled with artifacts and staining variances introduced during sample preparation, generates confusing and overlapping cell boundaries.

Although histological cell segmentation research has a rich history, few have been successfully applied to MIS. There are still several challenges which remain to be solved to achieve a robust and automatic MIS system. First, all exiting MIS methods can only handle small image patches with the size smaller than cropped from whole-slide muscle images. The main reason is that supervised methods usually rely on handcrafted features and pretrain classifiers to distinguish cell or noncell regions or pixels. However, the computation of well-designed handcrafted features and regionwise classification is usually expensive and the time cost is proportional to the image scale. Therefore, this limitation of existing methods makes them hardly be applied for large-scale, whole-slide muscle images.

Second, the special muscle cell shape and size as abovementioned make the segmentation methods hardly be generalized. For example, unsupervised methods, such as the deformable model [10, 11] and shape prior-based methods [12, 13], have been widely used in histological and microscopy cell segmentation. However, the arbitrarily transformed cell shapes and size increase the difficulty to use shape information for MIS.

Third, the densely touching fibres and staining artifacts make the fiber boundaries unclear and broken, which increases the difficulties for methods to separate multiple touching fibres by using boundary information. Fourth, current methods usually contain multiple mutually dependent steps; the failure of either step will largely affect other steps and the final results. On the other hand, the complex pipeline largely decreases the speed of MIS.

This paper addresses these challenges to achieve a both efficient and effective MIS method based on recently popular convolutional neural network (CNN). CNN-based methods have achieved unprecedented performance in various medical image applications. Different from conventional computer vision methods, CNN has strong capability to learn comprehensive representations via a deep architecture for effective classification. When using CNN for pixelwise classification, conventional CNN shows the efficiency shortcomings [14]. The end-to-end CNN training strategy has recently attracted a lot of research interest [15, 16]. However, a common problem is that the dense output is relatively coarse, and it is difficult to accurately classify each pixel [16, 17]. To generate more accurate and fine outputs, a refinement procedure needs to be considered.

In this paper, we propose a novel MIS method based on CNN trained in an end-to-end manner [16], which enables the CNN to better utilize the rich representations and directly predict fine-grained segmentation given an arbitrarily sized input image. Figure 1 shows some segmentation results of different image scales. Specifically, the main contributions of this paper are summarized as follows:(i)We propose a network whose architecture mainly contains two modules: the encoder and the decoder. The encoder captures rich representations through a very deep CNN architecture. The decoder leverages on the hierarchy characteristic of the encoder to enable multiscale prediction independently. A refinement procedure of the decoder automatically addresses the fine-grained dense outputs. Figure 2 illustrates our network.(ii)We propose a novel spatially weighted loss function to take care of the unbalanced class issue and unavoidable errors happened in ground truth, which encourage the convergence of the network.(iii)We propose a two-stage training approach to train the proposed very deep network, which facilitates the network to better use pretrained CNN for better convergence and preserve the weak boundary information of muscle cells.(iv)We conduct sufficient experiments on an expertise-annotated skeletal muscle image dataset demonstrating the significantly improved efficiency and accuracy compared with other state of the arts.

The growing interest in the computer-aided histological image diagnosis entails rich research literature. As one of the histological image analysis family, skeletal muscle image analysis is a new yet recently popular application which has built successful cooperations with clinics to accelerate their research and clinical trials [13, 18, 19].

As the prerequisite of skeletal muscle image analysis, various methods have been proposed for MIS. Klemencic et al. [10] proposed a semiautomatic muscle image segmentation approach based on the active contour model. Janssens et al. [8] proposed a top-down cell segmentation framework using supervised learning and clump splitting, which requires a long pipeline with the help of several low-level image processing techniques. However, the performance of these image processing techniques can be easily influenced by imaging artifacts and cell clumps. Smith and Barton [6] proposed SMASH—semiautomatic muscle image analysis software. Some other software applications such as CellProfiler [9] have obtained high exposure in histological image analysis community. However, these software applications show nonsatisfactory results for challenging muscle images. Practically, time-consuming manual adjustment is still needed. Liu et al. [4] proposed a deformable model-based segmentation algorithm. The success relays largely on the initial centers of the muscle cells. It is not able to handle cells with arbitrarily transformed fiber shape, and it requires complex postprocessing to refine the results, which is not robust in practice. Recently, Liu et al. [7] proposed a hierarchical tree-based region selection method to segment muscle fibres, which relies on elaborately designed features and high-level machine learning techniques. This method first detects fiber boundaries by using structured random forest [20]; then, it builds a hierarchical region tree based on the detected edge map. Finally, the dynamic programming is performed to select candidate regions from the tree structure [21]. This method shows obvious improvement compared with previous MIS approaches. However, this method still suffers from relatively expensive computation, so it is unable to be applied onto whole-slide images.

As a matter of fact, whole-slide MIS is still an unsolved problem. Although some literature discusses the usage of distributed computing [2224] to accelerate process for large-scale histological images, distributed computing is usually difficult to be deployed for practical usage in clinical practice.

Convolutional neural network (CNN) [25] is one major branch of the deep learning family. Its applications in pathology and histological image analysis domain became increasingly popular very recently [14, 2628]. CNN has shown strong ability to handle complex classification problem [29]. Recently, end-to-end CNN training concept is introduced for semantic image segmentation, termed fully convolutional neural network (FCN) [16]. Instead of performing patch-to-pixel prediction, it enables the network to perform spatial dense classification (i.e., a segmentation mask) given a test image. By taking advantaging of this strength, several methods have been proposed to handle various pixelwise classification tasks [15, 3034]. Our paper shares some similarity with the previous works of how to enable CNN to be trained in an end-to-end manner. Different from previous works, we have made several specific designs to handle fine-grained prediction, unbalanced class, multiscale features, and transfer learning from pretrained model for MIS. More details are discussed in the rest of the paper.

3. Methodology

In this section, we begin by introducing the proposed network architecture and then present proposed loss function for training the network. Finally, we introduce the two-stage learning to train the overall network.

3.1. Network Architecture

We briefly introduce the convolutional neural network (CNN) at first. CNN [25] is a variant of multilayer perception (MLP), which is mainly composed of multiple stacked computation layers from bottom to top, including convolutional, max-pooling and fully connected layers, activation layer, etc. The convolutional layer uses learnable convolutional filters to extract representations from locally connected image regions (receptive fields). The max-pooling layer reduces the dimensionality of the obtained representations from convolutional layers while keeping the feature translation invariance. The fully connected layer uses all features for high-level classification. From bottom layers to top layers, CNN gradually captures rich representations of input image from pixel level to content level so as to make accurate classification. Conventional CNN is used to perform high-level classification, i.e., assigning a category label to an input image patch. When it is applied to pixelwise prediction MIS task, extensive patch-to-pixel level prediction (CNN feedforward) is required, which will extensively limit the segmentation efficiency [14, 35].

To solve this problem, we train our network in an end-to-end manner, which enables the network to directly output the image segmentation given an input image. In this way, we no longer need to use patchwise classification to assign labels to all pixels via millions of CNN feedforward. Only one-time feedforward is needed to obtain the final segmentation. End-to-end training of CNN is used to enable the network to directly output dense image segmentation for a given input image [15, 16, 32].

However, modifying conventional CNN to perform end-to-end training brings a major side effect, i.e., substantial pixel-level information loss at top layers makes the pixelwise prediction inaccurate [17, 30]. It is because multiple max-pooling layers will dramatically decrease the spatial size of the output, so the predicted segmentation output is very coarse. Most proposed end-to-end CNN methods use upsampling [16, 17] or deconvolution operations [15] to resize back the output to the spatial size of the input image. Nevertheless, the max-pooling layer is essential to abstract the content-level representations for high-level category classification [25, 29, 36] and decrease the computation space of CNN.

As a matter of fact, when we generalize end-to-end CNN to MIS, content-level information becomes less important because the label of a single pixel does not rely on the knowledge of the whole muscle image. Different from semantic segmentation [16, 32] which needs content-level information to predict the category label per pixel, we are more interested in the fine-grained pixelwise prediction by taking advantage of the hierarchical representations of the encoder to improve the prediction accuracy. The hierarchy characteristic can be achieved by gradually enlarging the receptive field size after each max-pooling layer. To this end, we propose a novel network architecture, which is composed by one encoder module and multiple decoder modules. Generally, the decoder aims to use the rich and hierarchical representations obtained from the encoder for pixelwise classification.

3.1.1. Encoder Module

The encoder architecture is mostly identical to the conventional neural network. Instead of building our own layer combinations, we borrow the well-known VGG net [29] with fully connected layers truncated to capture the rich and hierarchical representations from pixel level at bottom layers to content level (i.e., category-specific knowledge) at top layers. VGG net is composed of a series of convolutional sets with each set having multiple convolutional layers followed by a max-pooling layer. VGG has two variants (one has 16 layers and the other has 19 layers); we use 16-layer VGG for efficiency consideration. We choose VGG for two reasons: (1) we can transfer the pretrained VGG model to help train our very deep network as described in the next section; (2) VGG net is very deep which extracts five different-scale feature maps, containing very rich multiscale representations for the usage of decoders.

3.1.2. Decoder Module

The decoder has two main purposes: (1) it utilizes the rich representations obtained from the encoder for pixelwise classification. So, the output of one decoder is a dense segmentation mask with each spatial position assigning a label to the corresponding pixel of the input image (cell or noncell in our case); (2) it refines the low-rescale coarse segmentation mask to efficiently generate fine-grained high-scale segmentation mask. The refinement procedure is achieved by multistep deconvolution and successive usage of same-scale feature maps obtained from other decoders.

We propose to connect multiple decoders prior to every max-pooling layer of the encoder; thus, the decoders can easily utilize the multiscale representations as input features as inspired by [15, 16, 31]. The decoder can be viewed as a small pixelwise classification network, which has an independent loss to update its parameters during training. Hence, the overall architecture is multitask CNN.

Our design of the decoder includes convolutional layers with intermediate deconvolution layers [15]. Specifically, the deconvolution is the backward convolution operation, which performs elementwise product with its filters (please note that some controversies arise in the naming of “deconvolution” in recent literature as the deconvolution layer used here is different from the previous definition of the deconvolution [37]; we maintain the same definitions as most of the literatures on end-to-end CNN). The output size of deconvolution will be enlarged by a factor of the stride. The filters of the deconvolution layers are learnable and updated by the loss of the decoder.

In this way, rather than enlarging the image with a large stride through a skip connection [16, 31, 38], our approach enlarges the feature map in multiple steps and progressively refines the feature maps at different scales via convolutional kernels, with the purpose of reducing the effects of pixel-level information loss. We use filter size as this small size has been proven effective widely. In the end, we concatenate multiscale predictions of all decoders, which generates a 5-dimensional feature map; we apply a convolutional layer to merge the feature map to generate the final output. Compared with how recent architecture [35, 39] uses multiscale information (resize input patch size and feed into multiple CNNs and merge all predictions [35, 39]), our approach enables multiscaling inside the network, requiring a single arbitrarily sized input and outputting the final segmentation result. Figure 3 specifies the parameters of each layer.

3.2. Spatially Weighted Loss for Backpropagation

This section describes the loss function training network through backpropagation. Our proposed spatially weighted loss plays an important role in network training.

Denote the training data as , where and N is the total number of pixels in the training image X. Y is the corresponding ground truth segmentation mask with each pixel (i.e., pixels inside and on the boundary the muscle cell and background otherwise). For an input image X, the main objective of our network is to obtain the pixelwise prediction :where is the prediction probability of pixel , i.e., the sigmoid function output of the network (denoted as afterwards for brevity). θ represents all parameters of our network.

Our network has multiple decoders with each having independent loss to update their parameters (see Section 3.1.2 for details). Denote the loss function of i-th decoder as . The extra convolutional layer after the concat layer is updated by another loss (see Figure 3), denoted as . Learning θ is achieved by minimizing the loss function , which is defined aswhere is the number of the decoder. Note that since and are both spatially computed on pixels of the dense output, both have the same formulation. The overall loss can be jointly minimized via backpropagation (specifically, when a layer has more than one path, such as the conv1-2 layer in Figure 3 which has two successive layers (decoder-1 and pool1), the gradients will be accumulated from multiple successive paths during backpropagation [40]).

In skeletal muscle images, there are several common problems which affect the network training: (1) the large proportion of pixels inside cell pixels will cause an unbalanced class such that the error effects occurred at the margins will be diminished during backpropagation; (2) usually cells are densely distributed and the boundaries between touching cells are thin and often unclear or broken due to muscle’s unique anatomy; based on our observations, the network often misclassifies the pixels at margins between fiber boundaries; (3) due to the staining issue, the boundary pixels are not smooth and continuous, so it is very difficult to ensure that annotations accurately label each pixel. It is necessary to reduce the ambiguity for network training.

We propose a loss function to ameliorate these problems by assigning different weights to the loss of each pixel. The loss function of a training data X, which is based on the cross-entropy loss, is defined as

The pixelwise weights are defined by the weight-assigning function f, which is defined as

The pixelwise weight-assigning function f has three terms, which play different roles to address the abovementioned three problems. The specific considerations make our proposed loss different from [16, 30].

In the first term, is the label frequency, which is a global term having the same value for same-class pixels. In second term, is the Euclidean distance of pixel to the boundary of the close cell. Similar to [30], the intention of f is to assign relatively high weights to pixels adjacent to boundaries to amplify the error penalty occurred at the margins and pixels close to fiber boundaries and 1 otherwise ( if ). We set and empirically. Compared with the “hard” error-balancing strategy in [16, 31], f produces soft error penalty so as to encourage better optimization convergence and enhance fine-grained prediction. The third term aims to reduce the reliability of the ground truth when the network predicts an opposite label with high probability. This term is a switch, so it forces the weight of the corresponding pixel to zero when the condition is not satisfied. In practice, we preserve this value during network feedforward, while the loss of the corresponding pixels does not get involved during network backpropagation.

3.3. Two-Stage Training

Training our deep network has some common difficulties:(i)The large number of parameters in both convolutional layers and deconvolutional layers makes the training difficult to achieve proper convergence [15, 41].(ii)Successful training from scratch requires extensive labeled data which are extremely difficult to obtain in medical image domain.

One typical solution is to apply transfer learning to reduce the training difficulty [41, 42], which reduces the difficulties of the tricky parameter initialization and tuning [25, 29] and heavy data acquirement procedure. The core idea behind is to use a pretrained model as the initialization and fine-tune the CNN to make it adapt to targeting tasks with new training data. The encoder of our network partially inherits the architecture of VGG [29], which is, however, trained on a large set of natural images for image classification. Transferring its knowledge to benefit the totally unrelated biological image analysis problem (i.e., MIS) seems impracticable. However, a recent literature coincides with our experiments. It demonstrates the advantage [41] using various biological imaging modalities transferring from AlexNet [25], a relatively shallow CNN for natural image classification. In terms of our MIS case segmentation, the network architecture is much more deeper with many new parameterized layers in decoders. More specific treatment needs to be considered.

It is well known that the bottom layers of CNN can be understood as various feature extractors attempting to capture the low-level image features such as edges and corners [25, 37, 41]. Actually, those low-level features are common between natural images and muscle images, of which the most common feature is image gradients (i.e., boundaries). In practice, we find that training the network to detect boundaries is relatively easier than directly training the network to segment muscle fibres.

We propose a two-stage training strategy to progressively train our network so as to utilize the powerful feature extractors of VGG and overcome the abovementioned problems. In the first stage, we apply transfer learning to use pretrained VGG to initialize the parameters of the encoder and randomly initialize the parameters of decoders. We then train the network to detect fiber boundaries, which is achieved by feeding the network with training muscle images associated with the ground truth boundary map (see Figure 2). This strategy will facilitate the network to converge swiftly. After the network becomes adapted to new muscle images, in the second stage, we fine-tune the model using the original training data (i.e., Y is the segmentation mask) to train the network to automatically segment muscle fibres, assigning in-cell pixels to 1 and other pixels to 0. More implementation details are described in the experimental section.

Another advantage of our proposed training strategy is that it further helps reduce the touch objects (due to thin boundaries) problem [30, 34] commonly occurred in end-to-end CNN segmentation (besides the pixel weight-assigning function f). The strategy of this literature [34] is to predict both a segmentation map and boundary map and merge two maps to solve touching glands. While in our method, the first stage training makes the network detect the cell boundaries. The second stage training is able to preserve this boundary information.

4. Experimental Results

4.1. Dataset

Our expert annotated skeletal muscle image dataset with H&E staining contains 500 annotated images, which are captured by the whole-slide digital scanner from the cooperative institution Muscle Miner. The images exhibit large appearance variances in color, fiber size, and shape. The image size roughly ranges from to pixels. We split the dataset into 100 testing images and 400 training images.

In order to evaluate the proposed method to handle large-scale images, we evaluate the runtime on a whole-slide image. Note that we use small image patches for segmentation accuracy evaluation because some comparative methods in the literature cannot handle whole-slide images. However, our proposed network is flexible to the input size during the testing stage because the decoder is able to adaptively adjust the output size to be consistent with the input size.

4.2. Implementation Details

Our implementation is based on the Caffe [40] framework with modifications for our network design. All experiments are conducted on a standard desktop with an Intel i7 processor and a single Tesla K40c GPU. The optimization is driven by stochastic gradient descent with momentum. For the first stage training, the network parameters are set to learning rate (divided by 10 every iteration), momentum , and minibatch size . In the second stage, we use the learning rate and keep the others the same.

Augmenting dataset is a normal step for training CNN. We apply a simple approach by randomly cropping 30 image patches from each of the training images to generate totally training data. We choose this patch size to take the memory capacity of GPU into account. Based on our observations, the segmentation accuracy will not be affected by increasing input size of test images. To simplify the computation of the weighting function f during training, we take another per-computed weighting map associated with each training data as network inputs.

4.3. Segmentation Accuracy Evaluation

For quantitative evaluation, we report Precision , Recall , and -score , where is the segmented cell region area and is the corresponding ground truth region area. For each test image, Precision and Recall are computed by averaging the results of all fibres inside. We report the three values with a fixed threshold (FT), i.e., a common threshold produces the best -score over the test set, and dynamic thresholds (DT) produce the best -score per image.

In Table 1, we compare the segmentation performance of our approach to several state-of-the-art methods. DC [43] and multiscale combinatorial grouping (MCG) [44] are recently proposed learning-based image segmentation methods. U-Net [30] is an end-to-end CNN for biomedical image segmentation. We use their public codes and carefully train the models over our training data with the same amount. DNN-SNM [14] is a well-known CNN-based image segmentation method. We regard it as a generic CNN for comparison with our end-to-end CNN approach. For our method, we directly use the network output as the segmentation results for evaluation without any extra postprocessing efforts.

Method-score ()Precision ()Recall ()

DC [43]48 ± 0.09360 ± 0.13841 ± 0.06654 ± 0.16467 ± 0.19473 ± 0.148
MCG [44]63 ± 0.20171 ± 0.10553 ± 0.13664 ± 0.13880 ± 0.30382 ± 0.091
DNN-SNM [14]76 ± 0.03378 ± 0.08083 ± 0.04285 ± 0.08970 ± 0.05873 ± 0.087
U-Net [30]80 ± 0.14381 ± 0.05487 ± 0.15586 ± 0.07674 ± 0.12677 ± 0.055
Liu et al. [7]82 ± 0.17284 ± 0.06181 ± 0.04384 ± 0.07185 ± 0.20285 ± 0.068
Our approach86 ± 0.18489 ± 0.04891 ± 0.17493 ± 0.05082 ± 0.17686 ± 0.058

σ is the standard deviation.

As shown in Table 1, our method achieves much better results than comparative methods. Although [7] has better Recall (FT), our method has around improvement on Precision (FT). DC and MCG are not robust to the image artifacts, which decreases their segmentation performance. Our method largely outperforms DNN-SNM and U-Net because (1) our network is deeper than DNN-SNM to capture richer representations, (2) the decoder better utilizes the multiscale representations than U-Net and is able to reduce the effects of the pixelwise information loss, and (3) two-stage training takes advantage of VGG for better training effectiveness rather than training from scratch as U-Net does. The outstanding Precision result demonstrates that our method produces more fine-grained segmentation than others. This superiority is better demonstrated by the qualitative evaluation as shown in Figure 4.

4.4. Whole-Slide Segmentation Runtime

In Table 2, we compare the runtime of our method to the comparative methods on images of different sizes cropped from a whole-slide image (see Figure 1). The runtime of non-deep learning-based methods (1st block) depends on both pixel and fiber quantities, so they cannot handle large-scale images. In contrast, deep learning-based methods (2nd and 3rd blocks) depend on the pixel quantity, so they have close-to-linear time complexity with respect to the image scale. We also implement a fast scanning version [45] of DNN-SNM on GPU. Although the speed has a large improvement, it is still much slower than ours. U-Net has more complicated layer connection configuration, so it is slower than ours, especially in large-scale cases. The significant speed improvement demonstrates the scalability of our proposed method to the application of whole-slide MIS with even larger scales.


DC [43]2079
MCG [44]727
Liu et al. [7]1059
DNN-SNM [14]26410562376422466009504129361689621384
U-net [30]
Our approach1.

The first three methods cannot handle images with 3x and larger sizes on our machine (represented with “—” in the table). DNN-SNM is a fast scanning implementation for prediction speed acceleration.

5. Conclusion

This paper presents a fast and accurate whole-slide MIS method based on CNN trained in the end-to-end manner. Our proposed network captures hierarchical and comprehensive representations to support multiscale pixelwise predictions inside the network. A two-stage transfer learning strategy is proposed to train such a deep network. Superior accuracy and efficiency are experimentally demonstrated on a challenging skeletal muscle image dataset. In general, our approach enables multiscaling inside the network, while just requiring a single arbitrarily sized input and outputting fine outputs. However, during the downsampling process of the encoding, due to the limitation of resolution of feature layer after downsampling, many important features, such as edge features of cells, are still lost. To further improve decoding efficiency, in the future work, we can design a module that complements important features to better improve network performance.

Data Availability

The data that support the findings of this study are available from the cooperative institution Muscle Miner, but restrictions apply to the availability of these data, which were used under license for the current study and so are not publicly available. Data are however available from the authors upon reasonable request and with permission of the cooperative institution Muscle Miner.

Conflicts of Interest

The authors declare that they have no conflicts of interest.


The authors would like to thank all study participants. This work was supported by the National Key R&D Program of China (grant no. 2017YFB1002504) and National Natural Science Foundation of China (nos. 81727802 and 61701404).


  1. C. S. Fry, J. D Lee, J. Mula et al., “Inducible depletion of satellite cells in adult, sedentary mice impairs muscle regenerative capacity without affecting sarcopenia,” Nature Medicine, vol. 21, no. 1, pp. 76–80, 2015. View at: Publisher Site | Google Scholar
  2. M. W. Lee, M. G. Viola, H. Meng et al., “Differential muscle hypertrophy is associated with satellite cell numbers and akt pathway activation following activin type IIB receptor inhibition in Mtm1 p.R69C mice,” The American Journal of Pathology, vol. 184, no. 6, pp. 1831–1842, 2014. View at: Publisher Site | Google Scholar
  3. H. Viola, P. M. Janssen, R. W. Grange et al., “Tissue triage and freezing for models of skeletal muscle disease,” Journal of Visualized Experiments: JoVE, vol. e51586, no. 89, 2014. View at: Google Scholar
  4. F. Liu, A. L. Mackey, R. Srikuea, K. A. Esser, and L. Yang, “Automated image segmentation of haematoxylin and eosin stained skeletal muscle cross-sections,” Journal of Microscopy, vol. 252, no. 3, pp. 275–285, 2013. View at: Publisher Site | Google Scholar
  5. H. Su, F. Xing, J. D. Lee et al., “Learning based automatic detection of myonuclei in isolated single skeletal muscle fibers using multi-focus image fusion,” in Proceedings of the IEEE International Symposium on Biomedical Imaging, pp. 432–435, San Francisco, CA, USA, April 2013. View at: Google Scholar
  6. L. R. Smith and E. R. Barton, “Smash–semi-automatic muscle analysis using segmentation of histology: a matlab application,” Skeletal Muscle, vol. 4, no. 1, pp. 1–16, 2014. View at: Publisher Site | Google Scholar
  7. F. Liu, F. Xing, Z. Zhang, M. Mcgough, and L. Yang, “Robust muscle cell quantification using structured edge detection and hierarchical segmentation,” in Lecture Notes in Computer Science, pp. 324–331, Shenzhen MICCAI, Shenzhen, China, 2015. View at: Publisher Site | Google Scholar
  8. T. Janssens, L. Antanas, S. Derde, I. Vanhorebeek, G. Van den Berghe, and F. Güiza Grandas, “Charisma: an integrated approach to automatic H&E-stained skeletal muscle cell segmentation using supervised learning and novel robust clump splitting,” Medical Image Analysis, vol. 17, no. 8, pp. 1206–1219, 2013. View at: Publisher Site | Google Scholar
  9. A. E. Carpenter, T. R. Jones, M. R. Lamprecht et al., “Cellprofiler: image analysis software for identifying and quantifying cell phenotypes,” Genome Biology, vol. 7, no. 10, pp. 1–11, 2006. View at: Publisher Site | Google Scholar
  10. A. Klemenčič, S. Kovačič, and F. Pernuš, “Automated segmentation of muscle fiber images using active contour models,” Cytometry, vol. 32, no. 4, pp. 317–326, 1998. View at: Publisher Site | Google Scholar
  11. N. Bova, V. Gál, Ó. Ibáñez, and Ó. Cordón, “Deformable models direct supervised guidance: a novel paradigm for automatic image segmentation,” Neurocomputing, vol. 177, pp. 317–333, 2016. View at: Publisher Site | Google Scholar
  12. T. F. Cootes, C. J. Taylor, D. H. Cooper et al., “Active shape models-their training and application,” Computer Vision and Image Understanding, vol. 61, no. 1, pp. 38–59, 1995. View at: Publisher Site | Google Scholar
  13. S. Zhang, Y. Zhan, M. Dewan, J. Huang, D. N. Metaxas, and X. S. Zhou, “Sparse shape composition: a new framework for shape prior modeling,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1025–1032, IEEE, Colorado Springs, CO, USA, June 2011. View at: Google Scholar
  14. D. Ciresan, A. Giusti, L. M. Gambardella, and J. Schmidhuber, “Deep neural networks segment neuronal membranes in electron microscopy images,” in Proceedings of the NIPS, pp. 2843–2851, Lake Tahoe, NV, USA, December 2012. View at: Google Scholar
  15. H. Noh, S. Hong, and B. Han, “Learning deconvolution network for semantic segmentation,” in Proceedings of the ICCV, Las Condes, Chile, December 2015. View at: Google Scholar
  16. J. Long, E. Shelhamer, and T. Darrell, “Fully convolutional networks for semantic segmentation,” in Proceedings of the CVPR, pp. 3431–3440, Santiago, Chile, December 2015. View at: Google Scholar
  17. P. O. Pinheiro, R. Collobert, and P. Dollar, “Learning to segment object candidates,” in Proceedings of the Advances in Neural Information Processing Systems, pp. 1981–1989, Montreal, Canada, December 2015. View at: Google Scholar
  18. J. Mula, J. D. Lee, F. Liu, L. Yang, and C. A. Peterson, “Automated image analysis of skeletal muscle fiber cross-sectional area,” Journal of Applied Physiology, vol. 114, no. 1, pp. 148–155, 2013. View at: Publisher Site | Google Scholar
  19. P.-Y. Baudin, N. Azzabou, P. G. Carlier, and N. Paragios, “Prior knowledge, random walks and human skeletal muscle segmentation,” in Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2012, pp. 569–576, Nice, France, October 2012. View at: Google Scholar
  20. P. Dollár and C. L. Zitnick, “Structured forests for fast edge detection,” in Proceedings of the ICCV, pp. 1841–1848, Sydney, Australia, December 2013. View at: Google Scholar
  21. F. Liu, F. Xing, and L. Yang, “Robust muscle cell segmentation using region selection with dynamic programming,” in Proceedings of the ISBI, pp. 521–524, Beijing, China, April 2014. View at: Google Scholar
  22. E. Van Aart, N. Sepasian, A. Jalba, and A. Vilanova, “Cuda-accelerated geodesic ray-tracing for fiber tracking,” Journal of Biomedical Imaging, vol. 2011, Article ID 698908, 12 pages, 2011. View at: Publisher Site | Google Scholar
  23. G. C. Kagadis, C. Kloukinas, K. Moore et al., “Cloud computing in medical imaging,” Medical Physics, vol. 40, no. 7, Article ID 070901. View at: Publisher Site | Google Scholar
  24. L. Yang, X. Qi, F. Xing, T. Kurc, J. Saltz, and D. J. Foran, “Parallel content-based sub-image retrieval using hierarchical searching,” Bioinformatics, vol. 30, no. 7, pp. 996–1002, 2014. View at: Publisher Site | Google Scholar
  25. A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” in Proceedings of the Advances in Neural Information Processing Systems, pp. 1097–1105, Lake Tahoe, NV, USA, December 2012. View at: Google Scholar
  26. H.-C. Shin, H. R. Roth, M. Gao et al., “Deep convolutional neural networks for computer-aided detection: CNN architectures, dataset characteristics and transfer learning,” IEEE Transactions on Medical Imaging, vol. 35, no. 5, pp. 1285–1298, 2016. View at: Publisher Site | Google Scholar
  27. J. Xu, X. Luo, G. Wang, H. Gilmore, and A. Madabhushi, “A deep convolutional neural network for segmenting and classifying epithelial and stromal regions in histopathological images,” Neurocomputing, vol. 191, pp. 214–223, 2016. View at: Publisher Site | Google Scholar
  28. X. Pan, L. Li, H. Yang et al., “Accurate segmentation of nuclei in pathological images via sparse reconstruction and deep convolutional networks,” Neurocomputing, vol. 229, pp. 88–99, 2017. View at: Publisher Site | Google Scholar
  29. K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” View at: Google Scholar
  30. O. Ronneberger, P. Fischer, and T. Brox, “U-net: convolutional networks for biomedical image segmentation,” in Lecture Notes in Computer Science, pp. 234–241, Shenzhen MICCAI, Shenzhen, China, 2015. View at: Publisher Site | Google Scholar
  31. S. Xie and Z. Tu, “Holistically-nested edge detection,” in Proceedings of the ICCV, pp. 1395–1403, Las Condes, Chile, December 2015. View at: Google Scholar
  32. D. Eigen and R. Fergus, “Predicting depth, surface normals and semantic labels with a common multi-scale convolutional architecture,” in Proceedings of the IEEE International Conference on Computer Vision, pp. 2650–2658, Las Condes, Chile, December 2015. View at: Google Scholar
  33. Q. Dou, H. Chen, L. Yu et al., “Automatic detection of cerebral microbleeds from mr images via 3d convolutional neural networks,” IEEE Transactions on Medical Imaging, vol. 35, no. 5, pp. 1182–1195, 2016. View at: Publisher Site | Google Scholar
  34. H. Chen, X. Qi, L. Yu, and P.-A. Heng, “Dcan: deep contour-aware networks for accurate gland segmentation,” View at: Google Scholar
  35. P. Moeskops, M. A. Viergever, A. M. Mendrik, L. S. de Vries, M. J. N. L. Benders, and I. Isgum, “Automatic segmentation of mr brain images with a convolutional neural network,” IEEE Transactions on Medical Imaging, vol. 35, no. 5, pp. 1252–1261, 2016. View at: Publisher Site | Google Scholar
  36. S. Hong, H. Noh, and B. Han, “Decoupled deep neural network for semi-supervised semantic segmentation,” in Proceedings of the Advances in Neural Information Processing Systems, pp. 1495–1503, Montreal, Canada, December 2015. View at: Google Scholar
  37. M. D. Zeiler and R. Fergus, “Visualizing and understanding convolutional networks,” in Proceedings of the Computer vision–ECCV 2014, pp. 818–833, Springer, Zurich, Switzerland, September 2014. View at: Google Scholar
  38. C. Szegedy, W. Liu, Y. Jia et al., “Going deeper with convolutions,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1–9, Boston, MA, USA, June 2015. View at: Google Scholar
  39. G. Bertasius, J. Shi, and L. Torresani, “Deepedge: a multi-scale bifurcated deep network for top-down contour detection,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4380–4389, Boston, MA, USA, June 2015. View at: Google Scholar
  40. Y. Jia, E. Shelhamer, J. Donahue et al., “Caffe: convolutional architecture for fast feature embedding,” in Proceedings of the International Conference on Multimedia, pp. 675–678, Orlando, FL, USA, November 2014. View at: Google Scholar
  41. N. Tajbakhsh, J. Y. Shin, S. R. Gurudu et al., “Convolutional neural networks for medical image analysis: full training or fine tuning?” IEEE Transactions on Medical Imaging, vol. 35, no. 5, pp. 1299–1312, 2016. View at: Publisher Site | Google Scholar
  42. J. Yosinski, J. Clune, Y. Bengio, and H. Lipson, “How transferable are features in deep neural networks?” in Proceedings of the NIPS, pp. 3320–3328, Montreal, Canada, December 2014. View at: Google Scholar
  43. M. Donoser and D. Schmalstieg, “Discrete-continuous gradient orientation estimation for faster image segmentation,” in Proceedings of the CVPR, pp. 3158–3165, Columbus, OH, USA, June 2014. View at: Google Scholar
  44. P. Arbelaez, J. Pont-Tuset, J. Barron, F. Marques, and J. Malik, “Multiscale combinatorial grouping,” in Proceedings of the CVPR, pp. 328–335, Columbus, OH, USA, June 2014. View at: Google Scholar
  45. A. Giusti, D. C. Cireşan, J. Masci, L. M. Gambardella, and J. Schmidhuber, “Fast image scanning with deep max-pooling convolutional neural networks,” 2013, View at: Google Scholar

Copyright © 2019 Lei Cui et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

More related articles

453 Views | 241 Downloads | 0 Citations
 PDF  Download Citation  Citation
 Download other formatsMore
 Order printed copiesOrder

Related articles

We are committed to sharing findings related to COVID-19 as quickly and safely as possible. Any author submitting a COVID-19 paper should notify us at to ensure their research is fast-tracked and made available on a preprint server as soon as possible. We will be providing unlimited waivers of publication charges for accepted articles related to COVID-19. Sign up here as a reviewer to help fast-track new submissions.