BioMed Research International

BioMed Research International / 2017 / Article
!A Erratum for this article has been published. To view the article details, please click the ‘Erratum’ tab above.

Research Article | Open Access

Volume 2017 |Article ID 2059036 |

Shibin Wu, Shaode Yu, Ling Zhuang, Xinhua Wei, Mark Sak, Neb Duric, Jiani Hu, Yaoqin Xie, "Automatic Segmentation of Ultrasound Tomography Image", BioMed Research International, vol. 2017, Article ID 2059036, 8 pages, 2017.

Automatic Segmentation of Ultrasound Tomography Image

Academic Editor: Cristiana Corsi
Received28 Mar 2017
Revised27 Jun 2017
Accepted07 Aug 2017
Published10 Sep 2017


Ultrasound tomography (UST) image segmentation is fundamental in breast density estimation, medicine response analysis, and anatomical change quantification. Existing methods are time consuming and require massive manual interaction. To address these issues, an automatic algorithm based on GrabCut (AUGC) is proposed in this paper. The presented method designs automated GrabCut initialization for incomplete labeling and is sped up with multicore parallel programming. To verify performance, AUGC is applied to segment thirty-two in vivo UST volumetric images. The performance of AUGC is validated with breast overlapping metrics (Dice coefficient (), Jaccard (), and False positive (FP)) and time cost (TC). Furthermore, AUGC is compared to other methods, including Confidence Connected Region Growing (CCRG), watershed, and Active Contour based Curve Delineation (ACCD). Experimental results indicate that AUGC achieves the highest accuracy ( and and ) and takes on average about 4 seconds to process a volumetric image. It was said that AUGC benefits large-scale studies by using UST images for breast cancer screening and pathological quantification.

1. Introduction

Breast cancer threatens women’s lives worldwide. It ranks as the second most common form of cancer with more than 1.3 million women diagnosed annually [1, 2]. In the United States, 12% of women will potentially develop this disease during their lifetime [1]. Consequently, breast cancer early detection is increasingly critical. Breast cancer screening plays an important role in early cancer detection, disease diagnosis, treatment planning, and therapeutic verification. In clinical applications, medical images serve as one of the primary means of breast cancer screening. Among the various modalities for breast cancer screening, mammography remains as the first choice, while supplementary modalities include hand-held ultrasound, computerized tomography (CT), and magnetic resonance imaging (MRI) [3, 4]. Of these commonly used modalities, mammography and hand-held ultrasound create two-dimensional (2D) images of the compressed breast, which leads to various deficiencies in clinical applications. Moreover, mammograms use X-ray imaging technology, exposing women to potentially harmful ionizing radiation. On the other hand, MRI provides three-dimensional (3D) images of the breast without exposure to ionizing radiation; however, its high cost prevents it from being widely adapted for breast cancer screening.

Practically, to promote an affordable and accurate 3D breast cancer screening imaging technique, Dr. Duric et al. developed a novel ultrasound tomography (UST) [5, 6]. It can scan the entire breast using ring array transducers with B-mode, which reduces breast compression and human subjectivity in image acquisition [7]. Moreover, no radiation is involved in UST imaging and breast anatomy is presented in 3D space [8]. As such, it aids in tumor differentiation in cases of obscured tumors or tumors located within dense breasts [5, 810]. In addition, UST volumetric images can be applied in breast density estimation [6, 11, 12], medicine response analysis [13], anatomical change, and breast tumor analysis [14, 15]. In summary, the image acquisition of UST is safe, cost-effective, and highly efficient. In clinical practice, breast segmentation affects follow-up image analysis for risk assessment, detection, and diagnosis [16, 17], as well as cancer treatment [18, 19]. Furthermore, extracting the breast region from surrounding water enhances tissue visualization and provides physicians and radiologist with superior understanding of breast tumor positioning [7].

To our knowledge, several algorithms have been developed toward UST image segmentation. Balic et al. [20] proposed an algorithm based on active contours, which is time consuming and its success depends on quality initial contour. Furthermore, this method is without a systematic evaluation of accuracy. Hopp et al. [21] presented a method integrating edge detection and surface fitting for breast segmentation. However, this method typically requires massive user interaction and postprocessing for outputting results. Sak et al. [22] took advantage of -means [23] and the thresholding methods [24] to reduce user interaction, though proper parameters are yet needed in these methods. Generally, major methods suffer from heavy time consumption and excessive interactions not applicable to large-scale studies. In order to overcome these issues, we proposed an approach based on GrabCut for automatic segmentation of UST images. GrabCut utilizes incomplete labeling to reduce user interaction and seeks efficient segmentation in an iterative manner of energy minimization [25]. It falls under the graph cut method [2631] and shows superiority in manual segmentation of 2D natural images, while in the presented algorithm, we provide an automatic approach for incomplete labeling of GrabCut, as well as deploying the algorithm on multicores to speed up the segmentation as demonstrated in [32].

The organization of this paper is as follows. The proposed method, experimental design, and evaluation criteria are presented in Section 2. Section 3 presents experimental results from perceived evaluation to objective evaluation. Discussion and conclusion are given in Sections 4 and 5, respectively.

2. Methods

2.1. Proposed Algorithm

As a semiautomatic algorithm, GrabCut is widely used in various scenarios [33, 34]. However, in clinical applications, an automatic method that can lighten workload and minimize subjective bias is always desirable. Figure 1 illustrates the proposed automatic method (AUGC) based on GrabCut. It integrates contrast enhancement, edge detection, convex hull searching, and curve fitting for automatic initialization of GrabCut.

To make the flowchart clearer, we present a case study shown in Figure 2. First, after contrast enhancement, the input image suppresses image content in the background and highlights breast boundaries and glandular tissues (b). Then, the major boundaries are detected in (c). After that, convex hull searching is used, and key points (red sparkers) are found (d). Generally, the boundaries of the breast in UST images are not complete to a large extent. Taking robustness into consideration, we further push these key points outward (green circle) to enclose the tissue region of interest shown in (d). Next, green circles are interpolated with Hermite curve to form a bimap (red polygon) for incomplete labeling (e). In the end, (f) shows the extracted breast region.

Image Preprocessing. This step includes contrast enhancement and boundary detection of UST images. It uses sigmoid function to sharpen the image followed by a median filter with a kernel of [33] to suppress speckle noise. As shown in Figure 2(b), this step not only suppresses background image content, but also reduces speckle noise. Moreover, it benefits edge detection as shown in Figure 2(c), because image contrast is enhanced.

Convex Hull Searching. A fast convex hull algorithm is used to determine a convex hull point set . With the point set , we calculated a centroid and a distance R between the farthest point in the point set and , as well as four extreme points of , , , and for the regions of interest. A set of points are uniformly generated on the circle centered at with the radius (). Note that outliers in Figure 2(c) are removed with morphology operation before convex hull searching.

As shown in Figure 2(c), one problem occurs because the breast boundary is incomplete. To tackle this problem, the convex hull point set and the pseudo-contour point set are combined to refine the breast mask. The algorithm is given in Algorithm 1.

Known: , , , , , ,
 for all do
  if   &   then
  else if   &   then
   for all do
   end for
  end if
 end for

Figure 2(d) shows the results after the convex hull searching and refinement, in which red sparkers are the original convex hull point set , green circles are adjusted points for generating breast mask, and the blue point is the centroid .

Closed Curve Fitting. Hermite cubic curve is powerful in smooth interpolation between control points [35] and four Hermite basis functions are described in

Moreover, the general form of Hermite curve is expressed in (2) below, where scale goes from 0 to 1 with equal spacing (). The closed curve in red is the Hermite curve with the inputting of the 36 points refined and shown in Figure 2(e). Pixels in the closed curve are the potential foreground, while outside of the curve is the definite background for GrabCut initialization.Here and represent the starting and the ending points of the curve, and and represent tangent to how the curve leaves the starting point and the ending point, respectively.

GrabCut. After closed curve fitting, Gaussian Mixture Models (GMMs) [31] are initialized with pixels inside and outside the closed curve and a flow network is built. In the network, each pixel represents a graph node. After that, a max-flow min-cut algorithm is applied for graph segmentation [27]. At last, a sample of the extracted breast is shown in Figure 2(f).

Overall, the procedure mentioned above handles only one slice in volumetric images and the entire breast UST volume is a stack of multiple gray-scale slices. As such, the proposed AUGC can be deployed with parallel programming as presented in [32]. Based on the advanced computer architecture, parallel programming can be realized on graphic processing units (GPUs) or on multicore central processing units (CPUs). GPU-based acceleration is difficult in algorithm development, in addition to costing extra programming time. On the other hand, parallel programming based on multicore CPUs is more promising as the technology is mature and comparatively easy to use. Personal computers with multicore CPUs are particularly easy to access; therefore, parallel programming based on multicore CPUs is utilized in the proposed method.

2.2. Algorithms for Comparison

Three algorithms are involved in this study. The first one, CCRG, utilizes simple statistics in region growth [36]. It calculates the median intensity and the standard deviation based on a given region. A multiplier should be supplied which defines a range around . In our experiment, the multiplier is adjusted to range from 3 to 5 times, the maximum iteration number is 500, and the seed radius is 4 mm.

The second algorithm, watershed, is a level set algorithm that classifies pixels into regions using gradient descent [37]. Additionally, a key parameter of watershed is water level (wl), tuned according to segmentation image. In our experiment, we start exploring it at 0.2. If too many small regions are obtained, we set it higher or else we tune it lower until a visually acceptable result is generated. Regarding different UST volumes, we found that water level ranges from 0.16 to 0.23. Since resultant regions are rendered by using different colors, a postprocessing step is used to merge these regions into two groups as the background and the breast region.

The final algorithm, ACCD, is derived from active contour evolution [38] and allows for control point delineation [39]. The number of control points is proportionally distributed to the region boundary length. Although basic active contour has more than ten tunable parameters, we focus on only two key parameters, and , which define the relative importance of the internal and external energy [38]. Note that we place control points near but not on the breast boundary. Compared to the original algorithm in [39], no refinement is involved.

At last, the classification and comparison of algorithms mentioned above were summarized in Table 1. For full knowledge of technical details, please refer to [3539].

CategoryInitializationTuned parameters

AUGCGrabCutCanny operator with adaptive thresholding and the structure element radii of morphological operators is initialized to 4 and 8

ACCDActive contour20 control points per slice and

WatershedLevel sets0.1wl

CCRGRegion growing10 seeds for each volume

2.3. Case Study and Evaluation

Data Collection. Thirty-two whole-breast UST volume images are collected (SoftVue™, Delphinus Medical Technologies, Michigan, USA). The size of image slice is and the physical resolution of UST volume is [0.5, 0.5, 2.0] mm3. An experienced radiologist defined the starting and ending slices following the procedure described in [14], and the average number of remaining slices in each volume is . The radiologist also manually delineated the breast region in each slice to build the ground truth for algorithm validation.

Software Platform. AUGC and CCRG are implemented with VS2010 ( in cooperation with OpenCV ( and ITK ( [40], and ACCD is previously built with MATLAB [35], while the watershed algorithm is manipulated on VolView ( All codes are running on Windows 7 workstation with 4 Intel (R) Cores (TM) of 3.70 GHz and 8 GB DDR RAM.

Accuracy Evaluation. Three criteria, Dice , Jaccard coefficients, and false positive (FP), are used to evaluate the accuracy of breast image segmentation [41]. These measures are defined in where and denote the segmentation result and the corresponding ground truth, respectively, while denotes the breast voxel number. The values of these equations range from 0 to 1. Higher values indicate better performance for and , while a value of zero is achieved when performing perfect breast volume segmentation for FP.

To evaluate the real-time capability, time cost (TC) is defined aswhere is the number of total image slices, is the number of breast volumes, and is the time cost for each volume. Note that time spent on parameters tuning and manual initialization for semiautomatic algorithms is not taken into account.

3. Results

3.1. Perceived Evaluation

Perceived evaluation of segmentation results is shown in Figure 3. From left to right are the ground truth and resultant breast regions from AUGC, ACCD, watershed, and CCRG, while from top to bottom are the coronal, sagittal, and transverse view, respectively. Note that images are cropped for display purposes. No visual difference is observed between algorithms on this case, except that watershed fails to detect bright pixels on the breast boundary and CCRG fails to segment the foreground content shown in red circles.

3.2. Quantitative Evaluation

Quantitative evaluation of all algorithms for UST image segmentation is shown in Figure 4 where different colors indicate different algorithms. Moreover, (a), (b), and (c) represent the values of , , and FP, respectively. It indicates that AUGC outperforms other algorithms, followed by ACCD. Furthermore, both AUGC and ACCD feature relatively robust values of , , and FP.

3.3. Real-Time Capability

Time cost for each slice in UST image segmentation is shown in Figure 5. Compared to manually building the ground truth (44.33 seconds per slice), all the algorithms speed up the process of breast image segmentation. Particularly, AUGC dramatically shortens time consumption and makes it possible for real-time UST image breast segmentation.

3.4. Comprehensive Performance Evaluation

Table 2 illustrates overall performance of four algorithms. It reveals that AUGC achieves the best performance, not only providing the highest volume overlap measures (D and J), but also leading to the least error (FP). In addition, AUGC demonstrates the real-time capability in image segmentation. Inferior to AUGC is ACCD. Both watershed and CCRG achieve J value less than 0.6. Additionally, CCRG produces the lowest accuracy with the highest FP.

Dice Jaccard False positive (FP)Time cost (TC)


4. Discussion

UST holds tremendous promise for breast cancer screening and examination and UST images are preferred in clinical applications, such as quantitative breast tissues analysis [5, 9, 10], breast mass growing monitoring [6, 11], and clinical pathologic diagnosis [1215]. In this paper, we presented a fully automated algorithm (AUGC) for breast UST image segmentation. The performance of four segmentation algorithms has been verified based on thirty-two volumetric images.

Quantitative evaluation of segmentation performance suggests that AUGC is superior to other three algorithms, ACCD, watershed, and CCRG, shown in Figures 3 and 4 and Table 2. Among these methods, CCRG resulted in the lowest accuracy and the highest amount of false positives. Moreover, watershed produced background content onto the final results. On the whole, ACCD is slightly inferior to AUGC. However, ACCD requires a user to locate several control points in the breast boundary. In addition, it contains more than ten parameters which need to be tuned manually, making the segmentation complicated and exhaustive. Generally, AUGC obtains the best performance in terms of segmentation accuracy.

AUGC is also superior to other approaches in terms of the real-time capacity. It can isolate an entire UST volumetric image within four seconds () on a four-core CPU system. Therefore, the greater the number of CPUs is, the less segmentation time it needs. It is known that real-time breast extraction plays a critical role in practical applications. For instance, breast density estimation is a routine task before rating breast cancer risk. At present, manual extraction of the whole breast in UST image hampers its large-scale experiments. Consequently, the proposed AUGC paves the way for large-scale studies in terms of high accuracy and real-time speed. It can accelerate the application of UST in anatomical change quantification, medicine response, and other related tasks.

The UST imaging technology is still under development and remarkable improvements have been made recently [42, 43]. These improved technologies are bound to enhance UST image quality and tissue contrast. High UST image quality can improve the performance of AUGC in breast segmentation, suggesting an even greater potential of AUGC to facilitate clinical diagnosis by using whole-breast UST images

5. Conclusion

UST image segmentation not only is time consuming, but also requires massive user interaction. An automated algorithm based on GrabCut is proposed and verified in this study. Experimental results have validated its good performance in UST image segmentation. Furthermore, it can segment one slice within less than 0.3 seconds. It is beneficial for large-scale studies and physicians can also be released from the tedious task of UST image segmentation.

Conflicts of Interest

The authors declare that they have no conflicts of interest.


This work is supported by the grants of the National Key Research Program of China (Grant no. 2016YFC0105102), the Union of Production, Study and Research Project of Guangdong Province (Grant no. 2015B090901039), the Technological Breakthrough Project of Shenzhen City (Grant no. JSGG20160229203812944), the Shenzhen Fundamental Research Program (JCYJ201500731154850923), the Natural Science Foundation of Guangdong Province (Grant no. 2014A030312006), and the CAS Key Laboratory of Human-Machine Intelligence-Synergy Systems, Shenzhen Institutes of Advanced Technology.


  1. C. Desantis, J. Ma, L. Bryan, and A. Jemal, “Breast cancer statistics, 2013,” CA: A Cancer Journal for Clinicians, vol. 64, no. 1, pp. 52–62, 2014. View at: Publisher Site | Google Scholar
  2. L. Fan, K. Strasser-Weippl, J.-J. Li et al., “Breast cancer in China,” The Lancet Oncology, vol. 15, no. 7, pp. e279–e289, 2014. View at: Publisher Site | Google Scholar
  3. L. Tabar, B. P. Dean, T. H. Chen et al., “The impact of mammography screening on the diagnosis and management of early-phase breast cancer,” in Breast Cancer: A New Era in Management, chapter 2, pp. 31–78, 2014. View at: Google Scholar
  4. N. C. Lee, F. L. Wong, P. M. Jamison et al., “Implementation of the national breast and cervical cancer early detection program: the beginning,” Cancer, vol. 120, no. 16, pp. 2540–2548, 2014. View at: Publisher Site | Google Scholar
  5. N. Duric, P. Littrup, L. Poulo et al., “Detection of breast cancer with ultrasound tomography: First results with the Computed Ultrasound Risk Evaluation (CURE) prototype,” Medical Physics, vol. 34, no. 2, pp. 773–785, 2007. View at: Publisher Site | Google Scholar
  6. C. Glide, N. Duric, and P. Littrup, “Novel approach to evaluating breast density utilizing ultrasound tomography,” Medical Physics, vol. 34, no. 2, pp. 744–753, 2007. View at: Publisher Site | Google Scholar
  7. N. Duric, P. Littrup, S. Schmidt et al., “Breast imaging with the SoftVue imaging system: First results,” in Proceedings of the Medical Imaging 2013: Ultrasonic Imaging, Tomography, and Therapy, vol. 8675, pp. 1–8, Lake Buena Vista, Fla, USA, February 2013. View at: Publisher Site | Google Scholar
  8. N. Duric, P. Littrup, C. Li et al., “Breast imaging with SoftVue: Initial clinical evaluation,” in Proceedings of the Medical Imaging 2014: Ultrasonic Imaging and Tomography, vol. 9040, San Diego, Calif, USA, February 2014. View at: Publisher Site | Google Scholar
  9. C. Li, N. Duric, P. Littrup, and L. Huang, “In vivo Breast Sound-Speed Imaging with Ultrasound Tomography,” Ultrasound in Medicine and Biology, vol. 35, no. 10, pp. 1615–1628, 2009. View at: Publisher Site | Google Scholar
  10. B. Ranger, P. J. Littrup, N. Duric et al., “Breast ultrasound tomography versus MRI for clinical display of anatomy and tumor rendering: Preliminary results,” American Journal of Roentgenology, vol. 198, no. 1, pp. 233–239, 2012. View at: Publisher Site | Google Scholar
  11. C. K. Glide-Hurst, N. Duric, and P. Littrup, “Volumetric breast density evaluation from ultrasound tomography images,” Medical Physics, vol. 35, no. 9, pp. 3988–3997, 2008. View at: Publisher Site | Google Scholar
  12. N. Duric, N. Boyd, P. Littrup et al., “Breast density measurements with ultrasound tomography: A comparison with film and digital mammography,” Medical Physics, vol. 40, no. 1, Article ID 013501, 2013. View at: Publisher Site | Google Scholar
  13. M. Sak, N. Duric, P. Littrup et al., “Breast density measurements using ultrasound tomography for patients undergoing tamoxifen treatment,” in Proceedings of the Medical Imaging 2013: Ultrasonic Imaging, Tomography, and Therapy, vol. 8675, pp. 1–8, Lake Buena Vista, Fla, USA, February 2013. View at: Publisher Site | Google Scholar
  14. Z. G. Khodr, M. A. Sak, R. M. Pfeiffer et al., “Determinants of the reliability of ultrasound tomography sound speed estimates as a surrogate for volumetric breast density,” Medical Physics, vol. 42, no. 10, pp. 5671–5678, 2015. View at: Publisher Site | Google Scholar
  15. E. O'Flynn, J. Fromageau, M. Ledger et al., “Breast density measurements with ultrasound tomography: a comparison with non-contrast MRI,” Breast Cancer Research, vol. 17, no. S1, 2015. View at: Publisher Site | Google Scholar
  16. C. G. Ortiz and A. L. Martel, “Automatic atlas-based segmentation of the breast in MRI for 3D breast volume computation,” Medical Physics, vol. 39, no. 10, pp. 5835–5848, 2012. View at: Publisher Site | Google Scholar
  17. M. Lin, J.-H. Chen, X. Wang, S. Chan, S. Chen, and M.-Y. Su, “Template-based automatic breast segmentation on MRI by excluding the chest region,” Medical Physics, vol. 40, no. 12, Article ID 122301, 2013. View at: Publisher Site | Google Scholar
  18. L. Gao, W. Yang, Z. Liao, X. Liu, Q. Feng, and W. Chen, “Segmentation of ultrasonic breast tumors based on homogeneous patch,” Medical Physics, vol. 39, no. 6, pp. 3299–3318, 2012. View at: Publisher Site | Google Scholar
  19. M. Mustra, M. Grgic, and R. M. Rangayyan, “Review of recent advances in segmentation of the breast boundary and the pectoral muscle in mammograms,” Medical and Biological Engineering and Computing, vol. 54, no. 7, pp. 1003–1024, 2016. View at: Publisher Site | Google Scholar
  20. I. Balic, P. Goyal, O. Roy, and N. Duric, “Breast boundary detection with active contours,” in Proceedings of the Medical Imaging 2014: Ultrasonic Imaging and Tomography, vol. 9040, pp. 1–8, San Diego, Calif, USA, February 2014. View at: Publisher Site | Google Scholar
  21. T. Hopp, M. Zapf, and N. V. Ruiter, “Segmentation of 3D ultrasound computer tomography reflection images using edge detection and surface fitting,” in Proceedings of the Medical Imaging 2014: Ultrasonic Imaging and Tomography, vol. 9040, San Diego, Calif, USA, February 2014. View at: Publisher Site | Google Scholar
  22. M. Sak, N. Duric, P. Littrup et al., “Using speed of sound imaging to characterize breast density,” Ultrasound in Medicine and Biology, vol. 43, no. 1, pp. 91–103, 2017. View at: Publisher Site | Google Scholar
  23. A. K. Jain, “Data clustering: 50 years beyond K-means,” Pattern Recognition Letters, vol. 31, no. 8, pp. 651–666, 2010. View at: Publisher Site | Google Scholar
  24. N. Otsu, “A threshold selection method from gray-level histograms,” IEEE Transactions on Systems, Man and Cybernetics, vol. 11, no. 285-296, pp. 23–27, 1975. View at: Google Scholar
  25. C. Rother, V. Kolmogorov, and A. Blake, ““GrabCut”: interactive foreground extraction using iterated graph cuts,” ACM Transactions on Graphics, vol. 23, no. 3, pp. 309–314, 2004. View at: Publisher Site | Google Scholar
  26. Y. Y. Boykov and M.-P. Jolly, “Interactive graph cuts for optimal boundary & region segmentation of objects in N-D images,” in Proceedings of the 8th International Conference on Computer Vision (ICCV '01), vol. 1, pp. 105–112, Vancouver, Canada, July 2001. View at: Publisher Site | Google Scholar
  27. Y.-Y. Chuang, B. Curless, D. H. Salesin, and R. Szeliski, “A Bayesian approach to digital matting,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR '01), pp. II264–II271, December 2001. View at: Google Scholar
  28. Y. Boykov, O. Veksler, and R. Zabih, “Fast approximate energy minimization via graph cuts,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 23, no. 11, pp. 1222–1239, 2001. View at: Publisher Site | Google Scholar
  29. Y. Boykov and G. Funka-Lea, “Graph cuts and efficient N-D image segmentation,” International Journal of Computer Vision, vol. 70, no. 2, pp. 109–131, 2006. View at: Publisher Site | Google Scholar
  30. A. Blake, C. Rother, M. Brown, P. Perez, and P. Torr, “Interactive image segmentation using an adaptive GMMRF model,” in Proceedings of the European Conference on Computer Vision, vol. 3021 of Lecture Notes in Computer Science, pp. 428–441, Springer, 2004. View at: Publisher Site | Google Scholar
  31. J. Carreira and C. Sminchisescu, “CPMC: Automatic object segmentation using constrained parametric min-cuts,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 34, no. 7, pp. 1312–1328, 2012. View at: Publisher Site | Google Scholar
  32. G. Wang, M. A. Zuluaga, R. Pratt et al., “Slic-Seg: Slice-by-Slice segmentation propagation of the placenta in fetal MRI using one-plane scribbles and online learning,” in Proceedings of the 18th International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI '15), vol. 9351 of Lecture Notes in Computer Science, pp. 29–37, 2015. View at: Publisher Site | Google Scholar
  33. S. Han, W. Tao, D. Wang, X.-C. Tai, and X. Wu, “Image segmentation based on GrabCut framework integrating multiscale nonlinear structure tensor,” IEEE Transactions on Image Processing, vol. 18, no. 10, pp. 2289–2302, 2009. View at: Publisher Site | Google Scholar | MathSciNet
  34. I. Na, K. Oh, and S. Kim, “Unconstrained object segmentation using grabcut based on automatic generation of initial boundary,” International Journal of Contents, vol. 9, no. 1, pp. 6–10, 2013. View at: Publisher Site | Google Scholar
  35. W. Zhou and Y. Xie, “Interactive contour delineation and refinement in treatment planning of image-guided radiation therapy,” Journal of Applied Clinical Medical Physics, vol. 15, no. 1, pp. 1–26, 2014. View at: Google Scholar
  36. M. Kass, A. Witkin, and D. Terzopoulos, “Snakes: active contour models,” International Journal of Computer Vision, vol. 1, no. 4, pp. 321–331, 1988. View at: Publisher Site | Google Scholar
  37. K. Martin, L. Ibáñez, L. Avila, S. Barré, and J. H. Kaspersen, “Integrating segmentation methods from the Insight Toolkit into a visualization application,” Medical Image Analysis, vol. 9, no. 6, pp. 579–593, 2005. View at: Publisher Site | Google Scholar
  38. P. A. Yushkevich, J. Piven, H. C. Hazlett et al., “User-guided 3D active contour segmentation of anatomical structures: significantly improved efficiency and reliability,” NeuroImage, vol. 31, no. 3, pp. 1116–1128, 2006. View at: Publisher Site | Google Scholar
  39. Y. Gao, R. Kikinis, S. Bouix, M. Shenton, and A. Tannenbaum, “A 3D interactive multi-object segmentation tool using local robust statistics driven active contours,” Medical Image Analysis, vol. 16, no. 6, pp. 1216–1227, 2012. View at: Publisher Site | Google Scholar
  40. H. J. Johnson, M. McCormick, and L. Ibanez, “The ITK Software Guide,” ITK version 4.10, 2016. View at: Google Scholar
  41. S. Choi, S. H. Cha, and C. Tappert, “A survey of binary similarity and distance measures,” Journal of Systemics, Cybernetics and Informatics, pp. 43–48, 2010. View at: Google Scholar
  42. K. Wang, T. Matthews, F. Anis, C. Li, N. Duric, and M. Anastasio, “Waveform inversion with source encoding for breast sound speed reconstruction in ultrasound computed tomography,” IEEE Transactions on Ultrasonics, Ferroelectrics, and Frequency Control, vol. 62, no. 3, pp. 475–493, 2015. View at: Publisher Site | Google Scholar
  43. G. Y. Sandhu, C. Li, O. Roy, S. Schmidt, and N. Duric, “Frequency domain ultrasound waveform tomography: Breast imaging using a ring transducer,” Physics in Medicine and Biology, vol. 60, no. 14, article no. 5381, pp. 5381–5398, 2015. View at: Publisher Site | Google Scholar

Copyright © 2017 Shibin Wu et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

More related articles

 PDF Download Citation Citation
 Download other formatsMore
 Order printed copiesOrder

Related articles

Article of the Year Award: Outstanding research contributions of 2020, as selected by our Chief Editors. Read the winning articles.