Research Article  Open Access
Patient Specific Dosimetry Phantoms Using Multichannel LDDMM of the Whole Body
Abstract
This paper describes an automated procedure for creating detailed patientspecific pediatric dosimetry phantoms from a small set of segmented organs in a child's CT scan. The algorithm involves full body mappings from adult template to pediatric images using multichannel large deformation diffeomorphic metric mapping (MCLDDMM). The parallel implementation and performance of MCLDDMM for this application is studied here for a sample of 4 pediatric patients, and from 1 to 24 processors. 93.84% of computation time is parallelized, and the efficiency of parallelization remains high until more than 8 processors are used. The performance of the algorithm was validated on a set of 24 male and 18 female pediatric patients. It was found to be accurate typically to within 12 voxels (2–4 mm) and robust across this large and variable data set.
1. Introduction
Measuring the radiation dose a patient accumulates through life is an important matter that has been receiving much attention recently, in particular for growing children (e.g., in the New England Journal of Medicine's recent critique of CT use [1], and the adoption of the Image Gently program [2] by the Society of Pediatric Radiology, the American Society of Radiologic Technologists, the American College of Radiology, the American Association of Physicists in Medicine, and others). While directly measuring dose to individual organs is impractical, the development of computational phantoms containing dosimetric information (e.g., [3]), such as the extended cardiactorso (XCAT) phantom used in this study [4] have begun to be a reliable substitute. A key shortcoming of this strategy is that standard phantoms cannot adequately reflect variability between patients, especially for children of different sizes and ages, and defining new phantoms for each patient manually would be unfeasible. The strategy used here consists of manually segmenting a small subset of organs from pediatric CT data and calculating a full body mapping to a similarly segmented adult XCAT phantom [5]. The resulting transformation is used to map rich anatomical and dosimetric information to the child's body.
To map dense image data as well as pointbased manifold data between adult and child, this application requires a smooth invertible transformation (a diffeomorphism) to be defined everywhere on the background space of the CT scan. Such transformations are an important focus of computational anatomy [6], where anatomical variability is understood by studying diffeomorphisms mapping anatomical manifolds to one another. Formally, anatomy is modelled as the quadruple , where is the background space (i.e., subsets of ), is a group of diffeomorphisms on , is the orbit of a template under , and is a family of probability measures on . Geodesic paths, for , are used to evolve a template according to , and a mapping to a target is defined when .
Large deformation diffeomorphic metric mapping (LDDMM) [7] generates such mappings () by integrating a smooth time dependent velocity field [8], with the initial condition being identity, . A functional of the velocity field, which enforces image matching as well as smoothness and ensures the path is a geodesic, is minimized as discussed below.
2. The Multichannel LDDMM Algorithm
There are existing algorithms for full body image registration, which are used (e.g.) in registering PET to CT data [9–11] and compensating for deformations such as breath holds. However, these tend to use elastic models, which are suitable for describing the small deformations that register two images of the same patient but are unable to accurately describe the widely varying deformations between adults and children of various ages. In addition to the constraints on smoothness and invertibility, transformations generated by LDDMM are well suited to this application, because its fluid model (rather than elastic) allows for large deformations to be generated [12] and because the submanifold preserving property of diffeomorphisms [13] allows a transformation calculated from a handful of segmented structures to be accurately applied to the thousands of anatomical structures defined in the XCAT phantom. Moreover, additional properties are well suited to future exploration. For example, LDDMM allows metric distances to be defined between template and target anatomies [8, 14] and allows statistical codification of anatomy [15, 16].
In this work, we use multichannel LDDMM (MCLDDMM), an algorithm which treats each segmented organ as a separate image linked by a common background space [17] to calculate diffeomorphisms. This is accomplished by calculating the velocity field minimizing the energy functional where and are the (out of ) channels (organs) of the target and template images, is a diffeomorphism generated by integrating the velocity field from to 1, and describes the contribution of the channel to the overall energy. The operator , where is fixed and is varied, is identity, and is the Laplacian operator, ensures smoothness of the velocity field and resulting deformations, with larger corresponding to smoother deformations, and smaller corresponding to more accurate transformations.
The energy gradient can be computed as [17] where is the operator inverse of , denotes determinant and denotes the Jacobian. The transformation generated by integrating (1) from time to time is denoted (i.e., ). The quantity is the template channel transformed up to time (i.e., ), is the target channel transformed backwards from time 1 to time (i.e., ), and is simply the spatial gradient.
It can be seen that the transformation and its inverse must be defined at all times, which was discretized here into 11 equally spaced time points from to . Calculating this transformation from the velocity field is a large part of the computational load. Integration in time is performed using semiLagrangian advection, a technique used in numerical weather prediction [18]. We use an implicit method for numerical integration, with 3 iterations per voxel at each timestep.
Moreover, a deformed target and template image must be calculated at each timestep. We use trilinear interpolation, which corresponds to another large computational load. To optimize calculations, the images for each channel were computed in the same loop (loop fusion).
Finally, application of the operator is implemented by multiplication in the Fourier domain. The FFT calculations were performed and parallelized using Intel Math Kernel Library's (MKL) FFT routines.
Since many steps of this algorithm involve independent calculations on a regular 3D voxel grid, it is well suited to parallelization. In our C++ implementation of the LDDMM algorithm, OpenMP (open multiprocessing) library routines were used. As stated in [19], “the OpenMP Application Program Interface (API) supports multiplatform sharedmemory parallel programming in C/C++ and Fortran on all architectures. …OpenMP is a portable, scalable model that gives sharedmemory parallel programmers a simple and flexible interface for developing parallel applications for platforms ranging from the desktop to the supercomputer.” In our algorithm, at each iteration of gradient descent, different operations defined on data over the voxel grid were parallelized using worksharing constructs, and loop iterations were split among the threads. The program was compiled using Intel C++ compiler version 12.0, with automatic compiler optimizations. It was run on a Dell R900, a 4 socket node with 6 cores per socket, with an Intel Xeon CPU E7450 at 2.40 GHz.
3. Methods
3.1. Calculation of Full Body Maps
In previous work [5], the feasibility of using multiMCLDDMM for this purpose was explored. A mapping to a single pediatric patient was calculated, and a reasonable subset of segmented organs was determined. However, generalizing this algorithm to a population of patients proved difficult. For example, where initial overlap of organs or bony details between template and target was poor, the diffeomorphism tended to shrink organs close to a point. Such distortions would also negatively affect the registration of nearby structures. Furthermore, when structures were shrunk by the diffeomorphism details were lost, and when structures were expanded, their initial voxelized character was spuriously reproduced at the larger scale. These difficulties are illustrated by showing a deformed adult template in Figure 1, where abdominal organs are seen contracting to a very small size, nearby structures in the neck and thorax are distorted, and features in the face and skull are lost. Further investigation resulted in the algorithm being made more robust [20] but at the expense of increased computation time.
In the modified MCLDDMM algorithm, (2) is minimized by initializing the velocity field to 0 and using a gradient descent routine with a large value of . At convergence, the value of is decreased, and minimization resumes, starting with the previously calculated velocity field. This procedure is iterated a total of four times. This sequential reduction of the parameter (denoted “cascading ”) allows for a coarse to fine registration and is responsible for the increased robustness as well as increased computation time of the modified algorithm. Beginning with a large value of is analogous to Tikhonov regularization, encouraging a desirable solution to an ill posed problem. The final small value for is chosen to give the desired level of accuracy in our mapping. Decreasing the value for abruptly often results in nondiffeomorphic transformations due to numerical instability. So, we include 2 intermediate values to mitigate this effect and unfortunately must bear the price of considerably increasing computation time.
The MCLDDMM algorithm with cascading was used to generate mappings between one of two adult templates (one male and one female), and pediatric patients (24 male and 18 female). Each was defined on an 2mm^{3} voxel grid. The patients varied in size between 0.072 and 0.472 times the volume of the adult, with an average of 0.233 times. Males ranged from 0.072 to 0.472 times the adult volume with a mean of 0.246, while females ranged from 0.076 to 0.372 times the adult volume with a mean of 0.215. The images were segmented into 8 channels with corresponding organs and weightings defined in Table 1, and 87 landmarks were placed automatically [4] mainly on easily reproducible bony structures. Images were initially aligned with an affine transformation minimizing distances between corresponding landmarks, followed by nonlinear landmark LDDMM [21]. Following this, cascading MCLDDMM was used with the four values . In previous work, we found this particular sequence to give qualitatively good results in 2D simulations and 3D full body data [20].

The sequence of transformations used to generate the final mapping is illustrated in Figure 2. Each transformations for each pediatric patient were combined to yield a double precision displacement vector at each voxel of the adult template images. This transformation was trilinearly interpolated to map NURBS (nonuniform rational Bspline) surfaces defined in the XCAT phantoms to the coordinate system of the child.
3.2. Analysis of Computation
The bulk of the computational work was performed during cascading MCLDDMM, and as such, its performance was investigated more thoroughly. Four patients were selected, 2 males and 2 females, corresponding to the largest, smallest, and 1/3 interquartile sizes, denoted “small”, “medsmall”, “medlarge”, and “large”. Mappings were calculated on these patients using each of 1, 2, 4, 8, 16, and 24 (the maximum readily available) processors. The total computation time excluding inputoutput (IO) operations was analyzed for each case as well as the time spent in specific functions. This allowed us an understanding of how computation time scales with the number of processors used, and in particular identify at what point computation time begins to increase beyond what would be expected.
To be more thorough, the portions of the program that were affected by parallelization, including IO operations, were analyzed. Speedup, , due to parallelization on processors was calculated (using “Amdahl's Law” [22] as in [23]) to be where is the total computation time for processors, and for a single processor, is the time spent that cannot be parallelized, and is the time spent that can be parallelized. These two quantities are easily estimated from a two parameter fit to the above equation, which allows determination of the fraction of the total computational time that can be parallelized. Furthermore, the efficiency of parallelization was calculated according to
3.3. Accuracy of Mappings
Finally, the quality of the mappings produced was validated. For each segmented organ, a triangulated surface was produced using isosurface generation via marching tetrahedra [24]. For each template (target) vertex, the minimum distance to a vertex on the target (template) surface was measured. Distances for template and target vertices were combined, and their distributions were analyzed. Breaking down this analysis into categories allows an understanding of the robustness of the algorithm. As such, distributions were analyzed separately for males, females, as well as for each segmented organ.
4. Results
4.1. Computational Performance
A summary of the 4 subjects used to analyze computation performance is included in Table 2. The number of voxels in each image is shown in the second column, giving more precise meaning to the labels “small”, “medsmall”, “medlarge”, and “large”. The total number of iterations of gradient descent across the 4 applications of MCLDDMM is shown in the third column. Due in part to adaptive stepsize selection in gradient descent, the number of iterations until convergence cannot be known before hand. In the fourth column, the product between number of voxels and number of iterations is shown as a rough approximation of the number of calculations used. This value can be used to better understand the timing results that follow. In particular the “medsmall” case required the most iterations to converge, and the approximate number of calculations was much less for the “small” patient than for the other three. We stress that these four patients were chosen with interquartile spacing of their total number of voxels, as opposed to uniform spacing across number of voxels, or uniform spacing across number of calculations. Such a choice is reflective of the pediatric population to be examined, rather than properties of the algorithm itself.

The total computational time in hours, excluding IO operations, is shown in Table 3. The two largest components of calculations are also shown. Numerically integrating the velocity field using semiLagrangian interpolation is shown in Table 4, and trilinearly interpolating the images is shown in Table 5. Surprisingly, the longest amount of time was spent on the “medsmall case”. While this is partially explained by the large number of iterations for this case shown in Table 2, other factors such as the specific implementation of the fast Fourier transform on a grid of this size, contribute as well.



To better understand this behavior, the same data is shown graphically, on a loglog axes in Figure 3. Figure 3(a) shows the total time, Figure 3(b) shows the time spent calculating semiLagrangian advection, and Figure 3(c) shows the time spent interpolating images. It appears that computation time scales with number of processors up until around 8, when efficiency starts to break down.
(a)
(b)
(c)
Again, this data must be interpreted with caution, because the images used are different sizes and a different number of iterations of gradient descent is required to converge in each case, as shown in Table 2. Therefore, the timing data was also plotted after being normalized by total number of voxels times total number of iterations in Figure 4. It should be noted that the smallest image actually takes the most time per voxel per iteration, while the largest image takes the least.
(a)
(b)
(c)
The speedup factor and efficiency were calculated according to (4) and (5) and are plotted in Figure 5. This analysis confirms and quantifies the sharp drop in efficiency beyond 8 processors. From a 2 parameter fit to the data in Figure 5(a), it was determined that of the computation time is parallelized, demonstrating the effectiveness of our implementation.
(a)
(b)
4.2. Accuracy of Transformations
To give a qualitative understanding of the mappings produced, an example of triangulated surfaces, for target and mapped template, are shown in Figure 6 with the body in Figure 6(a), the bones in Figure 6(b), and the other organs in Figure 6(c). One can see the quality of the mappings is good in most areas, the exceptions being the inferiormost regions, where the extent of template and target images vary, the scapula, where sliding motions between the nearby ribs and body surface are difficult to generate given the diffeomorphism constraint, and the sharp borders of some abdominal organs, whose curvature varies markedly from that of the template.
The mappings produced were used to generate customized dosimetry phantoms based on the adult XCATs. The adult male XCAT is shown in Figure 6(d) and an example pediatric dosimetry phantom is shown in Figure 6(e). Previous work has shown dosimetry measurements generated with these phantoms to agree within 10% percent of ground truth [5].
Cumulative distribution functions for final surface to surface distances are shown in Figure 7. They are shown for all patients pooled together as well as for males and females separately in Figure 7(a). The differences in accuracy, on average, between male and female patients is negligible. Additionally, distribution functions are shown for each organ in Figure 7(b). And they are shown for each of the 42 patients in Figure 7(c).
(a)
(b)
(c)
The results show that the majority of surfaces (a fraction of standard deviation of the vertices) agree within 2–4 mm or 12 voxels. Moreover, accuracy for females tends to be more variable than that for males, likely due to larger differences in body proportions between child and adult. Surprisingly, the least accurate case, apparent in Figure 7(c), is an average seeming patient of intermediate size between the medsmall and medlarge test cases. Furthermore, differences in accuracy for each organ are observed, where the brain is matched with the most fidelity and the stomach followed by lungs with the least fidelity. While these differences are small when compared to the voxel size, it is worth noting that the relatively poor performance for the stomach was likely due to its internal location and close proximity with many other abdominal structures, and the relatively poor performance of the lungs was likely due to large differences in curvature between the adult and child at the apexes and inferior borders.
5. Conclusions
This work presented an interesting application of diffeomorphic image registration, generating pediatric patient specific detailed dosimetry phantoms, made feasible on large scale due to parallel computing. The need for parallelization in deformable image registration is well recognized [23, 25, 26], and other authors have investigated parallelization of diffeomorphic registration from MASPAR [27] to GPU implementations [28].
The algorithm used here for generating full body maps involves a sequence of increasingly detailed transformations between adult templates and child images. This procedure ensures robustness to automate calculations across a wide range of pediatric patients but comes at the price high computational cost.
To overcome this cost, 93.84% of the algorithm computation time was parallelized. Running times for the various patients examined ranged from over 30 hours on a single processor to under 1 hour on 24 processors in parallel. An analysis of speedup and parallelization efficiency shows that performance begins to rapidly decline when implemented on more than 8 processors. As applications for LDDMM become more numerous and larger scale, an investigation of this issue will be necessary. It is likely that the effects of memory to cpu communication bandwidth, load balancing overhead (due to workload not evenly distributed across the available processors) play a major role.
The full body mapping algorithm is quite accurate for all the patients examined, with the majority of vertices defined on organ surfaces agreeing between template and target to within 2 voxels. Overcoming a main drawback of the diffeomorphism constraint, namely, forbidding sliding motions in the deformation, is the subject of ongoing research. One strategy we are currently investigating involves relabelling a strip of the segmented image, between two structures where sliding would be expected, as “background”. The XCAT phantoms generated are being further investigated for their accuracy and clinical utility.
While generating mappings using a sequence of transformations results in a robust algorithm for this application, it detracts from some of the theoretical appeal of LDDMM. Describing transformations by a single timedependent vector field allows a rigorous study of anatomical variability. Future work will involve combining these transformations, for example, as described in [29], and beginning to engage in shape analysis of full bodies.
Acknowledgments
The authors would like to extend thanks to Joseph Hennessey for the development of a parallel visualization application essential for this study, to Mike Bowers for developing and customizing an efficient parallelized implementation of landmark based LDDMM, and to Timothy Brown for assistance with computational infrastructure. D. J. Tward was supported by the the JuliePayette NSERC Research Scholarship (Canada). The authors gratefully acknowledge the support of NIH Grants nos. 1S10RR02505301, R01EB001838, and P41RR015241.
References
 D. J. Brenner and E. J. Hall, “Computed tomography—an increasing source of radiation exposure,” The New England Journal of Medicine, vol. 357, no. 22, pp. 2277–2284, 2007. View at: Publisher Site  Google Scholar
 The Alliance for Radiation Safety in Pediatric Imaging, “Image gently.”, 2009, http://www.pedrad.org/associations/5364/ig.
 C. Lee, D. Lodwick, J. L. Williams, and W. E. Bolch, “Hybrid computational phantoms of the 15year male and female adolescent: applications to CT organ dosimetry for patients of variable morphometry,” Medical Physics, vol. 35, no. 6, pp. 2366–2382, 2008. View at: Publisher Site  Google Scholar
 W. P. Segars, M. Mahesh, T. J. Beck, E. C. Frey, and B. M. W. Tsui, “Realistic CT simulation using the 4D XCAT phantom,” Medical Physics, vol. 35, no. 8, pp. 3800–3808, 2008. View at: Publisher Site  Google Scholar
 W. P. Segars et al., “Patient specific computerized phantoms to estimate dose in pediatric CT,” in Proceedings of the Medical Imaging: Physics of Medical Imaging, vol. 7258 of Proceedings of SPIE, p. 72580H, Lake Buena Vista, Fla, USA, February 2009. View at: Google Scholar
 U. Grenander and M. I. Miller, “Computational anatomy: an emerging discipline,” Quarterly of Applied Mathematics, vol. 56, no. 4, pp. 617–694, 1998. View at: Google Scholar
 M. F. Beg, M. I. Miller, A. Trouvé, and L. Younes, “Computing large deformation metric mappings via geodesic flows of diffeomorphisms,” International Journal of Computer Vision, vol. 61, no. 2, pp. 139–157, 2005. View at: Publisher Site  Google Scholar
 M. I. Miller, A. Trouvé, and L. Younes, “On the metrics and EulerLagrange equations of computational anatomy,” Annual Review of Biomedical Engineering, vol. 4, pp. 375–405, 1998. View at: Google Scholar
 R. Shekhar, V. Walimbe, S. Raja et al., “Automated 3dimensional elastic registration of wholebody PET and CT from separate or combined scanners,” Journal of Nuclear Medicine, vol. 46, no. 9, pp. 1488–1496, 2005. View at: Google Scholar
 C. Cohade, M. Osman, L. T. Marshall, and R. L. Wahl, “PETCT: accuracy of PET and CT spatial registration of lung lesions,” European Journal of Nuclear Medicine and Molecular Imaging, vol. 30, no. 5, pp. 721–726, 2003. View at: Google Scholar
 P. J. Slomka, D. Dey, C. Przetak, U. E. Aladl, and R. P. Baum, “Automated 3Dimensional registration of standalone 18FFDG wholebody PET with CT,” Journal of Nuclear Medicine, vol. 44, no. 7, pp. 1156–1167, 2003. View at: Google Scholar
 G. E. Christensen, R. D. Rabbitt, and M. I. Miller, “Deformable templates using large deformation kinematics,” IEEE Transactions on Image Processing, vol. 5, no. 10, pp. 1435–1447, 1996. View at: Google Scholar
 W. Boothby, An introduction to Differentiable Manifolds and Riemannian Geometry, Acedemic Press, 1986.
 M. I. Miller, A. Trouvé, and L. Younes, “Geodesic shooting for computational anatomy,” Journal of Mathematical Imaging and Vision, vol. 24, no. 2, pp. 209–228, 2006. View at: Publisher Site  Google Scholar
 M. Vaillant, M. I. Miller, L. Younes, and A. Trouvé, “Statistics on diffeomorphisms via tangent space representations,” NeuroImage, vol. 23, no. 1, pp. S161–S169, 2004. View at: Publisher Site  Google Scholar
 J. Ma, M. I. Miller, A. Trouvé, and L. Younes, “Bayesian template estimation in computational anatomy,” NeuroImage, vol. 42, no. 1, pp. 252–261, 2008. View at: Publisher Site  Google Scholar
 C. Ceritoglu, K. Oishi, X. Li et al., “Multicontrast large deformation diffeomorphic metric mapping for diffusion tensor imaging,” NeuroImage, vol. 47, no. 2, pp. 618–627, 2009. View at: Publisher Site  Google Scholar
 A. Staniforth and J. Cote, “SemiLagrangian integration schemes for atmospheric models  a review,” Monthly Weather Review, vol. 119, no. 9, pp. 2206–2223, 1991. View at: Google Scholar
 OpenMP, “The OpenMP API specification for parallel programming,” 2011, http://openmp.org/wp/. View at: Google Scholar
 D. J. Tward, C. Ceritoglu, G. Sturgeon, W. P. Segars, M. I. Miller, and J. T. Ratnanather, “Generating patientspecific dosimetry phantoms with wholebody diffeomorphic image registration,” in Proceedings of the IEEE 37th Annual Northeast Bioengineering Conference, Troy, NY, USA, April 2011. View at: Google Scholar
 S. C. Joshi and M. I. Miller, “Landmark matching via large deformation diffeomorphisms,” IEEE Transactions on Image Processing, vol. 9, no. 8, pp. 1357–1370, 2000. View at: Publisher Site  Google Scholar
 G. M. Amdahl, “Validity of the singleprocessor approach to achieving large scale computing capabilities,” Proceedings in AFIPS Converence, vol. 7, pp. 483–485, 1967. View at: Google Scholar
 T. Rohlfing and C. R. Maurer, “Nonrigid image registration in sharedmemory multiprocessor environments with application to brains, breasts, and bees,” IEEE Transactions on Information Technology in Biomedicine, vol. 7, no. 1, pp. 16–25, 2003. View at: Publisher Site  Google Scholar
 M. Joshi, J. Cui, K. Doolittle et al., “Brain segmentation and the generation of cortical surfaces,” NeuroImage, vol. 9, no. 5, pp. 461–476, 1999. View at: Publisher Site  Google Scholar
 Y. Liu, A. Fedorov, R. Kikinis, and N. Chrisochoides, “Realtime nonrigid registration of medical images on a cooperative parallel architechture,” in Proceedings of the IEEE International Conference on Bioinformatics and Biomedicine, pp. 401–404, Washington, DC, USA, November 2009. View at: Google Scholar
 F. Ino, K. Ooyama, and K. Hagihara, “A data distributed parallel algorithm for nonrigid image registration,” Parallel Computing, vol. 31, no. 1, pp. 19–43, 2005. View at: Publisher Site  Google Scholar
 G. E. Christensen, M. I. Miller, M. W. Vannier, and U. Grenander, “Individualizing neuroanatomical atlases using a massively parallel computer,” Computer, vol. 29, no. 1, pp. 32–38, 1996. View at: Google Scholar
 L. K. Ha, J. Krüuger, P. T. Fletcher, S. Joshi, and C. T. Silva, “Fast parallel unbiased diffeomorphic atlas construction on multigraphics processing units,” in Proceedings of the Eurographics Symposium on Parallel Graphics and Visualization, Munich, Germany, March 2009. View at: Google Scholar
 L. Risser, F. Vialard, R. Wolz, M. Murgasova, D. Holm, and D. Rueckert, “Simultaneous multiscale registration using large deformation diffeomorphic metric mapping,” IEEE Transactions on Medical Imaging. In press. View at: Publisher Site  Google Scholar
Copyright
Copyright © 2011 Daniel J. Tward et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.