- About this Journal ·
- Abstracting and Indexing ·
- Aims and Scope ·
- Annual Issues ·
- Article Processing Charges ·
- Author Guidelines ·
- Bibliographic Information ·
- Citations to this Journal ·
- Contact Information ·
- Editorial Board ·
- Editorial Workflow ·
- Free eTOC Alerts ·
- Publication Ethics ·
- Recently Accepted Articles ·
- Reviewers Acknowledgment ·
- Submit a Manuscript ·
- Subscription Information ·
- Table of Contents
Computational and Mathematical Methods in Medicine
Volume 2012 (2012), Article ID 389463, 6 pages
Modeling and Representation of Human Hearts for Volumetric Measurement
1College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou 310023, China
2Guangxi Academy of Sciences, 98 Daling Road, Nanning 530007, China
3DreamSciTech Consulting, Shenzhen 518054, China
Received 24 July 2011; Accepted 28 August 2011
Academic Editor: Carlo Cattani
Copyright © 2012 Qiu Guan et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
This paper investigates automatic construction of a three-dimensional heart model from a set of medical images, represents it in a deformable shape, and uses it to perform volumetric measurements. This not only significantly improves its reliability and accuracy but also makes it possible to derive valuable novel information, like various assessment and dynamic volumetric measurements. The method is based on a flexible model trained from hundreds of patient image sets by a genetic algorithm, which takes advantage of complete segmentation of the heart shape to form a geometrical heart model. For an image set of a new patient, an interpretation scheme is used to obtain its shape and evaluate some important parameters. Apart from automatic evaluation of traditional heart functions, some new information of cardiovascular diseases may be recognized from the volumetric analysis.
The research to diagnose and prevent cardiovascular diseases becomes more important than ever. Thanks to the newly developed technologies in medical imaging and computing, automatic evaluation of patient hearts now becomes possible. This is very useful for diagnosis and treatment. If all kind of cardiovascular diseases were cured, human life could be much longer . Modeling and volumetric measurement have the capability to improve the diagnostic value of cardiac images. Ventricular volume and representation make physician to evaluate the consequences of myocardial infarction according to a glance at the anatomy of the ventricles. Besides, Parameters rely on left ventricular volume such as the ejection fraction, which is a kind of measurement of ventricular ability to pump oxygenated blood through the body. Since cardiologists rely on these values as indicators of cardiac malfunction, it is useful for us to calculate the ventricular volumes [2–6].
Researchers have attempted much in this issue, such as medical image acquisition, image processing, feature enhancement and extraction, cardiac boundary segmentation, parameter computation, functional formulation, disease assessment, and so forth. A number of methods have also been attempted for carrying out these works. The contributions can be extensively explored in the literature [7–12]. Typically, Yamamuro et al. carried out a project on two-dimensional (2D) image processing . They evaluate accuracy of cardiac functional analysis with multidetector-row computed tomography (CT) and segmental reconstruction algorithm over a range of heart rates. Various functional parameters of the left ventricle are measured, and they are correlated and agreed with those obtained with magnetic resonance imaging (MRI). For the single-photon emission computed tomography, Germano et al. have developed an algorithm to quantitatively measure left ventricular ejection fraction from gated 99mTc-sestamibi myocardial perfusion images . The algorithm operates in the three-dimensional (3D) space and uses gated short-axis image volumes. It segments the ventricle, estimates and displays endocardial and epicardial surfaces for all gating intervals in the cardiac cycle, calculates the relative left ventricular cavity volumes, and derives the global ejection fraction from the end-diastolic and end-systolic volume. Results show that the automatic segmentation and contouring of the ventricle were very successful of the experimental studies. Relevant programs were also developed to provide clinically useful additional information to complement myocardial perfusion studies in hospitals. However, the limitation of these existing works lies in the lack of cardiac shape completeness, correctness or regularity, and fitting reliability.
This paper is concerned with modeling and analysis of 3D cardiac model (mainly the ventricle) of human heart and fitting the model to other patients’ 3D cardiac image volumes to form the patient’s subject-specified model. The model building includes semiautomated registration of shapes for all patients in the database, alignment, and decomposition into a flexible model. For a new case, the model is used to segment and fit 3D cardiac images. Finally quantitative functional analysis of the left ventricle is followed, which plays a very important role in the automatic/aided diagnosis of cardiac diseases. The method utilized in this research is a model-based approach. It includes two meanings. Firstly, the heart model is built by statistical analysis of many existing images from the database which are obtained from several hospitals, which creates a general heart model and possible variation. Secondly, the cardiac shape of a new patient is evaluated from the deviation of its shape from normal cases.
The remainder of this paper is organized as follows. In Sections 2 and 3, we introduce the method for model creation and training. Sections 4 and 5 describe issues of model representation and interpretation. Mathematic method of volumetric measurement is given in Section 6. Example experimental results are given in Section 7. Finally, conclusions are drawn in Section 8.
2. Model Creation
The key to quantitative analysis of cardiac functions is to recognize the dynamical and deformable cardiac shape from medical images. In recent decades, researchers have contributed many ideas for segmenting, fitting, and modeling of deformable shapes [13, 14]. The most frequently used ones are Deformable Models, Classification, Global Search, Snakes, Level-set, Eigen-faces and Eigen-patches, and so forth. Some of them are based on minimization of an energy function (e.g., Snakes). However, the energy function method searches the object boundary stopping on most rapid changes of gray-level image profiles. Due to the complexity of deformable object in human body where many other organs are located near it, the gray-level gradients sometimes cannot be used to describe the organ anatomical edges actually. Another typical method, the eigen-patches, is also used to model regions where the shape is assumed to be fixed. However, the problem comes when the medical organ is usually not fixed, for example, when the heart is beating. These require a method for not only modeling the object shape but also its shape variation. From our investigation, a method based on a flexible model is adopted to provide principled means to efficiently parameterize a cardiac shape and its variability. The model is initiated from the active shape modeling method described by Cootes et al. [15–17]. It not only satisfies the requirements of cardiac shape modeling and analysis, but also allows dimensionality reduction of the model for reducing the implementation cost.
In the first stage, a statistical model has to be created with a given set of cardiac examples for representing the general shape and its variation. For digital computation, each shape in the training set is represented by a set of labeled feature points, which must be consistent from one shape to another. For instance, a number of 3D points on a ventricle shape should always correspond to the same locations in the biological or anatomical sense. Figure 1 illustrates a segmentation program for training a general model of the ventricular shape. The segmentation results are also used as the input of the standard for evaluating automatic shape fitting by the proposed model-based approach.
To build a model of the general shape, it requires labeled training images to represent correspondences among the shape examples. Automatic landmarking methods for this purpose have been thoroughly studied by researchers in recently years. For example, Izard et al. presented a method for landmarking MR images in registering brain structures from different images using a generic algorithm . With a set of such labeled training examples, we need to align them into a common coordinate frame. The Generalized Procrustes Analysis can be used to align training shapes and minimize the sum of squared distances to the mean of the set. In fact, it is to find the transformations which minimize where and is a specific shape example in the training set which is represented by a 3 element vector (for points of landmarks in 3D space), that is, The aligned training set forms a cloud in the 3D space and can be considered as a sample from a probability density function. To reduce the computation cost and memory requirement, we use principal component analysis (PCA) to pick out the main axes of the cloud, and model only the first few, which account for the majority of the variation. For the ventricular model, the first 50 principal eigen vectors are good enough to represent the shape and variation. The general model is then represented as where is the mean of the aligned shapes, is a matrix whose columns are unit vectors along the principal axes, and is a element vector of shape parameters. This means the shape dimension is reduced from to by PCA analysis.
This creates a statistical model like the point distribution model, and such a model is used in the flexible model framework to locate new examples in new images. By varying the shape parameters in within limits learnt from the training set, we can generate new plausible shapes. Usually the variance of is (the eigen value of the th largest in the matrix ).
3. Model Training by Genetic Algorithm
Genetic algorithm (GA) is introduced as a computational analogy of adaptive systems. It is modeled loosely on the principles of the evolution via natural selection, employing a population of individuals that undergo selection in the presence of variation-inducing operators such as mutation and recombination (crossover). A fitness function is used to evaluate individuals, and reproductive success varies with fitness, and therefore GA is a better way in global search, and passed few years witness its widely applications. There are definitely some laws and sequence of species existing in the nature. Taking medical images, for example, when the artificial factors are excluded, the scale and rotation variety of required images obey the normal distribution. Besides, the information we get from the normal distribution can efficiently help us establish a more reasonable model.
Generally, the statistical model uses the first shape of the training set or an arbitrary shape to be the first meanshape. However, such meanshape is not a “mean” finally. Using GA to form the first meanshape is an alternative choice. We can then get the information of the normal distribution and make the meanshape so as to improve the searching efficiency. Below are some strategies that can be adopted in the fitness function [14, 19].
Strategy 1 (obtain the model parameters by GA)
As the amount of GA arithmetic parameters should be controlled when considering the factor of time. The first shape can be used as initial meanshape to form the model, and the model parameters are generated using the GA algorithm. The sum of Euclidean distances between each shape and the meanshape can be taken as the fitness function
Strategy 2 (generate the shape directly from GA)
Directly using the coordinates of the shape points together as the input parameter of GA, the sum of Euclidean distances between each shape and the meanshape can be used in the fitness function. It is an efficient method for the shape without many points. Its computation time will be increased rapidly with the more points
Strategy 3 (generate each point one by one)
Searching the optimum point using GA arithmetic with each point as the input parameter, the fitness of GA is the sum of the distance between the points generated by the GA and the other corresponding points in the training set. Suitable weights are added to reflect the significance of some shapes in the training set. All the points are fitted together as the first mean shape to participate in the align procedure GA searches for the optimum solution, and the running time expands more or less when using GA for model generation. Fortunately, this problem is not very serious in the modeling process.
4. Model Representation
In this paper, a B-Spline surface model is used to represent the shape and get its volume [20, 21]. By applying matrix form of B-Spline, we obtain a polynomial B-Spline representation with two unitary parameters. Polynomials of B-Spline surfaces make the integral possible. This approach provides an actual volume of B-Spline surface, and it is also convenient and quick. Actually, a part of the B-Spline surface can be represented by the following matrix form: where , , and are basis matrixes in the nonempty intervals for the B-Spline surface. contains the control points: Equation (8) can also be rewritten as (10) or split into three scalar equations (11)
After we get a statistical model trained from the sets of examples, it is ready to interpret new images. The heart is located in the chest between the lungs behind the sternum and above the diaphragm. It is surrounded by the pericardium. It has the great vessels: the superior and inferior vena cava, the pulmonary artery and vein, and the aorta. The aortic arch lies behind the heart. The esophagus and the spine lie further behind the heart. This knowledge can help the computer to roughly put the cardiac model into a new image volume keeping not too far from the true position for accelerating the interpreting process using the created model.
An iterative method is used for matching the model to images. It iteratively deforms to fit to image volume of the ventricle. The shapes are constrained by statistical derivations to vary only in ways seen in a training set of labeled examples. In addition to the shape model, we require models of the image appearance around each model point. It can be built to represent the statistical variation of the gradient along profiles through the points, normal to the boundary curve at that point. The true boundary position can be found by computing the distance for the statistical profile moving along the image profile [22–25]. Finally the ventricular shape of the patient is fitted by repeating the following two steps until convergence: (1) look along normals through each model point to find the best local match for the model of the image appearance at that point (with minimum distance); (2) update the pose and shape parameters to best fit the model instance to the found points. This is to say that the goal of model fitting is to search best candidate image points near the model and update global transformation, , and parameters, , to minimize where is the temporary model obtained in the immediate steps. This minimization can be achieved by some nonlinear optimizers with iterative approaches. Finally the pose parameters in are fixed, and we get the corresponding shape parameters of the patient.
6. Volumetric Measurement
There are mainly three traditional approaches to calculate the ventricular volume. First, the ventricular volume is represented as the volume of a simple shape (e.g., truncated ellipse) or a combination of different figures. This method is simple to perform while the volume is coarse. Secondly, the ventricular volume is regarded as the sum of multiple smaller volumes of similar configuration. Thirdly, the ventricular volume is represented as the volume of B-Spline surface, and the volume is usually computed by using numerical integral such as Simpson’s rule and Gauss’s rule.
To calculate the volume of B-Spline surfaces, the polynomial expressions of B-Spline surface are given . Let and denote the derivatives of with respect to and , respectively. The volume can be determined by With a order B-Spline surface, it can be finally written as where is a matrix which is the product of the three polynomials:
Medical images give functional information about the heart while having less information on its anatomy. As it is well known that one of the main diagnostic parameters of interest for physician is its volume. In this paper, some experiments are carried out to construct the left ventricular surfaces fitted by B-spline model (Figure 2). From one cardiac cycle, the changes in volume can be obtained while the heart beats. The ventricular volumes are determined by the algorithms. Table 1 shows some of the volumes sampled from a cardiac cycle.
This paper presented a model-based approach for volumetric analysis of human hearts, especially for the ventricles, which is very important for diagnosis and treatment of cardiovascular diseases. The method is based on a flexible model combined with genetic strategies. Based on the flexible model trained from hundreds of patient images, a new patient will be actively analyzed to obtain its individual shape. We also adopted an efficient method for representation of 3D surfaces and provided a corresponding volumetric measurement algorithm of the B-Spline surface. The volumetric algorithm of a B-Spline model is important in working out other functional parameters of human hearts. The pipeline proposed in this paper takes advantage of complete segmentation of the heart shape. It can automatically construct a 3D model from a set of medical images. This not only significantly improves its reliability but also makes it possible to derive valuable information to doctors, such as dynamic volumetric measurements.
Thanks to Z. Xu, T. T. Jiang, J. Zhang, Y. Q. Xu, Y. H. Du, and C. Jiang for their helpful research and laboratory work in the project during their study in the university. This work was supported by the National Natural Science Foundation of China (NSFC nos. 61103140 and 61105073) and Zhejiang Provincial S&T Department (2010R10006, 2010C33095).
- S. Y. Chen, J. Zhang, H. Zhang et al., “Myocardial motion analysis for determination of tei-index of human heart,” Sensors, vol. 10, no. 12, pp. 11428–11439, 2010.
- D. Moroni, S. Colantonio, O. Salvetti, and M. Salvetti, “Heart deformation pattern analysis through shape modelling,” Pattern Recognition and Image Analysis, vol. 19, no. 2, pp. 262–270, 2009.
- X. Huang and D. N. Metaxas, “Metamorphs: deformable shape and appearance models,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 30, no. 8, pp. 1444–1459, 2008.
- R. I. Ionasec, I. Voigt, B. Georgescu et al., “Patient-specific modeling and quantification of the aortic and mitral valves from 4-D cardiac CT and TEE,” IEEE Transactions on Medical Imaging, vol. 29, no. 9, Article ID 5458068, pp. 1636–1651, 2010.
- S. Y. Chen and Q. Guan, “Parametric shape representation by a deformable NURBS model for cardiac functional measurements,” IEEE Transactions on Biomedical Engineering, vol. 58, no. 3, pp. 480–487, 2011.
- T. Arts, F. W. Prinzen, T. Delhaas, J. R. Milles, A. C. Rossi, and P. Clarysse, “Mapping displacement and deformation of the heart with local sine-wave modeling,” IEEE Transactions on Medical Imaging, vol. 29, no. 5, Article ID 5437350, pp. 1114–1123, 2010.
- M. Yamamuro, E. Tadamura, S. Kubo et al., “Cardiac functional analysis with multi-detector row CT and segmental reconstruction algorithm: comparison with echocardiography, SPECT, and MR imaging,” Radiology, vol. 234, no. 2, pp. 381–390, 2005.
- G. Germano, H. Kiat, P. B. Kavanagh et al., “Automatic quantification of ejection fraction from gated myocardial perfusion SPECT,” Journal of Nuclear Medicine, vol. 36, no. 11, pp. 2138–2147, 1995.
- C. Lorenz and J. V. Berg, “A comprehensive shape model of the heart,” Medical Image Analysis, vol. 10, no. 4, pp. 657–670, 2006.
- M. E. Barley, A. M. Galea, A. A. Armoundas, T. S. Rosbury, and G. B. Hirschman, “Validation of a novel catheter guiding method for the ablative therapy of ventricular tachycardia in a phantom model,” IEEE Transactions on Biomedical Engineering, vol. 56, no. 3, Article ID 4663616, pp. 907–910, 2009.
- S. Y. Chen, “Cardiac deformation mechanics from 4D images,” Electronics Letters, vol. 43, no. 11, pp. 609–611, 2007.
- G. Luo and P. A. Heng, “LV shape and motion: B-spline-based deformable model and sequential motion decomposition,” IEEE Transactions on Information Technology in Biomedicine, vol. 9, no. 3, pp. 430–446, 2005.
- O. Ecabert, J. Peters, H. Schramm et al., “Automatic model-based segmentation of the heart in CT images,” IEEE Transactions on Medical Imaging, vol. 27, no. 9, Article ID 4505365, pp. 1189–1202, 2008.
- H. Lu, S. Liu, W.-L. Wang, and S. Y. Chen, “Generation of a point distribution model using genetic algorithms,” in Proceedings of the 4th International Conference: Sciences of Electronic, Technologies of Information and Telecommunications, Tunisia, March 2007.
- T. F. Cootes and C. J. Taylor, “Statistical model of appearance for computer vision,” Tech. Rep., University of Manchester, Manchester, UK, 2004.
- A. Andreopoulos and J. K. Tsotsos, “Efficient and generalizable statistical models of shape and appearance for analysis of cardiac MRI,” Medical Image Analysis, vol. 12, no. 3, pp. 335–357, 2008.
- J. Koikkalainen, T. Tolli, K. Lauerma et al., “Methods of artificial enlargement of the training set for statistical shape models,” IEEE Transactions on Medical Imaging, vol. 27, no. 11, Article ID 4591396, pp. 1643–1654, 2008.
- C. Izard, B. Jedynak, and C. Stark, “Automatic landmarking of magnetic resonance brain images,” in Proceedings of the SPIE International Symposium on Medical Imaging, vol. 5747, pp. 1329–1340, San Diego, Calif, USA, 2005.
- C. Mclntosh and G. Hamarneh, “Genetic algorithm driven statistically deformed models for medical image segmentation,” in Proceedings of the Genetic and Evolutionary Computation Conference, (GECCO '06), Image Processing and Computer Vision, Seattle, Wash, USA, July 2006.
- T. T. Jiang, S. Y. Chen, and Y. Xu, “3-D representation and volumetric measurement of human heart from a cylindrical B-spline surface model,” in Proceedings of the International Conference on BioMedical Engineering and Informatics, pp. 765–769, Sanya, Hainan, China, May 2008.
- B. Zhang and J. F. M. Molenbroek, “Representation of human head with bi-cubic B-Spline technique based on the laser scanning technique in 3D surface anthropometry,” Applied Ergonomics, vol. 35, no. 5, pp. 459–465, 2004.
- S. Y. Chen and Y. F. Li, “Determination of stripe edge blurring for depth sensing,” IEEE Sensors Journal, vol. 11, no. 2, Article ID 5585653, pp. 389–390, 2011.
- S. Y. Chen, H. Tong, Z. Wang, S. Liu, M. Li, and B. Zhang, “Improved generalized belief propagation for vision processing,” Mathematical Problems in Engineering, vol. 2011, Article ID 416963, 12 pages, 2011.
- A. I. Veress, W. P. Segars, B. M. W. Tsui, and G. T. Gullberg, “Incorporation of a left ventricle finite element model defining infarction into the XCAT imaging phantom,” IEEE Transactions on Medical Imaging, vol. 30, no. 4, pp. 915–927, 2011.
- K. Punithakumar, I. B. Ayed, A. Islam, I. G. Ross, and S. Li, “Tracking endocardial motion via multiple model filtering,” IEEE Transactions on Biomedical Engineering, vol. 57, no. 8, Article ID 5471238, pp. 2001–2010, 2010.