About this Journal Submit a Manuscript Table of Contents
Computational and Mathematical Methods in Medicine
Volume 2013 (2013), Article ID 834192, 7 pages
http://dx.doi.org/10.1155/2013/834192
Research Article

Nonrigid Registration of Lung CT Images Based on Tissue Features

1Key Laboratory of Health Informatics, Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, 1068 Xueyuan Avenue, Shenzhen 518055, China
2Harbin Institute of Technology Shenzhen Graduate School, Shenzhen 518055, China

Received 29 June 2013; Revised 3 September 2013; Accepted 2 October 2013

Academic Editor: Tianye Niu

Copyright © 2013 Rui Zhang et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

Nonrigid image registration is a prerequisite for various medical image process and analysis applications. Much effort has been devoted to thoracic image registration due to breathing motion. Recently, scale-invariant feature transform (SIFT) has been used in medical image registration and obtained promising results. However, SIFT is apt to detect blob features. Blobs key points are generally detected in smooth areas which may contain few diagnostic points. In general, diagnostic points used in medical image are often vessel crossing points, vascular endpoints, and tissue boundary points, which provide abundant information about vessels and can reflect the motion of lungs accurately. These points generally have high gradients as opposed to blob key points and can be detected by Harris. In this work, we proposed a hybrid feature detection method which can detect tissue features of lungs effectively based on Harris and SIFT. In addition, a novel method which can remove mismatched landmarks is also proposed. A series of thoracic CT images are tested by using the proposed algorithm, and the quantitative and qualitative evaluations show that our method is statistically significantly better than conventional SIFT method especially in the case of large deformation of lungs during respiration.

1. Introduction

Lung cancer is the most common cause of cancer-related death all over the world, with exceeding 1 million deaths annually [1]. Image-guided radiation therapy (IGRT) plays an important role in both the curative and palliative treatment of lung cancer, and precise targeting of lung tumors is an essential step in IGRT [2]. However, it is difficult to target tumors in lungs by taking into account respiration and tumor motion. Deformable or nonrigid image registration has been recognized as a key technology to locate position of tumor precisely [3]. By definition, image registration is a process of establishing spatial correspondences between two images. It can be classified into rigid registration and nonrigid registration. As the motion and shape change of lung is nonlinear, it is appropriate to register the lung CT images by using nonrigid registration.

Various nonrigid image registration methods have been applied for the lung CT images. In general, the registration methods can be divided into intensity-based and feature-based methods. There are a lot of intensity-based methods used in the thoracic CT image registration, such as fast intensity-based freeform registration [4], multiresolution optical flow technique [5], regional narrow shell model [6], and demons algorithm [7]. However, intensity-based methods tend to misregister small structures in the lung like vessels and airways as they only rely on image intensity [8]. Feature-based methods are typically applied when the local structural information is distinctive and salient, and the performance of feature-based registration is closely related to the accuracy of feature extraction and matching in images.

In feature-based registration methods which were commonly used in thoracic images, both Rohr [9] and Coselmon et al. [10] used manually extracted landmarks and an approximating thin-plate spline to describe the mapping function. However, manual landmark extraction often takes much time and is tedious, and the accuracy greatly depends on the experience of doctors. Han [2] proposed a hybrid method that used features detected by Forstner operator as constrains to guide an intensity-based deformable registration. Gorbunova et al. [11] combined intensity, curves, and surfaces to register lung CT images. These methods were hybrid techniques that combined feature-based and intensity-based methods. Recently, Xie et al. [12] and Urschler et al. [8] showed the feature-based registration based on scale-invariant feature transform (SIFT) [13] and thin-plate spline (TPS) achieved promising results in thoracic CT images. The robustness of SIFT has been properly evaluated, which shows well performance and stability under arbitrary affine transformations for MR, CT, and ultrasound images due to its distinctive and superior advantages in feature detection and description [14, 15]. However, the accuracy of extracted key points and the matching strategy have not been fully explored. Our work mainly focuses on these two questions.

In the present work, we develop a hybrid feature extraction method that can detect anatomic tissue features of lungs effectively based on Harris and Stephens [16] and SIFT. To effectively remove the mismatched features, a novel method based on cross-correlation and structural invariance is also proposed.

2. Methods

The proposed method consists of four major steps: tissue feature detection, feature description, feature matching and mismatched points removing, and thin-plate spline (TPS) transformation [17]. We will illustrate the method in the following sections in detail. As the voxel spacings in three directions are different, we first preprocess the images into a volume with isotropic spacing. This process can improve computation speed and reduce memory usage due to the increase of voxel spacing of each slice. In addition, having an isotropic voxel spacing can increase the accuracy of SIFT descriptor.

2.1. Tissue Feature Detection

The SIFT algorithm detects key points of extremes in the scale space, and the detected key points are blob features which cannot fully reflect the movement of lungs. The features that we need to find are lung boundaries, vessel bifurcations, and alveoli. These features usually have high intensity gradient; thus we use Harris algorithm which can find points with high intensity gradient.

The Harris detector is based on the local autocorrelation function of a signal, which measures the local changes of the signal with patches shifted by a small amount in different directions [18]. In 3D images, the local autocorrelation is defined as where denotes the image function and are the points in the Gaussian window centered on .

The shifted image can be approximated by a Taylor expansion truncated to the first order terms: where  , , and denote the intensity gradient in -, -, and -axises.

The eigenvalues of the matrix contain enough local information related to the neighborhood structure, and points with three high eigenvalues are selected as the features. To reduce computational complexity and get a good distribution of feature points in the lungs, the strategy we used is to select the key points which have three high eigenvalues and each key point has a distance larger than a threshold with others.

2.2. Local SIFT Feature Descriptor

The SIFT descriptor is robust to local deformations and to errors in feature detection. It is considered as one of the most effective descriptors currently available [19]. For each detected feature, a distinctive local SIFT feature descriptor is built [20, 21]. The 3D SIFT descriptor is characterized by using the gradient orientation distribution in a grid surrounding the feature position, and the cube is divided into subregions. The descriptor structure is shown in Figure 1.

834192.fig.001
Figure 1: The 3D SIFT descriptor covers a voxel region, the whole region divided into subregions. For each subregion, bins histogram is calculated to summarize the gradient orientation of voxels in the subregion.

Each voxel has two values which represent the direction of the gradient in three dimensions. One is from 0° to 360°, and the other is from −90° to 90°. To build the orientation histograms, each bin indicates 45°. Thus each subregion has an bins histogram to summarize the gradients of the voxel in it. Therefore, a total of 2048 vectors are calculated for a given feature.

2.3. Feature Matching and Removing Mismatched Points

In the process of feature matching, the best candidate match for each key point is found by identifying its nearest neighbor in the dataset of key points from images. The nearest neighbors are defined as the key points with minimum Euclidean distance from the given descriptor vector. The probability that a match is correct can be determined by taking the ratio of distance from the closest neighbor to the distance of the second closest [13]. To find corresponding feature points, we first convert the histograms of a feature to a single vector and compare the distance between a feature vector in template image against every feature vector of the target image. Suppose and are the best match (which has the lowest distance) and the second best match of feature , and , are the corresponding distance between features, with . If the ratio is below a threshold , then is chosen tentatively as the corresponding feature of . The threshold Lowe [13] rejected all matches in which the distance ratio is greater than 0.8, which eliminates 90% of the false matches while discarding less than 5% of the correct matches. In our tests, we set to be 0.8 experientially. To further improve the accuracy of above vector matching, we perform the matching process twice by reversing the roles of the two volumes and consider only those matches as valid for which the features are still matched from volume to and from to . However, there are still many wrong matched pairs after the above process.

To remove these mismatched points, we proposed a novel method based on cross-correlation and structural invariance for the matching verification. Cross-correlation is a similarity metric between two signals. We can remove those mismatches which have slightly large differences in the neighborhood areas around candidate corresponding key points by cross-correlation. The correlation coefficient of two corresponding features is calculated according to where are the coordinates of voxels within a distance surrounding the features position, and are the intensity values of corresponding voxel, and and are the mean intensity values. These intensity values are weighted by a Gaussian window and normalized. If the correlation coefficient CC is larger than a threshold, we deem that they are similar to each other. We experientially set d to be 5 voxels.

Even by using cross-correlation, there are still some mismatched points. As we know, the relative position of structures in lungs would not change too much during respiration. Therefore, if there are a number of key points surrounding a feature in the lung during inspiration, these key points would remain surrounding this feature during expiration. The approach based on structural invariance consists of two steps. First, a number of points are selected which have the shortest distance to the tested feature in an image. Second, the same number of points are selected which are nearest to the corresponding feature in the other image. If the number of the matched point pairs is larger than a given threshold, we affirm that the tested features are corresponding. This process is illustrated in Algorithm 1.

alg1
Algorithm 1: Structural invariance removing method.

Figure 2 shows the result before and after using the above method. These control point pairs are detected using Harris detector and intensity thresholds are used to insure the main points detected are in the region of lungs; we can see that the above method can remove mismatched point pairs effectively. Table 1 shows the number of key point pairs in the process of matching verification using cross correlation and structural invariance.

tab1
Table 1: The number of key point pairs in the process of matching verification.
fig2
Figure 2: (a) The matched pairs only using symmetrical nearest neighborhood method. (b) The retained pairs after using our method. (c) The removed point pairs using our method. The red points are points detected in one thoracic volume, and the blue points are in the other. The black lines link matched pairs.
2.4. Thin-Plate Spline Transformation

The process of nonrigid registration warps the template image to the target image in a way that they can best match on a voxel-by-voxel basis. Mathematically, this is an optimization problem, in which a set of transformation parameters transform the voxels in the template image to their corresponding voxels in the target image [22].

To find the transformation matrix that maps an arbitrary voxel on the template image to that on the target image, TPS transformation is employed in this study. This method can be used to establish voxel-to-voxel correspondence of a region of interest according to the paired control points. A detailed description of the TPS method can be found in the studies by Bookstein [17].

3. Results and Discussion

The performance of the above method was evaluated by a series of thoracic CT images. Each image had 80 slices, and was reconstructed with a 2.5 mm slice thickness. Each CT slice was discretized into voxels, and the voxel spacing is 0.977 mm. All the images are handled on a personal computer with Pentium 2.8 GHz Dual-Core and 3 GB memory, and the proposed methods are implemented using C++. Insight Segmentation and Registration Toolkit (ITk) [23] is also used. Some programs are referred to Xie et al. [12] and Cheung and Hamarneh [21]. The mean of the square sum of intensity differences (SSD) and image intensity cross-correlation coefficient (CC) is used to evaluate the quality of the registration images quantitatively.

Figure 3(a) shows the fusion image of the template and target phases before registration. Figure 3(b) shows the fusion image of the two phases using the proposed approach. For comparison, the results of the conventional SIFT approach are also shown in Figure 3(c). The red region stands for the target image, and the green region stands for the template image.

fig3
Figure 3: Fusion images of two phases of digital phantom: (a) before registration, (b) after registration using the proposed method, and (c) after registration using the conventional SIFT method.

To be quantitative, we listed the results of these two methods in Table 2. It is clear that by using the proposed method, SSD was reduced from 56727 to 37650, and the CC was increased from 0.646 to 0.740, demonstrating that the proposed method is much better than the conventional SIFT.

tab2
Table 2: Comparison of the proposed method and the conventional SIFT.

We also choose 15 points to reflect the average errors of the above registration methods. The results are listed in Table 3. The errors are computed by point coordinates in target images minus the point coordinates in register images. On average, using our hybrid method, the mean absolute deviation in three directions of the 15 points was reduced from 1.6 to 0.6 mm, from 1.9 to 0.7 mm, and from 2.8 to 0.7 mm, respectively. The standard deviation (SD) in three directions of the 15 points was reduced from 2.6 to 0.9 mm, from 2.2 to 1.0 mm, and from 5.2 to 1.2 mm, respectively. From the data in Table 3, it is clear that the proposed method is more accurate than conventional SIFT.

tab3
Table 3: Errors of 15 representative points using the proposed and conventional SIFT.

Generally, the registration results by using TPS method depend on the number and locations of the control points. In our work, as we only use an intensity threshold rather than a true segmentation to preprocess the volume, the detected points by using SIFT were spread in the phantom, only a small part of them was located on the lungs. However, the detected points by using proposed method were mostly on the lungs. Both of these two methods got about 500 matched point pairs (SIFT: 539 pairs, proposed method: 490 pairs). Though the matched point pairs by using the proposed method were fewer than those by using SIFT, the detected points on the lungs by using the proposed method were statistically significantly more than those by using SIFT. Figure 4 shows results of detected feature points using these two methods.

fig4
Figure 4: Feature point detection: (a) SIFT, (b) proposed method. The white dots stand for the detected feature points.

4. Conclusions

In the present work, a hybrid feature-based nonrigid image registration method is proposed. It can effectively detect the tissue features. Promising results were obtained using clinical thoracic data. Furthermore, our proposed method to remove mismatched points is automatic and robust. It can guarantee the correctness of point pairs and can be a significant supplementary of former feature-based registration methods.

Acknowledgments

This work was supported by the Grant from China Postdoctoral Science Foundation (2013M530740), in part by grants from National Natural Science Foundation of China (NSFC: 81171402), NSFC Joint Research Fund for Overseas Research Chinese, Hong Kong, and Macao Young Scholars (30928030), National Basic Research Program 973 (2010CB732606) from Ministry of Science and Technology of China, and Guangdong Innovative Research Team Program (no. 2011S013) of China.

References

  1. J. Ferlay, H.-R. Shin, F. Bray, D. Forman, C. Mathers, and D. M. Parkin, “Estimates of worldwide burden of cancer in 2008: GLOBOCAN 2008,” International Journal of Cancer, vol. 127, no. 12, pp. 2893–2917, 2010. View at Publisher · View at Google Scholar · View at Scopus
  2. X. Han, “Feature-constrained nonlinear registration of lung CT images,” in Medical Image Analysis Clinic: A Grand Challenge, pp. 137–146, 2010.
  3. L. Xing, B. Thorndyke, E. Schreibmann et al., “Overview of image-guided radiation therapy,” Medical Dosimetry, vol. 31, no. 2, pp. 91–112, 2006. View at Publisher · View at Google Scholar · View at Scopus
  4. W. Lu, G. H. Olivera, Q. Chen, M.-L. Chen, and K. J. Ruchala, “Automatic re-contouring in 4D radiotherapy,” Physics in Medicine and Biology, vol. 51, no. 5, pp. 1077–1099, 2006. View at Publisher · View at Google Scholar · View at Scopus
  5. G. Zhang, T.-C. Huang, T. Guerrero et al., “Use of three-dimensional (3D) optical flow method in mapping of 3D anatomical structure and tumor contours across four-dimensional computed tomography data,” Journal of Applied Clinical Medical Physics, vol. 9, no. 1, pp. 59–69, 2008. View at Scopus
  6. M. Chao, T. Li, E. Schreibmann, A. Koong, and L. Xing, “Automated contour mapping with a regional deformable model,” International Journal of Radiation Oncology Biology Physics, vol. 70, no. 2, pp. 599–608, 2008. View at Publisher · View at Google Scholar · View at Scopus
  7. H. Wang, L. Dong, J. O'Daniel et al., “Validation of an accelerated “Demons” algorithm for deformable image registration in radiation therapy,” Physics in Medicine and Biology, vol. 50, no. 12, pp. 2887–2905, 2005. View at Publisher · View at Google Scholar · View at Scopus
  8. M. Urschler, J. Bauer, H. Ditt, and H. Bischof, “SIFT and shape context for feature-based nonlinear registration of thoracic CT images,” in Computer Vision Approaches to Medical Image Analysis, vol. 4241 of Lecture Notes in Computer Science, pp. 73–84, 2006.
  9. K. Rohr, Landmark-Based Image Analysis: Using Geometric and Intensity Models, vol. 21, Springer, New York, NY, USA, 2001.
  10. M. M. Coselmon, J. M. Balter, D. L. McShan, and M. L. Kessler, “Mutual information based CT registration of the lung at exhale and inhale breathing states using thin-plate splines,” Medical Physics, vol. 31, no. 11, pp. 2942–2948, 2004. View at Publisher · View at Google Scholar · View at Scopus
  11. V. Gorbunova, S. Durrleman, P. Lo, X. Pennec, and M. de Bruijne, “Lung CT registration combining intensity, curves and surfaces,” in Proceedings of the 7th IEEE International Symposium on Biomedical Imaging: From Nano to Macro (ISBI '10), pp. 340–343, April 2010. View at Publisher · View at Google Scholar · View at Scopus
  12. Y. Xie, M. Chao, and L. Xing, “Tissue feature-based and segmented deformable image registration for improved modeling of shear movement of lungs,” International Journal of Radiation Oncology Biology Physics, vol. 74, no. 4, pp. 1256–1265, 2009. View at Publisher · View at Google Scholar · View at Scopus
  13. D. G. Lowe, “Distinctive image features from scale-invariant keypoints,” International Journal of Computer Vision, vol. 60, no. 2, pp. 91–110, 2004. View at Publisher · View at Google Scholar · View at Scopus
  14. M. Moradi and P. Abolmaesumi, “Medical image registration based on distinctive image features from scale-invariant (SIFT) keypoints,” in Proceedings of the International Congress and Exhibition on Computer Assisted Radiology and Surgery (CARS '05), p. 1292, 2005.
  15. M. Moradi, P. Abolmaesoumi, and P. Mousavi, “Deformable registration using scale space keypoints,” in Medical Imaging 2006: Image Processing, vol. 6144 of Proceedings of SPIE, pp. 61442G1–61442G8, February 2006. View at Publisher · View at Google Scholar · View at Scopus
  16. C. Harris and M. Stephens, “A combined corner and edge detection,” in Proceedings of the 4th Alvey Vision Conference, pp. 147–151, 1988.
  17. F. L. Bookstein, “Principal warps: thin-plate splines and the decomposition of deformations,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 11, no. 6, pp. 567–585, 1989. View at Publisher · View at Google Scholar · View at Scopus
  18. I. Sipiran and B. Bustos, “Harris 3D: a robust extension of the Harris operator for interest point detection on 3D meshes,” International Journal of Visual Computer, vol. 27, no. 11, pp. 963–976, 2011. View at Publisher · View at Google Scholar · View at Scopus
  19. K. Mikolajczyk and C. Schmid, “A performance evaluation of local descriptors,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 27, no. 10, pp. 1615–1630, 2005. View at Publisher · View at Google Scholar · View at Scopus
  20. P. Scovanner, S. Ali, and M. Shah, “A 3-dimensional sift descriptor and its application to action recognition,” in Proceedings of the 15th International Conference on Multimedia (ACM '07), pp. 357–360, New York, NY, USA, September 2007. View at Publisher · View at Google Scholar · View at Scopus
  21. W. Cheung and G. Hamarneh, “n-SIFT: n-sdimensional scale invariant feature transform,” IEEE Transactions on Image Processing, vol. 18, no. 9, pp. 2012–2021, 2009. View at Publisher · View at Google Scholar · View at Scopus
  22. Y. Xie, M. Chao, and G. Xiong, “Deformable image registration of liver with consideration of lung sliding motion,” Medical Physics, vol. 38, no. 10, pp. 5351–5361, 2011. View at Publisher · View at Google Scholar · View at Scopus
  23. L. Ibanez, W. Schroeder, and L. Ng, The ITK Software Guide, Kitware, 2005.