Biomedical Signal and Image Processing for Clinical Decision Support Systems
View this Special IssueResearch Article  Open Access
Feng Li, Fatih Porikli, "Tracking Lung Tumors in Orthogonal XRays", Computational and Mathematical Methods in Medicine, vol. 2013, Article ID 650463, 7 pages, 2013. https://doi.org/10.1155/2013/650463
Tracking Lung Tumors in Orthogonal XRays
Abstract
This paper presents a computationally very efficient, robust, automatic tracking method that does not require any implanted fiducials for lowcontrast tumors. First, it generates a set of motion hypotheses and computes corresponding feature vectors in local windows within orthogonalaxis Xray images. Then, it fits a regression model that maps features to 3D tumor motions by minimizing geodesic distances on motion manifold. These hypotheses can be jointly generated in 3D to learn a single 3D regression model or in 2D through back projection to learn two 2D models separately. Tumor is tracked by applying regression to the consecutive image pairs while selecting optimal window size at every time. Evaluations are performed on orthogonal Xray videos of 10 patients. Comparative experimental results demonstrate superior accuracy (~1 pixel average error) and robustness to varying imaging artifacts and noise at the same time.
1. Introduction
Tumor tracking is an essential component of imageguided radiation therapy (IGRT) systems for treating lung and abdominal area tumors, which move significantly due to respiration. With accurate tracking, large margins added to the planning target volumes can be eliminated, and thus, adverse effects of radiation on healthy tissue can be minimized. Duration of treatment sessions can be shortened considerably by tracking tumors and applying radiation continuously.
Conventional tracking approaches determine tumor position using internal and/or external surrogates. Often, multiple passive (metallic) [1] and active (electromagnetic) [2] internal fiducials are implanted around tumor to continuously monitor its motion in Xray videos. However, surgical implantation of such invasive internal fiducials are harmful as they destroy healthy tissues, and patients have greater chance of developing pneumothorax during CTguided fiducial placement. Besides, internal fiducials slide and relocate during the course of multiple treatment sessions causing uncertainty in their reference positions.
It is also possible to apply correspondence models between external markers (chest and abdominal area pointers) and internal fiducials to indirectly estimate tumor position in optical images [3, 4]. Yet, the correlation between external markers and tumor position may be violated easily as a result of complex respiratory biomechanics [5]. Alternatively, parametric models of motion patterns are used to track tumors [6]. Such methods, however, require manual labeling of surrogate regions and tumor positions for a long period of time in order to train models. A multiple template matching method for Xray images is described in [7]. It should be noted that template matching may fail for lowcontrast tumor regions where the image quality is low.
Here, we present a tumor tracking method that does not require any invasive internal fiducials or external markers. Unlike the traditional methods, we consider the tumor tracking problem as a regression model fitting task between orthogonal Xray videos and underlying tumor motion as illustrated in Figure 1. Our intuition is that tumor motion in orthogonal planes can be approximated by affine motion (or a similar parametric motion) and image features can be linearly correlated with these tumor motion parameters. In other words, the feature vector we compute within the tumor region is supposed to be an indicator of the tumor motion. For this, we use an image feature that is sensitive to motion unlike the insensitive features such as intensity histograms that conventional motion estimation methods often require.
We learn an online regression model, which is a single matrix that maps image features to motion parameters using the initial pair of the orthogonal Xray images. We randomly generate a set of 3D motion hypotheses (affine motion matrices) around the initial tumor location. To efficiently cover the parameter space, we generate motion hypotheses conditioned on the respiratory biomechanics (e.g., depending on the position of the tumor, the motion can be constrained to be translational only).
These hypotheses map an initial support window volume tumor in both orthogonal Xray image pairs by simple 3D2D projection. In addition, we determine an optimal tracking window size in each orthogonal view. We then compute two image feature vectors (e.g., histogram of oriented gradients) in the corresponding image windows, and we concatenate these vectors.
To learn the regression model, we solve an overcomplete leastsquares fitting problem between the motion hypotheses and concatenated feature vectors using the geodesic distances. In [8, 9], a linear regression function is considered. Since affine motion matrices constitute a Riemmanian manifold, motion hypotheses distances should be computed through geodesics. The benefits of using Riemmanian manifold over previously used motion models in Euclidean space (e.g., just vectorizing the motion hypothesis matrix) can be found in [10] for nonmedical examples. We are inspired by [10], yet we significantly extend its basic idea to 3D tumor motion, multiple orthogonal videos, and joint feature computation while further refining it with an adaptive optimal window selection.
To estimate the current tumor position in a new Xray pair, we simply apply the learned regression model to the concatenated feature vector computed in the new Xray images within the previous tumor windows.
Note that we track the tumor position but not its boundary. Boundary tracking can be done by segmentation, registration, boundary fitting, Bsplines, and so forth. However, for a lowcontrast and invisible tumor, boundary tracking may not be possible since the tumor boundary is not distinct from its surrounding tissues. Our method can track the position of lowcontrast tumors and does not require registration.
The presented regression method is possibly one of the fastest tumor tracking algorithms (and the fastest that we tested) as it only requires a feature vector and a matrix multiplication without any expensive search operation (template matching, etc.), iterative updates (mean shift), optimization with smoothing or total variation constraints (opticalflow), or testing a large number of hypotheses during the tracking process (particle filtering) as other tracking techniques. It does not require any offline training or patientspecific adaptation. It does not require a tumor detector either.
Extensive experiments with challenging real patient datasets demonstrate that the proposed method is robust to different tumor locations and very accurate for lowcontrast tumors.
2. Regression Tracking
Regression tracking learns an online regression model to correlate the observed image features to underlying tumor motion . This can be done either learning a single model in 3D Euclidean space (3DR) or learning two separate regression models in 2D planar space then intersecting them in 3D space (2DR).
We formulate the model learning as a fitting problem, which finds the regression coefficients through minimizing the geodesic distance between the tumor motion matrices and its estimates as where determines the weight of Tikhonov regularization, that is, preference to solutions with smaller smooth norms. We solve (1) through ridge regression to determine . The solution is given as At the first image pair () of orthogonal Xray sequences, we learn using the initial location . Then, for any given image pair , we simply compute a feature vector and apply to find the tumor position: A flow diagram of our method is given in Figure 2. Next, we explain how to compute , , and and details of our 3DR/2DR formulations.
2.1. Training
For a single Xray sequence, the tumor motion between two consecutive frames can be modeled as a 3 × 3 affine matrix (represented by 6 independent motion parameters in 9 coefficients). We use affine motion (translation, rotation, scale change, and skew) to represent the incremental movement of the tumor. Considering the biomechanical tissue models and respiratory mechanics, such an affine model is an adequate model. [11] further states that the tumor motion can be modeled as a simpler periodic function, featuring more time spent at the exhale phrase. When there is no hysteresis with the patient’s respiratory system, the pathways for inspiration and expiration of one respiratory cycle are approximately the same and almost linear. Even with the presence of hysteresis, the maximum deviation between the two pathways is usually much smaller than the distance along the primary direction of the tumor motion. Thus, a simple translation only motion would be sufficient in most cases; however, we consider full affine motion in our analysis for the completeness of the discussion.
Let represent the transformation from a unit square in the object space to the affine region enclosing the target tumor in the image space; that is, . The tumor position at time can be computed efficiently from by where is the incremental motion.
Affine motion matrices lie on a Riemmanian manifold. This means that we cannot simply vectorize the motion matrices to compute an Euclidean distance between them. Geodesic distances should be measured, thus we apply Lie group exponential map to determine where is a feature vector corresponding to the unit square in object space warped from through , is a matrix of regression coefficients, and is the number of motion coefficients. For 2D (3D) affine transformation can be set to 6 (12) independent parameters, or all 9 (16) coefficients of the 3 × 3 (4 × 4) affine motion matrix. In either case, each row of should be reshuffled to obtain the corresponding motion matrix.
To find the optimal solution of (1), we first generate a training set of random affine transformation hypotheses together with their corresponding feature vectors extracted from image , as shown in Figure 3. Then, we construct and as
Keep in mind that we generate more hypotheses than the feature vector size and significantly more than the number of motion parameters. This means that we obtain an overcomplete system of equations in above minimization. Random sampling serves to achieve an effective and reliable training set and avoid overfitting. One can also quantize the motion space and use those quantized motion parameter values in training. We observed that the value of has little impact on the performance as long as it is larger than the feature vector dimension; that is, . In our experiments, we set for the 2D regression algorithm.
Since the incremental motion of the tumor is small [11], we can also limit the sampling bounds using the maximum translational and rotational motion constraints typical for tumor motion. This not only improves the tracking accuracy but also stabilizes the tumor trajectories by removing possible jitters.
We use Histograms of Oriented Gradients (HOG) to describe the tumor window features . HOG has 8bin histograms for each 5 × 5 block within the tracking window, concatenated into a single column vector.
During the tracking process, the model can be relearned to adapt changes if necessary. In our simulations we have not observed any drift or model distortion issue with only one training at the initialization.
Note that the regression model built using the first images of the orthogonal sequences where the initial tumor position is given (by table alignment, etc). Then, it automatically tracks the tumor in the newly given images. Our learning method is blind to patient data; it does not make any patientspecific assumption or require any patientspecific information. It does not use any offline training and does not require manual marking of the tumor trajectory either.
For 512 × 512 image resolution, automatic generation of motion hypotheses, computing the corresponding concatenated feature vectors and solving ridge regression, takes 0.05 seconds.
2.2. Regression (3DR)
To learn a joint regression model that correlates the 3D tumor motion directly with the orthogonal Xray sequences, we group the feature vectors from the two views and estimate a 3D affine motion corresponding to them directly. We construct the combined feature vectors from the two Xray views and based on the labeled tumor position at time and randomly generate 3D affine matrices in 3D Euclidean space. Next, we project on the orthogonal Xray image planes. Within the corresponding regions, we extract the features and for the first and second Xray views. In this case, maps combined feature vectors to their corresponding 3D affine motion matrices , where is a 4 × 4 affine matrix. Thus, , .
2.3. 2D Regression (2DR)
Instead of learning a 3D regression model, we could simply learn two separate regression matrices and for the two Xray views, and we apply (4) and (5) iteratively to track the tumor in each Xray view. Once we have the two individual tracking results and of the two orthogonal Xray views at time , we can simply compute the tumor position in 3D through back projection. To do this, we first connect and to form lines , , and to form line and then compute as the intersection of and , while and are the source points of the two orthogonal Xray radiation. In practice, and may not necessarily intersect with each other; thus we choose the midpoint of the shortest path connecting these lines to represent the tumor positions.
2.4. Optimal Tracking Window
Tracking window size plays an important role in tumor tracking in soft tissues, especially in lung and abdomen areas. To yield valid tracking results for template based methods, the tracking window should not be too big (may underestimate motion) or too small (may lose track). We examine the selfsimilarity in the local tumor region to find the optimal tracking window size for each patient data.
Given a candidate window size , we define the local search region of size with the same center, and , where is the bounding size of the tumor. We compute the feature distance for any pair of image patches within the searching region using norm and use the mean error as an indicator to describe the discriminatory power of the candidate window size.
Intuitively, the larger this mean error the more discriminatory power the candidate size has. However, there may exist different distance distributions, which have the same mean error. To take into account these cases, we give preference to the small (20% of) feature distances and use their mean as the indicator. In our experiments, the ratio of the optimal windows size to the tumor bounding box varies from 0.9 to 1.2 for the coronal view and from 1.1 to 1.3 for the sagittal view for different Xray data sets.
3. Results
For objective performance evaluations, we test our tracking algorithm on digitally reconstructed radiograph (DRR) sequences obtained from realpatient 4DCT data. These DRR sequences have manually labeled groundtruth 3D landmark positions.
Using Xray videos has several issues. To annotate groundtruth motion, Xray videos should depict tissues with embedded metallic marker. However, it is problematic to make a tracking algorithm to ignore high contrast marker regions, which are often close to tumor and yet compute uncontaminated image features for an unbiased evaluation. Besides, markers themselves introduce uncertainty on the groundtruth data since they may dislocate from the initial calibrated positions or occlude each other in Xray videos.
We use orthogonal DRR sequences obtained from 10 patients’ 4DCT data [12]. This data has different tumor locations, shapes, and internal volume characteristics.
Using a stateoftheart simulator [13], we embed lowcontrast tumors in different shapes, sizes, and locations in the original 4DCT data and then generate DRR sequences representing different breathing patterns. Tumor shapes range from spheroids to very intricate 3D polytopes. Each test case is tested with a different regular breathing signal and two irregular breathing patterns. Since we do not impose any temporal smoothing or linear dynamical model (Kalman filter), the performance is not affected by the different breathing patterns. Each DRR sequence we test has around 900 frames. Figure 4 shows sample coronal and sagittal views.
Compared with higher resolution flatpanel Xray digitizer, this data presents considerable challenges. (i)Since the patient CT has limited number of slices, it is low resolution and drastically blurred particularly in the cranialcaudal direction of the coronal view; for example, ~100 × 256 pixels (underlying DRR are of 256 × 256) for Patients 1–5 and ~128 × 512 (DRR 512 × 512) for Patients 6–10. On the other hand, a typical of the flat panel Xray digitizer has 2048 × 2048 pixels. Limited resolution causes less discriminative features leading potential tracking failures.(ii)It has low contrast. The typical dynamic range of a commercial flatpanel digitizer is 16 bits. However, the DRRs we use are 8 bits encoded to push the algorithm to its limit.(iii)It contains noise and imaging artifacts. Unlike Xray videos from digitizer, DRR sequences inherit all 4DCT imaging artifacts due to limited CT scanning speed.(iv)We add white random Gaussian noise to the DRR sequences considering that DRR images might not suffer from Xray image acquisition noise. These issues certainly make tumor tracking more difficult in our dataset.
We compare the performance of our 3DR and 2DR methods with the stateoftheart including the best existing optical flow implementation (OF) [14] and based particle filter (P) [15]. [14] combines the “classical” flow formulation with image boundaries and designs an optimization framework that utilizes median filtering for flow field estimation. [15] is a templatebased robust visual tracking method, which enforces sparse representation on the template set and follows a particle filterbased Bayesian state inference.
We also implemented a HOG based particle filter (PHOG) algorithm, which uses the same number of 3D particles as the P method but computes the observation likelihood from HOG feature matching. The OF algorithm first finds the tumor motions on the two Xray views and then estimates the 3D tumor position through back projection. We compute the optical flow between two consecutive frames in full resolution since that the magnitude of the tumor motion is unknown and far away feature points can also contribute to the motion estimation of tumor regions.
Table 1 presents the detailed performance comparison of these algorithms as well as their average processing time per frame. For each test case, we also list the tumor motion magnitude in the last column. We use the Euclidean distance between the estimated and the ground truth (GT) tumor center in 3D as the error measurement (in pixels). The GT tumor center is calculated as the mass center of the GT tumor, while the estimated tumor center is the center of the estimated tracking window.

Our results using 10 different patient data show that 2DR gives 1.05 pixel, 3DR 1.16 pixel, OF 3.57 pixel, P 5.01, and PHOG 5.68 pixel error on average where the average tumor displacement in the GT is 13.86 pixels. This means that 2DR estimates are 92.5% accurate (1.05 is 7.5% of 13.86).
From Table 1, we can see that 3DR and 2DR consistently achieve most accurate tracking results. Other algorithms vary significantly for different test cases and may lose track of the tumor under certain scenarios (e.g., for Patient 10).
We do not claim that just because we achieve a 1.05 pixel average error on lowerresolution images we may obtain the same error at higher resolutions. Our experimental results on the 256 × 256 and 512 × 512 datasets clearly show that the estimation errors in terms of pixels remains the same (in fact becomes lower: from 1.63 pixel average error for 256 × 256 to 0.69 pixel average error for 512 × 512) when the image resolution increases.
In Table 1, 4DCT data for Patients 1–5 have 256 × 256 DRR image size, while Patients 6–10 have 512 × 512. Even though the underlying total displacement doubled from 8.49 pixel to 19.23 pixel on average (implying the tracking problem becomes more challenging), the tracking errors in terms of pixels are better for 512 × 512 sequences. In other words, we can confidently expect similar pixel errors when we use even higher resolution images (as a result, get lower mm error). Instead of DRR, if we used a commercial product (e.g., Siemens Axiom Luminos dRF flat detector), 1.05 pixel error would correspond to 0.14 mm to 0.42 mm error.
Here, we also like to mention that, for a fair assessment of tracking methods, it is essential to report pixel errors rather than mm errors that is commonly conveyed in medical literature due to bottomline clinical requirements. However, by measuring mm error, the same algorithm can produce different mm errors using different resolution input data. This does not mean that the algorithm gets any better or worse as nothing algorithmically changes.
In Figure 4, we show the tracking results of different algorithms on data set Patient 7, and our 3DR/2DR tracks the tumor very well for both inhale and exhale phases. In Figure 5, we draw the estimated tumor positions in the CranialCaudal (CC) direction along with the ground truth for Patient 7. It is apparent that the optical flow based tracker tends to underestimate the tumor motion causing significant errors when the tumor motion is large. The particle filter based algorithms (P and PHOG), on the other hand, exaggerates the tumor motions at the two extremities and produces jittery results. Another issue with the particle filter based trackers is that the lack of sufficient texture in Xray images sometimes causes the selection of wrong motion hypothesis as the mode of the approximated pdf after the importance sampling.
Computationally optical flow based tracker (OF) is the slowest one among all trackers taking about 2.8 minutes to process a single frame on an Intel 3.4 GHz CPU, which is prohibitive for realtime tasks. There are of course faster OF methods, but their accuracy is worse. Our method can track the tumor in real time: 3DR (2DR) in less than 0.03 (0.06) seconds. 3DR algorithm can be preferred over 2DR for two reasons. First, 3DR generates tracking results twice faster than 2DR. Second, 3DR can avoid the divergence between two orthogonal views by learning a joint 3D regression model and maintaining the tumor positions in 3D space as opposed to two 2D planes.
Note that we do not train and test on the same images. We train on the initial pair where the tumor position is known and then track on the rest of the sequence automatically. Since the training step takes less than 0.05 seconds, this algorithm can run under any realtime clinical setting.
4. Conclusion
We presented a noninvasive tumor tracking method and demonstrated that this tracker outperforms the stateoftheart both in accuracy (~1 pixel error) and speed (0.03 sec). This corresponds to 7.5% tumor positioning error with respect to maximum tumor dislocation. Such a small tumor location error significantly reduces the prescribed treatment volume margins from several centimeters to millimeter range, and thus, prevents radiating healthy tissue in IGRT systems [16]. As a future study, we will extend the regression model to incorporate biomechanical tissue constraints for very complex tumor shapes.
References
 H. Shirato, T. Harada, T. Harabayashi et al., “Feasibility of insertion/implantation of 2.0mmdiameter gold internal fiducial markers for precise setup and realtime tumor tracking in radiotherapy,” International Journal of Radiation Oncology Biology Physics, vol. 56, no. 1, pp. 240–247, 2003. View at: Publisher Site  Google Scholar
 T. R. Willoughby, P. A. Kupelian, J. Pouliot et al., “Target localization and realtime tracking using the Calypso 4D localization system in patients with localized prostate cancer,” International Journal of Radiation Oncology Biology Physics, vol. 65, no. 2, pp. 528–534, 2006. View at: Publisher Site  Google Scholar
 Y. Seppenwoolde, R. I. Berbeco, S. Nishioka, H. Shirato, and B. Heijmen, “Accuracy of tumor motion compensation algorithm from a robotic respiratory tracking system: a simulation study,” Medical Physics, vol. 34, no. 7, pp. 2774–2784, 2007. View at: Publisher Site  Google Scholar
 F. Ernst, V. Martens, S. Schlichting et al., “Correlating chest surface motion to motion of the liver using esvr  a porcine study,” in Proceedings of the 12th International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI '09), pp. 356–364, September 2009. View at: Google Scholar
 S. Ahn, B. Yi, Y. Suh et al., “A feasibility study on the prediction of tumour location in the lung from skin motion,” British Journal of Radiology, vol. 77, no. 919, pp. 588–596, 2004. View at: Publisher Site  Google Scholar
 T. Lin, L. I. Cervĩo, X. Tang, N. Vasconcelos, and S. B. Jiang, “Fluoroscopic tumor tracking for imageguided lung cancer radiotherapy,” Physics in Medicine and Biology, vol. 54, no. 4, pp. 981–992, 2009. View at: Publisher Site  Google Scholar
 Y. Cui, J. G. Dy, G. C. Sharp, B. Alexander, and S. B. Jiang, “Multiple templatebased fluoroscopic tracking of lung tumor mass without implanted fiducial markers,” Physics in Medicine and Biology, vol. 52, no. 20, pp. 6229–6242, 2007. View at: Publisher Site  Google Scholar
 J. Matas, K. Zimmermann, T. Svoboda, and A. Hilton, “Learning efficient linear predictors for motion estiamtion,” Technical Report, Research Reports of CMP, Czech Technical University in Prague, Prague, Czech Republic, 2006. View at: Google Scholar
 L. Ellis, N. Dowson, J. Matas, and R. Bowden, “Linear regression and adaptive appearance models for fast simultaneous modelling and tracking,” International Journal of Computer Vision, vol. 95, no. 2, pp. 154–179, 2011. View at: Publisher Site  Google Scholar
 O. Tuzel, F. Porikli, and P. Meer, “Learning on lie groups for invariant detection and tracking,” in Proceedings of the 26th IEEE Conference on Computer Vision and Pattern Recognition, pp. 1–8, June 2008. View at: Publisher Site  Google Scholar
 Y. Seppenwoolde, H. Shirato, K. Kitamura et al., “Precise and realtime measurement of 3D tumor motion in lung due to breathing and heartbeat, measured during radiotherapy,” International Journal of Radiation Oncology Biology Physics, vol. 53, no. 4, pp. 822–834, 2002. View at: Publisher Site  Google Scholar
 R. Castillo, E. Castillo, R. Guerra et al., “A framework for evaluation of deformable image registration spatial accuracy using large landmark point sets,” Physics in Medicine and Biology, vol. 54, no. 7, pp. 1849–1870, 2009. View at: Publisher Site  Google Scholar
 F. Li and F. Porikli, “Robust visual tracking using L1 minimization,” in Proceedings of the MICCAI Workshop on BioImaging and Visualization for PatientCustomized Simulations, Nagoya, Japan, September 2013. View at: Google Scholar
 D. Sun, S. Roth, and M. J. Black, “Secrets of optical flow estimation and their principles,” in Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 2432–2439, June 2010. View at: Publisher Site  Google Scholar
 X. Mei and H. Ling, “Robust visual tracking using L1 minimization,” in Proceedings of the IEEE 12th International Conference on Computer Vision, pp. 1436–1443, Kyoto, Japan, September 2009. View at: Google Scholar
 R. Luchtenborg, Realtime dose compensation methods for scanned ion beam therapy of moving tumors [Ph.D. thesis], CU Darmstadt, Darmstadt, Germany, 2012, Technical Report.
Copyright
Copyright © 2013 Feng Li and Fatih Porikli. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.