Research Article  Open Access
Deep Adaptive LogDemons: Diffeomorphic Image Registration with Very Large Deformations
Abstract
This paper proposes a new framework for capturing large and complex deformation in image registration. Traditionally, this challenging problem relies firstly on a preregistration, usually an affine matrix containing rotation, scale, and translation and afterwards on a nonrigid transformation. According to preregistration, the directly calculated affine matrix, which is obtained by limited pixel information, may misregistrate when large biases exist, thus misleading following registration subversively. To address this problem, for twodimensional (2D) images, the twolayer deep adaptive registration framework proposed in this paper firstly accurately classifies the rotation parameter through multilayer convolutional neural networks (CNNs) and then identifies scale and translation parameters separately. For threedimensional (3D) images, affine matrix is located through feature correspondences by a triplanar 2D CNNs. Then deformation removal is done iteratively through preregistration and demons registration. By comparison with the stateoftheart registration framework, our method gains more accurate registration results on both synthetic and real datasets. Besides, principal component analysis (PCA) is combined with correlation like Pearson and Spearman to form new similarity standards in 2D and 3D registration. Experiment results also show faster convergence speed.
1. Introduction
The aim of image registration is to establish spatial correspondences between two or more images of the same/or different scene acquired at different times, from different viewpoints, and/or by different sensors. Usually the ability to capture complex and large image deformations is vital to many computer vision applications including image registration and atlas construction. The problem becomes more challenging when the object in the image or edge of the image undergoes severe deformation [1].
Take medical image registration for example, tissues and organs or body itself are prone to deform, move, and rotate under most circumstances. Most methods iteratively reach a satisfying overlap under specific mathematical criterions, maximizing or minimizing deformation energy as described in 1. Fixed image is defined as , while moving image as . Registration aims to find the optimal model that best satisfies energy . As a result, model , objective function (similarity metric), and optimization method constitutes the three main components of image registration. Consider According to a stateoftheart survey [2], registration can be classified into rigid and nonrigid registration. Rigid models restrain the optimum to a few parameters to achieve global registration, while nonrigid models recover local deformation through physical model like elastic or viscous, or statistical model or support vector regression framework, and so forth. In order to fully overlap two images, researchers commonly adopt the twostep strategy, which contains initial registration and following iterative registration [3].
In the twostep strategy, registration firstly begins with a global affine transformation for initial global alignment, take stateoftheart method FLIRT [4] and ELASTIX [5, 6] for example. Or fiducial markers are firstly detected through feature descriptors, for example, the SIFT method [7], so the initial registration is carried out to establish correspondences between these point sets. In preregistration procedure, rotation, scale and translation of the moving image are modified by the calculated affine matrix. After that, nonrigid registration iteratively goes on. One severe problem of preregistration in affine matrix is that when large distortion and rotation both exists, accuracy is limited by correspondences between those regionbased descriptors. If descriptor itself is not accurate, problem becomes more severe. Once descriptors fail to discover point correspondences, accuracy of following registration would be badly influenced. As a result, imprecision may also be introduced misleading the following procedure. Besides, traditional FLIRT and ELASTIX method declares that images for registration must be with the best quality, otherwise poor registration may occur.
In order to address these above limitations and capture very complex and large deformations, we proposed a new approach for image registration based on a twolayer deep adaptive registration framework. Firstly, in the preregistration procedure, rotation, scale and translation extent between two images are obtained separately to achieve initial registration. This is quite different from traditional “one time calculated” affine matrix. For rotation parameter, a CNN classifier is trained offline in order to identify the level of current image rotation under sever distortion. Then scale and translation parameters are obtained. An optimum preregistration is calculated relating to above gained parameters. As for 3D images, a triplanar 2D CNNs [8] around each voxel is utilized for calculation of final affine matrix. Until now, preregistration is done. Secondly, the rectified images are further recovered through the following nonrigid demons registration procedure. In the next circle, the former registration further facilitates results of the later registration of last iteration. This iterative procedure is carried out until an optimum overlap between the two images is achieved. Besides, PCA is introduced to extract the most valuable features, and detected features are put into SSD, Kendall, Pearson and Spearman, and so forth to form new similarity metric. Also, a triplanar 2D PCA is proposed to process 3D registration problem and Figure 1 gives details of the algorithm. As a result, convergence speed is accelerated while maintaining the same registration accuracy. Figures 1, 2 and Algorithm 3 illustrate work flow of our framework of processing 2D and 3D image registration.
The work introduced in this paper contributes in the following aspects:(i)Preregistration is improved through estimation of rotation, scale, and translation separately. A multisource CNNs is developed to precisely classify various levels of rotation under sever distortion and help identify rotation extent with high accuracy. For 3D images, triplanar 2D CNNs is constructed to estimate parameters of affine matrix. This new preregistration performs better than stateoftheart ELASTIX and SURFbased methods.(ii)A twolayer adaptive registration framework is constructed and it performers better than other socalled twostep strategies.(iii)PCA is used to extract valuable features and introduced into traditional similarity metric as SSD, Pearson, and so forth. For 3D images, triplanar 2D PCA is proposed to process 3D registration problem. Experiment results show that convergence speed is accelerated with the new similarity standard.(iv)The proposed framework is tested under both synthetic and nature 2D and 3D images under various extent deformation. Experiment results show that our twolayer deep adaptive registration framework is able to identify the extent of rotation under sever deformation more precisely and correct large and complex distortions with high dice ratio than the comparative methods as it adaptively modify differences between images while others does not have any deep insight of deformation between images.
The rest of the paper is organized as follows. The whole architecture of the proposed twolayer adaptive registration framework for 2D and 3D images is illustrated in Section 2; Section 3 explains methodology of our CNNs classifier preregistration; Section 4 introduces our preregistration in combination with demons nonrigid registration and our new PCA related similarity metric; the proposed methods are evaluated in Section 5 under different datasets and evaluation principles; finally, the conclusion of this work is given in Section 6.
2. Architecture
2.1. CNNs for 2D Images
The whole workflow of our 2D image preregistration compared with traditional method is illustrated in Figure 1. In traditional algorithms, an affine matrix is calculated through correspondences between detected features, containing information of rotation, scale, and translation. This procedure is significantly influenced by accuracy of detected feature points. And under sever deformation images, traditional feature methods usually corrupt. Our algorithm processes each of the three above elements separately. By refining each procedure, accurate correspondences between fixed and moving image is obtained. It works as follows.(i)For rotation, firstly the CNNs classifier is trained offline in order to rectify rotation extent of image under sever deformation. The trained CNNs classifier can identify as much as 360 classes of rotation.(ii)For scale, image size information is utilized to achieve consistency between fixed and moving image.(iii)For translation, centroid of each image is calculated through statistical algorithm and translation is achieved by utilizing position information of centroids.
2.2. Triplanar CNNs for 3D Images
Different from the 2D image preregistration, CNNs classifier here is used for the slice location of one voxel (, , three directions) instead of the rotation identifier. The work flow of 3D image preregistration is showed in Figure 2. The main procedure includes sampling, slices classification, transform matrix calculation, and image transformation by the matrix. Using CNNs on 3D image registration is a new attempt to resolve image registration for high deformation. Detailed method is introduced in Section 5.2.
3. Preregistration
Our strategy consists of firstly preregistration through CNNs classifier on both 2D and 3D images and then utilizing CNNs and demons algorithm adaptively in the following nonrigid registration and finally improving similarity metric for acceleration of registration convergence speed. In this section, we show our preregistration methodology by introducing our CNNs rotation classifier.
3.1. Why We Use CNNs
(1) The Robustness of Classification. CNNs are a kind of data based classification method which undertakes training by appropriate amount of data. CNN is suitable for nearly any types of data and can make classification with high accuracy, especially for the low quality of fMRI, CT images or images under high deformation (Experiment in Section 5 shows these two kinds of real data are suitable for CNN processing method). Detailed CNNs structure and back propagation training method will be described in Section 3.5.
(2) Automatic Image Feature Perception. Nearly all kinds of preregistration method are based on precise feature perception so that the different feature perception methods are playing the key role in this procedure. Traditional image feature perception method is usually based on expert designated data feature. Usually experts give some fixed method to detect specific features of limited kinds of images. For example FLIRT method using intermodel voxel similarity measures where correlation ratio and mutual information are used to detect voxel relationships of different parts. This method has high limitation to the image sources, quality and variable settings. When exceptional case happens, some large deformation images are input for example, it will not work well. While the features from CNNs method are learned by network itself from training data such as edge, brightness, high or low frequency feature, distribution features and so on. Once the training data is updated, the network will get fit for more features automatically at the same time. Although long training time and complicated network variable learning makes CNN method not so easy to use, because of its high accuracy, it is still the image processing trend and future.
(3) High Efficiency Classification. Although the data training time of CNNs is long (depending on the detailed training method, network layer structure and hardware equipment like GPU), the total time spent on testing or classification is very short. Once the network is trained well, the only time consuming for processing is as short as linear operation.
Above all, even though there are some good affine transformation methods based on expert knowledge, we still need a smarter one to adapt to more complex image processing tasks in the future.
3.2. Theory of CNNs
The concept of deep learning was raised by Hinton and Salakhutdinov [9] in 2006, and it has brought great advances to machine learning since then. Deep learning aims to construct/use brain simulations to recognize data such as image/video, audio and text in an unsupervised way. Deep learning framework uses a multilayer “encoder” network to transform the highdimensional data into a lowdimensional code and a similar “decoder” network to recover the data from the code. Outputs of low layer network acts as inputs of higher layer network. The whole network aims to equal inputs and outputs without loss of information. By using lower layer features to represent higher layer feature/classification, distributed feature representation of data is found. Auto encoder, Sparse coding, Restricted Boltzmann Machine (RBM), Deep Belief Networks (DBNs) and CNNs are five kinds of deep learning framework. Convolutional neural networks are excellent deep learning architectures, which were firstly introduced by Fukushima [10] and applied for handwritten digit recognition. Image recognition and segmentation tasks have also successfully used CNNs since then, with an error rate as low as 0.23 percent on the MINST database [11]. Besides, it is of high speed and accuracy for image classification in [12]. In facial recognition [13, 14] and video quality analysis [15], CNNs also gained large decrease in error rate and root mean square error.
A CNN is a multilayer perceptron consisting of multilayers, each layer with a convolutional layer followed by a subsampling layer. Through locally connected networks, stationary features of natural images are exploited by the network topology. Firstly, images are sampled into small patches. In the convolutional layers, small feature detectors are learned based on these extracted samples. Then, a feature is calculated by convolution of the feature detector and the image at that point. In the sampling layer, the number of features is reduced to reduce computational complexity and introduce invariance properties. One significant property of features learnt by CNNs is invariance to translation, rotation, scale and other deformations. This twice feature extraction structure enables CNNs with high distortion tolerance when identifying input samples.
3.3. CNNs Methodology
The goal of CNNs has no difference with other classification methods. They both focus on minimal total square error. Here we use to denote the class number, and to denote the training dataset, the total square error function can be shown:
Here is the dimension of the dataset, the stands for the output from the network, activation function in CNNs is sigmoid function for faster convergence rate. For each single dataset , 2 can be describes as 3. The final aim of CNNs is to achieve smallest total square error between and . Consider
For traditional full connection neural network, BP (Back propagation method) is used to calculate partial derivative to get the minimum square error, usually the current layer, the output of can be shows as 4, where is sigmoid function. Consider
Unlike 4, as 5 shows, for the convolutional layer , the image features () from prior layer is convoluted by kernel which is different in different layers, is the offset of sigmoid function . Consider
For the sample layer, the image feature numbers and styles are the same with prior layer except the feature size is scaled down. Each feature contains a multi and addition kind offset. The down sample size in this paper is 2 which means the next layer image size is shrink two times by both weight and height. So through combination of 4 and 5, we can get sample equation 7 in which stands for the value of no. output with no. input features. By calculating and training kernels by back propagation method we can finally get the best features from different layers with high classification accuracy. Consider Constraint condition , and .
As shown in Figure 3, input images are defined as input layer; detailed introduction can be found in Sections 3.4 and 3.5. Hidden layer is the four pairs of convolutional and subsampling layer, which are denoted as , and called local connection layer. The output layer is a combination of full connection layer and softmax classifier for classification. Each layer of and is constructed with multimaps and each map is consisted of multi independent neural cells. Let and be the input and output for the th layer, and be the size of the input and output map, and be the number of input and output maps respectively of that layer. According to CNNs, .
3.4. CNNs Structure Design
We adopt a tenlayer CNNs perceptron network (input and output layers are included; convolutional and sample layers are separately calculated). Key variables setting including kernel size and sample rate of different layers in proposed CNN is showed in Table 1 and Figure 3. Learning rate alpha = 1, variable update batch size = 10, iteration times = 1000, any training and test images are normalized to size gray images with pixel size.

3.5. Training Image Rotation Classifier through CNNs
Our input images for training are difference images between fixed and moving image: . is under deformation with different extent of rotation. Each rotation angle of 360° is defined as one class, producing as much as 360 classes. Two distinguishing characters of CNNs are perception field and shared weights. Perception field means each neural cell in each layer is not connected wtih all neural cells in adjacent layers, but limited to a local area of neural cells ( as in Figure 3). Shared weights means the connection weight parameters () of every neural cell to the local area cell are the same. As shown in Figure 3, suppose size of input image is . After convolution with filters, the kernel size of which is , image changes into Ts1 of size. Image then scales into Tc1 in layer S1. After four pairs of and , the original image is represented as Ts4 of only matrix. In this hidden layer, all neural cells on feature maps are not all connected but with same weights. As a result, only weight parameters need to be calculated, greatly reducing computation complexity. An all connection exists between Ts4 matrix and output layer, eliminating disparity caused by partial connection in the hidden layer. Then softmax classifier identifies the matrix and outputs the detected results. After that, the parameters are finetuned through back propagation of 1000 times until convergence. After all these steps, a finite classifier is obtained.
4. TwoLayer Deep Iterative Registration Framework
4.1. Diffeomorphic Log Demons Registration
In the 19th century, Maxwell firstly introduced the concept of demons to illustrate a paradox of thermodynamics. In 1998, Thirion [16] proposed a registration algorithm under demons model, which had a high registration precision and efficiency through pixel velocities caused by edge based forces.
(i) Theory and Improvements of Demons Registration. Demons registration utilizes optical flow equation as basis forces for finding tiny deformations in temporal image sequences. For point in space, let and be intensity values in fixed image and moving image respectively. According to Thirion’s theory, 8 shows calculation of velocity allowing point to match the corresponding point in . Here, called internal edge force is the gradient image of fixed image and called the external force. In order to make the equation more stable and appropriate for image registration, Thirion added term . Later on, He Wang et al. added image forces of the moving image in the equation to improve convergence speed and stability of the registration as shown in 9. Parameter was proposed by Cachier et al. to adjust force strength. Consider
Vercauteren et al. [17] proposed nonparametric diffeomorphic demons algorithm. It considers the demons algorithm as a procedure of optimization on the whole space of velocity fields and adapts that procedure in a space of diffeomorphic transformations. The transformation result is smoother and more accurate. Then Vercauteren et al. [18] brings the process into logdomain, that is, he uses a stationary velocity field. Besides, the algorithm is symmetric with respect to the order of the input images. Lorenzi et al. [19] implements a symmetric local correlation coefficient to logdemons diffeomorphic algorithm. Lombaert et al. [1] proposed spectral logdemons to capture large deformations. Peyrat et al. [20] implements multichannel demons to register 4D timeseries cardiac images.
(ii) Diffeomorphic Log Demons Algorithm. Here, diffeomorphic log demons algorithm is briefly reminded. A diffeomorphic transformation is related to the exponential map of the velocity field (Algorithm 1) [1]. The logdemons framework alternates between optimization of a similarity metric updated by EulerLagrangian function in 10. In general, procedure of diffeomorphic log demons framework is described in Algorithm 2. Consider



4.2. New Similarity Metric by Combination of PCA
Mathematically, PCA is defined as an orthogonal linear transformation that transforms the data to a new coordinate system to extract the greatest variance in the data set. As a result, it is able to avoid influences caused by image biases. Traditionally, PCA is used for dimensionality reduction to facilitate classification, visualization, communication, and storage of highdimensional data. Here, PCA is applied in both 2D and 3D medical and usual images, and the detected feature representations are used as inputs of similarity metric to achieve anatomical correspondence and assist optimization procedure in registration.
There are many classical metric measures, such as SSD, mutual information (MI), cross correlation (CC), pattern intensity and also their corresponding improved edition. In this paper, Pearson, Spearman, kendall, SSD together with extracted features by PCA are utilized as the new similarity metric. Pearson, Spearman and Kendall are concepts in statistics and are frequently used in data mining. Pearson is short for Pearson productmoment correlation coefficient (PPMCC), which was developed to measure the linear correlation between two variables. Spearman’s rank correlation coefficient is a nonparametric measure of statistical dependence between two variables. Both of their value is between +1 and −1. Spearman has no requirement on variables, while pearson insists variables meets normal distribution. Our utilization of log demons registration avoids the influence brought by this.(i)For 2D images of size , firstly, PCA is applied to both fixed image and registered moving image , gaining and . Thus, most important information of image can be fully utilized by combination of and as inputs of pearson, spearman, and so forth, forming new similarity metric.(ii)For 3D images of size , firstly, PCA is applied to every slice of axis and gains a series of . By summarizing each of , is obtained. The same operation is carried out on and axis data, obtaining , . Then, PCA of both fixed and registered moving image is calculated. Thus, information of image can be fully utilized by combination of and as inputs of PPMCC, Spearman, and so forth. Workflow of this part is shown in Figure 4.
4.3. TwoLayer Iterative Registration Framework
Traditionally, the two step registration means an initial affine registration in the very beginning to coarsely rectify deformation and a following iterative registration to optimize a similarity metric achieving fine registration. We also adopt the two step strategy. But before the two step registration, we build up a classifier offline under CNNs training to identify rotation between fixed and moving image under very large distortions, then scale and translation. Also in each iteration, the initial and following registration are carried out iteratively. This feedback procedure assists achieving higher registration accuracy comparing with traditional SURF and affine method.
Besides, at the end of each iteration, we utilized a new similarity metric by combining PCA with traditional SSD, pearson, and so forth, fully containing most important features of image. As a result, convergence speed is highly accelerated than traditional SSD without PCA while maintaining the same registration accuracy. Algorithm 3 shows the over flow of the framework.
5. Experiment Results
In this section, the performance of the whole twolayer registration method is evaluated on both 2D and 3D images, synthetic and nature datasets. For comparison, traditional two step methods, ELASTIX and SURF related algorithm are used to preregistrate moving and fixed image. Then demons nonrigid registration is conducted. These methods are set as the baseline methods, which are denoted as ELASTIX+demons and SURF+demons. They all firstly use detected features initially to register images through affine transformation and original SSD as similarity metric under the diffeomorphic log demons framework. Our method is different from their framework both in preregistration and following nonrigid registration framework. For 2D images, firstly train a rotation classifier through CNNs and preregister moving image under large distortion and rotation, then together with scale and translation transformation, preregistration is done. For 3D images, a pretrained triplanar 2D CNNs is utilized to locate voxels, establishing feature correspondences. Finally, PCA related similarity metric iteratively registering images under diffeomorphic log demons framework.
The improvement of our twolayer method in registration accuracy, robustness to large deformation and rotation, and convergence speed are all assessed with ground truth data. Our matlab code is under Lombaert’s work [21] and Toolbox [22].
Specifically, we downloaded brain and lung dataset from BrainWeb MRI Simulated Normal Brain Database [23, 24] and Empire 10 challenge [25] for training and testing. Besides, training and tests were also carried out on classical Lena image, which is mostly used for image processing. T1, T2 image from [26] and ITK image [27] is also utilized in our experiment. BrainWeb MRI dataset contains 20 both 2D and 3D new normal anatomical models. Empire 10 challenge lung dataset includes 20 3D scans, each containing fixed and moving image pair. Description of lung dataset is shown in Table 2 [28].

5.1. 2D Image Registration
5.1.1. Synthetic Deformation Tests
A lot of registration has been evaluated on synthetic deformation images for algorithm test according to previous work [1, 26].
(1) Training CNNs Deformation Classifier. In this work, tenlayer CNNs are constructed to train various sources of images. We also tested other number layer CNNs, results showed that tenlayer CNNs achieved highest score when classifying rotation of deformed images. Four kinds of 2D source images [24–27] served as samples. An example of sample image is shown in Figure 5. Take image for example, linear transformation like rotation or translation is added to image by multipling rotation matrix coded through matlab; then four kinds of large and complex nonlinear transformation is added to through special processing by photoshop. image with only rotation is denoted as , with only deformation noted as , and with both rotation and deformation denoted as . The same notation is with and Lena image. Figure 6 is an illustration after all those processing. In order for accurate identification of rotation, here for training, difference image of and (with only rotation) is input of CNNs. After training, each angle of 360° is defined as one class, obtaining 360 class of distortion. For other CNNs, number of classes is 180, 90, 36.
Our test is carried out on computer of windows 7 system, with 8 GB RAM, i74770 CPU @ 3.4 GHz. Take BrainWeb data [23, 24], for example, Table 3 shows test results of the classifier according to these data.

As we can see from Table 3, when input images are resized into pixels, identification of rotation can reach as much as 99.86% for classifier 90; while images are resized into pixels, the identification accuracy for classifier 36 is 99.97%. All these are done under condition that training data is also for testing. When the testing data of BrainWeb is put into the trained classifier, accuracy reaches 99.56%, even lower than the training data itself, but still very high according to many usual classifiers. For Lena, ITK and T1 training data, classifier 36 gains an accuracy rate of 99.94%. Number of iteration is set to 1000 for every training.
(2) CNNs Preregistration Test. SURF related method, ELASTIX and our CNNs method are tested. Here, SURF related method means using firstly SURF algorithm to detect features and then affine transformation to initially register images.(i)When only rotation exists as in Figure 7, ELASTIX method failed; SURF method is able to identify rotation invariant features and establish accurate correspondences between Lena and . Established correspondences are shown as Lena_{corr} in Figure 7. Vertices of the yellow lines stand for feature correspondences. Red circle vertex stands for feature in original moving image while green cross stands for corresponding feature point in registered moving image Lena_{SURF}. Registered image is denoted as Lena_{ELASTIX} and Lena_{SURF}.(ii)However, when rotation and large deformation simultaneously appears in moving image as in Figure 8, both ELASTIX and SURF method crushed. Under such circumstances, in our tests SURF only found one pair of correspondence points. As there are not enough feature correspondences, initial registration failed.(iii)On the contrary, our trained CNNs classifier and following scale and translation operation directly identified Lena image’s rotation angle accurately (rotation), and turned it back to as in Figure 9. For better comparison, we used software to show ways of rotation processing in CNNs as SURF’s manner, feature detecting and matching in Lena_{analogy  CNN}. As enough number of socalled features are detected, CNNs is able to recover rotation added to .
(3) Accuracy Evaluation of Registration. Mathematically, dice ratio is used to evaluate overlap between two datasets. It is defined in 11. In this section, both dice ratio and subjective human evaluation method is used to assess accuracy of ELASTIX and SURF related registration and our method result
After preregistration in Section 5.1.1, ELASTIX and SURF related method performs diffeomorphic log demons algorithm iteratively to achieve for best registration; while our method iteratively carries out CNNs classifier and diffeomorphic log demons algorithm to optimize registration. This new twolayer registration framework makes full use of both preregistration and following demons method and registration results show that it indeed improves accuracy.
Figure 10 shows registration procedure and result of ELASTIX (Figure 10(c)) and SURF (Figure 10(b)) related method, while Figure 11 shows that of our method. When both rotation and deformation exists in image , our registration result is much better than and apparently. Besides, to test dice ratio of registration, original fixed image and registered moving image of two methods are put into function 11 separately. Dice ratio of ELASTIX and SURFdemons method is 0.889 and 0.88, while our CNNsdemonsiterative method achieves as much as 0.8964.
(a)
(b)
(c)
5.1.2. Lung Atlases
Description of lung dataset can be found in Table 2 [28]. Empire 10 lung datasets are firstly used for the MICCAI conference 2010. It contains 20 intrapatient thoracitic CT image pairs. Figures 12 and 13 shows our preregistration results of slice image 8 and 6 of 4D image pairs compared with that of Elastix tool. All images are shown with the help of tool vv [29]. From left to right in the figures are fixed image, moving image, preregistration result of ELASTIX and our proposed CNNS method, final registration result of above two methods with demons. It is obvious that our method can accurately rectify rotation, scale and translation deformation added to moving image. While ELASTIX preregistration failed to rectify rotation differences between fixed and moving images. Registered images are denoted as Elastixdemons and CNNsdemons. Diff images between fixed and registered image are denoted as Elastixdiff and CNNsdiff.
ELASTIX preregistration consuming time of each slice is shown in Figure 14. The shortest time of one slice is more than 1000 ms (1 s) and time for slice 8 is 3500 ms. Although training of our CNNs classifier costs long time, it is offline. And our CNNs rotation, scale and translation operation costs a total of only 39 ms. As a result, it is quite attractive for realtime clinical applications.
5.1.3. Brain Atlases
We select the cross section 2D image of the BrainWeb MRI 20 object, 10 for training and the other 10 for testing. From Figure 15, we can see that our proposed preregistration can rectify both rotation and translation more successfully than traditional Elastix affine registration.
5.2. An Attempt on 3D Image Registration by Using CNNs
For the 3D image registration part we focus on the brain atlases registration and give a CNNs 3D image registration method. We train brain atlas from 18 people’s 3D image data in BrainWeb Brain database by four steps: (1) Randomly select 10 label points by Normal distribution in 3D image. (2) Adjust 3D brain image and separate it to 2D image on three directions (, , ). (3) Test each 2D slice position by triplanar aftertrained CNNs classifier (each dimension enjoys one CNN network) and get the right slice position (predicted voxels). (4) Adjusting the 3D image to make label voxels and predict voxels that enjoy smallest hamming distance. Experiments shows the high accuracy CNNs classify results will greatly improve moving 3D image’s similarity to the fixed 3D image. The detailed procedure is shown in Figure 16.
5.3. Convergence Speed Evaluation of Registration
Sections 5.1 and 5.2 improve both registration accuracy and speed. In this section, we test registration accuracy on T2 brain medical data and focus on accelerating convergence speed of registration. PCA is introduced to extract valuable features and by combining features with SSD, Pearson, Spearman, Kendall, we get new similarity PCASSD, PCApearson, PCAspearson and Kendall. Original SSD is denoted as energy in Figure 17(e). The coursetofine (in here, three level is recommended) registration strategy is adopted in here. In Figure 17, horizontal axis stands for iteration times and vertical axis stands for the values of metric. Firstly, mean convergence extent of the threelevel registration is calculated. Then normalization is carried out on the mean value. Several conclusions can be gained:(1)both PCA related and original SSD methods converge regularly,(2)as a whole, PCASSD and PCAPearson methods perform best and converge faster than original SSD metric;(3)PCAspearman metric firstly converges fastest, but latterly it slows down;(4)Kendall metric performs worst compared with other metrics.
(a)
(b)
(c)
(d)
(e) Convergence extent
6. Conclusion
In this paper, a comprehensive method of constructing rotation classifier for images under severe deformation and rotation was proposed through CNNs. The classifier is able to identify distortion as much as 360 classes according to analysis of rotation angles. The classifier is utilized to assist our proposed twolayer deep adaptive registration framework. In each registration iteration, preregistration with identification of the trained classifier, scale, and translation operator and following diffeomorphic log demons registration facilitates each other one after another. Besides, proposed PCA related similarity metric helps achieve faster convergence speed. The new twolayer registration framework is compared with traditional diffeomorphic log demons registration in combination with stateoftheart ELASTIX and SURF preregistration. As baseline method carries out preregistration only once, large deformations cannot be fully modified. From tests on different image resources containing various kinds of both 2D and 3D, MRI, and CT datasets, our framework indeed outperforms the baseline method on both registration quality and convergence speed.
In the following work, we would combine other kinds of deep learning framework as independent subspace analysis (ISA) [30], sparse coding [31], and so forth to improve current registration. Also, more performance tests of the proposed twolayer registration framework should be carried out on more data resources. Besides, the proposed method performance should be compared with other deep learning models.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
Acknowledgments
This paper is supported by The Project for the National Key Technology R&D Program under Grant no. 2011BAC12B03, The Innovation Team of Beijing, The National Natural Science Foundation of China under Grant nos. 81370038 and 61100131, The Beijing Natural Science Foundation under Grant no. 7142012, The Beijing Nova Program under Grant no. Z141101001814107, The Science and Technology Project of Beijing Municipal Education Commission under Grant nos. KZ201310005004 and km201410005003, The Rixin Fund of Beijing University of Technology under Grant nos. 2013RXL04 and 2012RX03, and The Basic Research Fund of Beijing University of Technology under Grant no. 002000514312015.
References
 H. Lombaert, L. Grady, X. Pennec, N. Ayache, and F. Cheriet, “Spectral logdemons: diffeomorphic image registration with very large deformations,” International Journal of Computer Vision, vol. 107, no. 3, pp. 254–271, 2014. View at: Publisher Site  Google Scholar
 P. Markelj, D. Tomaževič, B. Likar, and F. Pernuš, “A review of 3D/2D registration methods for imageguided interventions,” Medical Image Analysis, vol. 16, no. 3, pp. 642–661, 2012. View at: Publisher Site  Google Scholar
 G. Song, N. Tustison, B. Avants, and J. Gee, “Lung CT image registration using diffeomorphic transformation models,” in Medical Image Analysis for the Clinic—A Grand Challenge, pp. 23–32, 2010. View at: Google Scholar
 http://fsl.fmrib.ox.ac.uk/fsl/fslwiki/FLIRT.
 S. Klein, M. Staring, K. Murphy, M. A. Viergever, and J. P. W. Pluim, “Elastix: a toolbox for intensitybased medical image registration,” IEEE Transactions on Medical Imaging, vol. 29, no. 1, pp. 196–205, 2010. View at: Publisher Site  Google Scholar
 D. P. Shamonin, E. E. Bron, B. P. F. Lelieveldt, M. Smits, S. Klein, and M. Staring, “Fast parallel image registration on CPU and GPU for diagnostic classification of Alzheimer's disease,” Frontiers in Neuroinformatics, vol. 7, article 50, 2014. View at: Publisher Site  Google Scholar
 C. Liu, J. Yuen, and A. Torralba, “SIFT flow: dense correspondence across scenes and its applications,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 33, no. 5, pp. 978–994, 2011. View at: Publisher Site  Google Scholar
 A. Prasoon, K. Peters, C. Lgel, F. Lauze, E. Dam, and M. Nielsen, “Deep feature learning for knee cartilage segmentation using a triplanar convolutional neural network,” in Proceedings of the International Conference on Medical Image Computing and ComputerAssisted Intervention (MICCAI '13), vol. 8150, pp. 246–253, 2013. View at: Google Scholar
 G. E. Hinton and R. R. Salakhutdinov, “Reducing the dimensionality of data with neural networks,” Science, vol. 313, no. 5786, pp. 504–507, 2006. View at: Publisher Site  Google Scholar  MathSciNet
 K. Fukushima, “Neocognitron: a selforganizing neural network model for a mechanism of pattern recognition unaffected by shift in position,” Biological Cybernetics, vol. 36, no. 4, pp. 193–202, 1980. View at: Publisher Site  Google Scholar
 D. Ciregan, U. Meier, and J. Schmidhuber, “Multicolumn deep neural networks for image classification,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR '12), pp. 3642–3649, June 2012. View at: Publisher Site  Google Scholar
 D. C. Cireşan, U. Meier, J. Masci, L. M. Gambardella, and J. Schmidhuber, “Flexible, high performance convolutional neural networks for image classification,” in Proceedings of the 22nd International Joint Conference on Artificial Intelligence (IJCAI '11), vol. 22, pp. 1237–1242, July 2011. View at: Publisher Site  Google Scholar
 S. Lawrence, C. L. Giles, A. C. Tsoi, and A. D. Back, “Face recognition: a convolutional neuralnetwork approach,” IEEE Transactions on Neural Networks, vol. 8, no. 1, pp. 98–113, 1997. View at: Publisher Site  Google Scholar
 M. Matsugu, K. Mori, Y. Mitari, and Y. Kaneda, “Subject independent facial expression recognition with robust face detection using a convolutional neural network,” Neural Networks, vol. 16, no. 56, pp. 555–559, 2003. View at: Publisher Site  Google Scholar
 P. Le Callet, C. ViardGaudin, and D. Barba, “A convolutional neural network approach for objective video quality assessment,” IEEE Transactions on Neural Networks, vol. 17, no. 5, pp. 1316–1327, 2006. View at: Publisher Site  Google Scholar
 J.P. Thirion, “Image matching as a diffusion process: an analogy with Maxwell's demons,” Medical Image Analysis, vol. 2, no. 3, pp. 243–260, 1998. View at: Publisher Site  Google Scholar
 T. Vercauteren, X. Pennec, A. Perchant, and N. Ayache, “Nonparametric diffeomorphic image registration with the demons algorithm,” in Medical Image Computing and ComputerAssisted Intervention—MICCAI 2007, vol. 4792 of Lecture Notes in Computer Science, pp. 319–326, Springer, Berlin, Germany, 2007. View at: Publisher Site  Google Scholar
 T. Vercauteren, X. Pennec, A. Perchant, and N. Ayache, “Symmetric logdemons diffeomorphic registration: a demonsbased approach,” in Proceedings of the 11th International Conference on Medical Image Computing and ComputerAssisted Intervention (MICCAI '08), vol. 5241 of Lecture Notes in Computer Science, pp. 754–761, Springer, New York, NY, USA, 2008. View at: Publisher Site  Google Scholar
 M. Lorenzi, N. Ayache, G. B. Frisoni, and X. Pennec, “LCCDemons: a robust and accurate symmetric diffeomorphic registration algorithm,” NeuroImage, vol. 81, pp. 470–483, 2013. View at: Publisher Site  Google Scholar
 J.M. Peyrat, H. Delingette, M. Sermesant, C. Xu, and N. Ayache, “Registration of 4D cardiac CT sequences under trajectory constraints with multichannel diffeomorphic demons,” IEEE Transactions on Medical Imaging, vol. 29, no. 7, pp. 1351–1368, 2010. View at: Publisher Site  Google Scholar
 http://www.mathworks.cn/matlabcentral/fileexchange/39194diffeomorphiclogdemonsimageregistration.
 https://github.com/rasmusbergpalm/DeepLearnToolbox.
 C. A. Cocosco, V. Kollokian, R. K.S. Kwan, G. B. Pike, and A. C. Evans, “BrainWeb: online interface to a 3D MRI simulated brain database,” NeuroImage, vol. 5, p. 425, 1997. View at: Google Scholar
 B. AubertBroche, A. C. Evans, and L. Collins, “A new improved version of the realistic digital brain phantom,” NeuroImage, vol. 32, no. 1, pp. 138–145, 2006. View at: Publisher Site  Google Scholar
 http://empire10.isi.uu.nl/.
 http://www.mathworks.cn/matlabcentral/fileexchange/21451multimodalitynonrigiddemonalgorithmimageregistration.
 http://www.itk.org/ITK/resources/software.html.
 K. Murphy, B. van Ginneken, J. M. Reinhardt et al., “Evaluation of registration methods on thoracic CT: the empire 10 challabnge,” IEEE Transactions on Medical Imaging, vol. 30, pp. 1901–1920, 2011. View at: Google Scholar
 http://elastix.isi.uu.nl/index.php.
 G. Wu, M. Kim, Q. Wang, Y. Gao, S. Liao, and D. Shen, “Unsupervised deep feature learning for deformable registration of MR brain images,” in Proceedings of the 16th International Conference on Medical Image Computing and ComputerAssisted Intervention (MICCAI '13), vol. 8150 of Lecture Notes in Computer Science, pp. 649–656, Springer, Nagoya, Japan, 2013. View at: Publisher Site  Google Scholar
 M. Liu, D. Zhang, P. T. Yap, and D. Shen, “Treeguided sparse coding for brain disease classification,” Medical Image Computing and ComputerAssisted Intervention, vol. 7512, no. 15, part 3, pp. 239–247, 2012. View at: Google Scholar
Copyright
Copyright © 2015 Liya Zhao and Kebin Jia. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.