Research Article  Open Access
A Novel Fusion Framework Based on Adaptive PCNN in NSCT Domain for WholeBody PET and CT Images
Abstract
The PET and CT fusion images, combining the anatomical and functional information, have important clinical meaning. This paper proposes a novel fusion framework based on adaptive pulsecoupled neural networks (PCNNs) in nonsubsampled contourlet transform (NSCT) domain for fusing wholebody PET and CT images. Firstly, the gradient average of each pixel is chosen as the linking strength of PCNN model to implement selfadaptability. Secondly, to improve the fusion performance, the novel summodified Laplacian (NSML) and energy of edge (EOE) are extracted as the external inputs of the PCNN models for low and highpass subbands, respectively. Lastly, the rule of max region energy is adopted as the fusion rule and different energy templates are employed in the low and highpass subbands. The experimental results on wholebody PET and CT data (239 slices contained by each modality) show that the proposed framework outperforms the other six methods in terms of the seven commonly used fusion performance metrics.
1. Introduction
Medical images are very significant to clinical diagnosis and treatment. However, only one modality of image could not provide sufficient clinical information. For example, the Positron Emission Tomography (PET) images only reflect functional information and the Computed Tomography (CT) images only reflect the anatomical information [1]. Therefore, it is necessary to use different modalities of images to provide complementary information to physicians for better diagnosis. Medical image fusion is the process of collaboratively combining the complementary information from multimodal source medical images into one single fused image for further process. The fused image is suitable for visual perception, analysis, and diagnosis which is of great clinical meaning [2, 3].
According to different image domains, image fusion methods are generally classified into two categories: spatial domainbased and transform domainbased methods [4]. The spatial domainbased methods perform on the pixels of the original image directly. They are simple but with poor performance as they would lead to reduction of the contrast and distortion of the spectral characteristics [2]. Currently, the image fusion methods are mainly based on the transform domain, such as discrete wavelet transform (DWT), nonsubsampled shearlet transform (NSST), contourlet transform, and nonsubsampled contourlet transform (NSCT) [1, 4]. By transforming, source images could be decomposed into high and lowpass subbands. After that, different fusion rules could be designed for different subbands. Finally, the inverse transform would be adopted to obtain the final fusion result.
It is known that the fusion result could be improved by making better use of image information. NSCT is a multiscale, multidirectional, and fully shiftinvariant transform which could well suppress pseudoGibbs phenomena [5, 6]. Besides, more information could be obtained by NSCT leading to better fusion performance. Thus, NSCT is suitable for image fusion [7–10]. For example, [5] proposed a fusion method based on NSCT which outperformed the DWTbased methods. However, the method did not make enough use of the feature information (e.g., edges, textures) contained by source images as only the gray value of pixel was considered in the lowpass subband. Besides, the fusion rule of lowpass subband, averaging method, could lead to loss of contrast and hence possible cancellation of few patterns in source images. Reference [8] aimed to improve the fusion performance by using maximum entropy of squares of the coefficients in a local window for the lowpass subband and maximum weighted summodified Laplacian for the highpass subbands. Quantitative evaluation results demonstrated that the method was superior to the compared methods. In [10], the average weighting and maximum regional energy rules were adopted as the fusion rules for low and highpass subbands, respectively. Results showed the effectiveness of the method on the fusion of images of brains.
However, there is still room to improve for the fusion performance of NSCTbased methods. PCNN, a visual cortexinspired neural network, involves the global features of the source images that could extract and contain the detailed information effectively [11, 12]. Therefore, many researchers are devoted to combine PCNN with NSCT for a better fusion performance [13–16]. For instance, [13] adopted PCNN in the highpass subbands of NSCT domain and achieved better fusion results than methods based on Laplacian, DWT, and NSCT. Nevertheless, this PCNN model had lack of adaptability as there were many parameters needed to be adjusted manually. To improve the adaptability, [14] developed a fusion method based on adaptive dualchannel unitlinking PCNN in NSCT domain to decrease the number of parameters and achieved satisfactory fusion performance. Human eyes are sensitive to edge, direction, and texture information rather than single pixels; however, methods in [13, 14] only considered the normalized gray level of image pixels as the external input of PCNN, which could affect the fusion effectiveness to some extent. Reference [15] employed an adaptive PCNNNSCT based fusion method where the orientation information was utilized as the linking strength in PCNN model. To make better use of edge information, [15] adopted the modified spatial frequency (SF) as the external input of PCNN, which achieved a preferable fusion performance. However, in [15], the PCNN was only adopted in the highpass subbands where the feature information contained by lowpass subbands was not considered enough.
In conclusion, it is observed that although the PCNNNSCT based methods could achieve good results, there are still some critical factors which could affect the fusion performance. One is that many parameters need to be set manually in PCNN, that is, lack of adaptability. Another is how to design the external input of PCNN and the fusion rules for more ideal fusion results. To address these issues, in this paper, a modified fusion framework based on PCNNNSCT is proposed for fusing wholebody PET and CT images. There are mainly three contributions. Firstly, for the sake of implementing the adaptability of PCNN, the average gradient of each pixel is utilized as the linking strength. Secondly, to take better advantages of feature information contained by input images, we use PCNN model in both the low and highpass subbands, where the novel summodified Laplacian (NSML) and the energy of edge (EOE) are selected as the external input of the PCNN model in low and highpass subbands, respectively. This is because NSML [14] could well reflect the edge details of lowpass subband and the EOE [16] could well retain the details of source input as it could denote the edge features in horizontal, vertical, and diagonal directions. Lastly, the region energy is chosen as the fusion rule and two different energy templates are used for the low and highpass subbands, respectively. Experiments on wholebody PET and CT images showed the effectiveness of our framework.
The remaining sections of this paper are organized as follows. Section 2 reviews the theory of NSCT and PCNN. Section 3 introduces the proposed framework in detail. Section 4 presents the experimental results and discussion. Finally, the conclusions are given in Section 5.
2. Background Knowledge
2.1. Nonsubsampled Contourlet Transform (NSCT)
NSCT is a multiscale, multidirectional, and translation invariant transform [17, 18]. Different from contourlet transform [19], NSCT does not employ downsamplers/upsamplers so that it could ensure the translation invariance and could effectively represent the edge and contour information. As a result, the pseudoGibbs phenomena could be well overcome which could improve the performance of image fusion. The structure of a NSCT with twostage decomposition is shown in Figure 1.
As shown in Figure 1, NSCT is composed by nonsubsampled pyramid (NSP) decomposition and nonsubsampled directional filter banks (NSDFB). These two parts could ensure the multiscale property and the multidirectional property, respectively. Firstly, the NSP is performed on the source image to achieve multiscale decomposition. Through NSP decomposition, one lowpass subband and one highpass subband could be obtained at each NSCT decomposition stage. Then, the NSDFB is employed on the highpass subbands at each stage to produce highpass directional coefficients. Through this, the more detailed directional information which is important for fusion could be extracted, with, finally, iterating the former steps on the lowpass subband until the defined decomposition levels are reached. As a result, one lowpass subband coefficient and several highpass subband coefficients with the same size as the source image could be obtained.
2.2. Pulse Couple Neuron Network (PCNN)
PCNN is a single layered, twodimensional array of laterally connected network of integrateandfire neurons [20]. Each particular neuron corresponds to one particular pixel, which would be also affected by the surrounding neurons. As the traditional PCNN model is complicated, a simplified PCNN model [21] is used in this paper. As shown in Figure 2, a PCNN neuron contains three parts: receptive field, linking modulation field, and pulse generator. Mathematically, the PCNN model could be descripted as
Firstly, the input of the neuron consists of two parts: the external input (external stimulus) and the pulse output of its neighboring neurons . Traditionally, is the normalized gray value of the corresponding pixel . Subsequently, the nonlinear modulation is performed in the linking modulation filed and then the internal activity could be obtained. Finally, is compared with the dynamic threshold . If is larger than , then the neuron would be ignited. In addition, is the total iteration times. and are the time constants and normalizing constants for the linked input and dynamic threshold, respectively. The linking strength reflects the weight of linking field that plays a key role in fusion. Traditionally, is chosen according to experiences which is lack of selfadaptability.
3. Our Proposed Fusion Framework
The linking strength plays a key role in PCNN which determines the lifting range and exciting character. It is generally determined manually according to experience which is lack of selfadaptability. In addition, the external input of the PCNN is usually the pixel coefficient in spatial domain or in the transform domain which does not make full use of the edge information of the source image. To address these shortcomings, this paper proposes a novel PCNNNSCT based fusion framework where the regional average gradient is as the linking strength to achieve selfadaptability. Moreover, the novel summodified Laplacian (NSML) and energy of edge (EOE) are calculated as the external inputs for the low and highpass PCNNs, respectively. Besides, it is known that the fusion rules determine the fusion effectiveness. Generally, the maximum or averaging rules are used as the fusion rules which will lead to loss of contrast or information. Thus, the rule of max region energy is employed as the fusion rule in this paper. To further improve the fusion effectiveness, different energy templates are used for low and highpass subbands.
Since devices that capture CT and PET images are different from each other, it is necessary to correct spatial displacements such as offset, scale, and geometric distortion in advance. In image fusion, the alignment or registration process is important. In this paper, we adopt the registration framework that we proposed in [22] including preprocessing, feature extraction, and registration, which would not be introduced in this paper.
Figure 3 illustrates the process flow of our proposed fusion framework. The detailed fusion process consists of the following steps:(1)The NSCT decomposition is performed on both the source PET and CT images. Then one lowpass subband image and a series of highpass subband images could be obtained.(2)The NSML feature and the EOE feature are computed for the low and highpass subbands, respectively. These features would be used as the external inputs of the subsequent PCNN models. After that, both the input coefficients of the PCNN models of the low and highpass subbands are normalized to .(3)Through PCNN, the firing maps could be produced for these subbands.(4)The firing maps of the low and highpass subbands could be fused according to the rule of max region energy. Note that different energy templates are used in low and highpass subbands.(5)After getting the fused subbands, the inverse NSCT is applied and then the final fused image could be produced.
3.1. Novel SumModified Laplacian (NSML)
Laplacian energy could well reflect the edge features of the lowpass subband images, so the NSML is employed in this paper as the external input of PCNN to improve the fusion performance. The definition of NSML is as where means the coefficients of the lowpass subband image at . and denote the sizes of neighbor window which are generally , , and . In this paper, is chosen for computing NSML. In addition, means the weighted template, which emphasizes the coefficient at the center of window.
3.2. Energy of Edge (EOE)
Traditionally, the gray value of the original pixel is utilized as the input of PCNN which does not consider the effect of the neighborhoods. Moreover, it is known that human eyes are more sensitive to edge and directional and texture information rather than single pixel information. In order to make better use of the edge information, the EOE is selected as the external stimulation for PCNN model for highpass subbands. The definition of EOE is shown as where is the weighted template. denotes the neighborhood of . means the coefficients of the highpass subband image at . , , and mean the directional filtering operators. The EOE could well reflect the edge information in horizontal, vertical, and diagonal directions.
3.3. Region Energy
The rules of weighted average and larger absolute value are commonly used as the fusion rules to calculate the fused coefficients. However, these methods might lead to losing part of details of the source images and reducing the contrast. To improve the fusion performance, the relationships of the neighboring regions should be considered. In this paper, the rule of max region energy is adopted as the fusion rule. The definition of region energy is shown in (4) and (5): where means the coefficients of the firing map image at . is the energy template. Different energy templates are used for low and highpass subbands. is the energy template used for lowpass subband images. And is used for highpass subbands to strengthen the coefficients at the window center. In addition, the rule of max region energy is defined as (6), where is the fused coefficient and and are the subband images which are obtained by employing NSCT on the source images and . and are the region energy for the corresponding firing maps of and , respectively.
3.4. Improved PCNN
The linking strength , reflecting the variance of the coefficients of the subband images, plays a key role in fusion process. In traditional PCNN based fusion process, is usually assigned manually according to experiences and all neurons in PCNN are set with the same linking strength value. However, according to [17], the values of should not be all the same in different neurons. Besides, it has been verified that is relevant to the image features of the corresponding pixels of the input images. If the external input coefficient is larger, then a higher value should be assigned to . As a result, considering the edge information of the source image, the region average gradient is selected as the linking strength to improve its selfadaptability in this paper. The higher value of the average gradient, the higher clarity of the image. The definition of region average gradient is shown as where means the coefficients of the input image at . is the average gradient at . The larger value of means that the image has a higher clarity. Moreover, the higher the value of is, the earlier the ignition of the correspondent neuron would be which will result in better use of the detailed information of the input image and increase the fusion effectiveness.
4. Results and Discussions
4.1. Experimental Data and Platform
To verify the effectiveness of the proposed fusion framework, we conduct experiments on wholebody PET/CT data which are provided by the General Hospital of Shenyang Military Area Command, Shenyang, China. All of our data is in conformity with laws and ethical standards. Each model of data consists of 239 slices. The original PET and CT images captured from devices are different from each other, such as the images sizes (the sizes of PET and CT images are 128 × 128 and 512 × 512, resp.) and the scale and geometric distortion, which would affect the effectiveness of fusion. Thus, it is necessary to do an interpolation and a registration process before fusion so that all PET images would be resized to 512 × 512 and the geometric displacement would be corrected. In this paper, we use the registration framework that we proposed in [22] to preprocess the PET and CT images before fusion. This framework could achieve a good performance on the registration of wholebody PET and CT images. The experimental platform is Intel® Core TM i72600 CPU @ 3.40 GHz, 8 G RAM, 1 T hard disk, Windows 7 OS. The integrated development environment is the MATLAB 2015b. Besides, the ITKSNAP is utilized to view images in this paper.
4.2. Performance Measures
To quantitatively evaluate fusion performance of the proposed framework and the other compared methods, seven commonly used metrics are applied in this paper including average gradient (AG), Shannon entropy (EN), joint entropy (JE), cross entropy (CE), image quality index (IQI), , and [2, 10, 23].
The average gradient reflects the variance of the gray value which could be used to evaluate the clarity of one image. The higher value of gradient value denotes that the image is clearer and the fusion performance is better. The definition is shown as where means the gray value of the input image at . and mean the size of the input image.
The Shannon entropy measures the information content in the source image. The higher value of Shannon entropy of the fused image, the more information it contained, the better fusion performance it has. The definition of Shannon entropy is shown as where means the probability of the gray level in the source image.
The joint entropy denotes the similarity of the fused image with the source images. The definition is as where means the joint probability of the gray level in image and the gray level in image . The larger joint entropy denotes better performance.
The cross entropy measures the difference between the fused image and the source images. The cross entropy image fusion metric is defined as the average of the relative entropies between the source images and and the fused image ; see (11)where means relative entropy of two probability distribution functions and . and are the normalized histograms of the source images and , respectively. The lower cross entropy means the better fusion performance.
The IQI between the fused image and the source images and are defined as where , , and are the mean of , variance of , and the covariance between and . If the IQI achieves higher values, it means the better fusion quality of the fused image.
is an edgedependent fusion quality metric based on the IQI. The definition of is introduced as where the definition of IQI is given in (13). expresses the contribution of the edge images compared with the original images (usually ). and denote the structural similarity measures between the source images and the fused image in a sliding local window . is defined as , where the overall saliency of a window is defined as . and denote saliency of image and image in window . The larger means better fusion performance.
is a fusion performance metric based on the edge information to measure the similarity among images. The definition of is defined as where and denote the edge strength and orientation preservation values, respectively. The larger value means the better fusion performance.
4.3. Results Evaluation and Discussions
To evaluate the performance of the proposed PET and CT fusion framework, we compare the fusion results of the proposed framework with the following six methods on the same dataset.
Method 1. It is DWTbased method where the fused coefficients are selected according to the rule of maximum absolute values.
Method 2. It is NSCTbased method (NSCT) where the maximum selection rule is used for both the low and highpass subbands [24].
Method 3. It is NSCTPCNN based method (NSCT_PCNN_1) where the same rules with Method 2 are used here.
Method 4. It is image fusion method in the NSCT domain using spatial frequencybased PCNN method (NSCTSFPCNN) where the maximum selection rule is used for both the low and highpass subbands [25].
Method 5. It is an improved NSCTPCNN based fusion method [10, 12, 14] named as NSCT_PCNN_2, where the NSML and modified spatial frequency (MSF) are as the external input of PCNN for low and highpass subbands, respectively.
Method 6. It is an improved NSCTPCNN based fusion method [16] named as NSCT_PCNN_3, where EOE is as the external input of the PCNN.
Figure 4 shows three examples of the fusion results obtained by these methods. (a) and (b) are the source CT images and PET images to be fused. (c)–(i) are the fusion results obtained by different compared methods. (c) shows the fusion results based on DWT. (d) presents the fusion results based on NSCT. (e) exhibits the fusion results produced by NSCT_PCNN_1. (f) shows the fusion results produced by NSCT_SF_PCNN. (g) and (h) are the fusion results produced by NSCT_PCNN_2 and NSCT_PCNN_3, respectively. (i) exhibits the fusion results produced by the proposed fusion framework. From the visual analysis of Figure 4, it is observed that the proposed framework could successfully preserve both the feature information of the CT images (e.g., the bony structures) and the PET images (e.g., high metabolic areas). Specifically, the fusion results produced by NSCT have the worst contrast and there are many artifacts introduced in the fusion results produced by DWT. Besides, the rest of the compared methods have similar performance with the proposed framework and it is difficult to determine which one is better by eyes.
(a)
(b)
(c)
(d)
(e)
(f)
(g)
(h)
(i)
In order to evaluate the fusion results more intuitively, the 3D reconstruction is performed on the source data and the fusion result sequences (see Figure 5). In addition, the pseudocolor coding is also employed on these reconstructions for visualizing the results better. In Figure 5, the first row represents transverse plane of the reconstruction. The second row and third row represent the sagittal plane and the coronal plane, respectively. (a) and (b) are the reconstructions of the source CT and PET image sequences, respectively. (c)–(h) are the reconstructions of the fusion result sequences by DWT, NSCT, NSCT_PCNN_1, NSCT_SF_PCNN, NSCT_PCNN_2, and NSCT_PCNN_3, respectively. (i) is the reconstruction of the fusion results obtained by the proposed framework. From Figure 5, the fusion results obtained by the proposed framework could preserve both the structural feature information of CT and the functional feature information of PET. Besides, there are many artifacts in the results produced by DWT. Similar with Figure 4, the fusion results of NSCT are still the worst with the lowest contrast.
(a)
(b)
(c)
(d)
(e)
(f)
(g)
(h)
(i)
According to Figures 4 and 5, it is difficult to distinguish which method achieves the best performance directly. In order to evaluate the fusion results between different algorithms objectively and quantitatively, this paper adopts seven metrics for performance evaluation which have been introduced in Section 4.2. Table 1 shows the average values of these metrics on 239 pairs of PET and CT images. Figures 6 and 7 are the bar charts corresponding to Table 1. Note that the lower the value of CE, the better the fusion performance. The higher the values of the rest of metrics, the better the fusion performance. It can be seen that the performance of our proposed framework on these metrics is always on the top two among all algorithms which outperforms other methods. Although the DWT performs best on AG, CE, and , its values of EN, HE, and are worse than the others. In addition, the NSCT gives poorer results than other NSCTPCNN based algorithms because PCNN could make use of the global feature information of the image which would improve the fusion performance. Following the analysis and discussion, a conclusion can be drawn that the proposed fusion framework outperforms the other image fusion methods for the fusion of wholebody PET and CT images. The fusion results of our framework could combine more information which is useful for diagnoses.

5. Conclusions
In this paper, a novel fusion framework based on adaptive PCNN in NSCT domain is proposed for fusing the wholebody PET and CT images. Our framework utilizes the average gratitude of each pixel as the linking strength in PCNN model to make it more adaptive. Besides, in order to make full use of the feature information contained by the input images, the NSML and EOE are as the linked input of the PCNN model in lowpass and highpass subbands, respectively, to improve the fusion performance. Moreover, the rule of max region energy is adopted as the fusion rule and different energy templates are used for low and highpass subbands. The experiments on wholebody PET and CT images with well alignments demonstrate the good performance of the proposed framework. An evaluation on seven metrics including AG, EN, JE, CE, IQI, , and illustrates objectively that the proposed framework outperforms the other six methods.
Conflicts of Interest
The authors declare that there are no conflicts of interest regarding the publication of this paper.
Acknowledgments
This work was supported by the National Natural Science Foundation of China (no. 61472073, no. 61272176).
References
 V. Bhateja, A. Srivastava, A. Moin, and A. LayEkuakille, “NSCT based multispectral medical image fusion model,” in Proceedings of the IEEE International Symposium on Medical Measurements and Applications (MeMeA '16), pp. 1–5, Benevento, Italy, May 2016. View at: Publisher Site  Google Scholar
 S. Singh, D. Gupta, R. S. Anand, and V. Kumar, “Nonsubsampled shearlet based CT and MR medical image fusion using biologically inspired spiking neural network,” Biomedical Signal Processing and Control, vol. 18, pp. 91–101, 2015. View at: Publisher Site  Google Scholar
 X. Xu, D. Shan, G. Wang, and X. Jiang, “Multimodal medical image fusion using PCNN optimized by the QPSO algorithm,” Applied Soft Computing Journal, vol. 46, pp. 588–595, 2016. View at: Publisher Site  Google Scholar
 Z. Fu, X. Wang, J. Xu, N. Zhou, and Y. Zhao, “Infrared and visible images fusion based on RPCA and NSCT,” Infrared Physics and Technology, vol. 77, pp. 114–123, 2016. View at: Publisher Site  Google Scholar
 Q. Zhang and B.L. Guo, “Multifocus image fusion using the nonsubsampled contourlet transform,” Signal Processing, vol. 89, no. 7, pp. 1334–1346, 2009. View at: Publisher Site  Google Scholar  Zentralblatt MATH
 Y.Q. Wu, C. Wu, and S.H. Wu, “Fusion of remote sensing images based on nonsubsampled contourlet transform and region segmentation,” Journal of Shanghai Jiaotong University (Science), vol. 16, no. 6, pp. 722–727, 2011. View at: Publisher Site  Google Scholar
 G. Bhatnagar, Q. M. J. Wu, and Z. Liu, “A new contrast based multimodal medical image fusion framework,” Neurocomputing, vol. 157, pp. 143–152, 2015. View at: Publisher Site  Google Scholar
 P. Ganasala and V. Kumar, “CT and MR image fusion scheme in nonsubsampled contourlet transform domain,” Journal of Digital Imaging, vol. 27, no. 3, pp. 407–418, 2014. View at: Publisher Site  Google Scholar
 N. Amini, E. Fatemizadeh, and H. Behnam, “MRIPET image fusion based on NSCT transform using local energy and local variance fusion rules,” Journal of Medical Engineering and Technology, vol. 38, no. 4, pp. 211–219, 2014. View at: Publisher Site  Google Scholar
 Z. Y. Jin and Y. J. Wang, “Multimodality medical image fusion method based on nonsubsampled contourlet transform,” Chinese Journal of Medical Physics, vol. 33, no. 5, pp. 445–450, 2016. View at: Google Scholar
 S. Das and M. K. Kundu, “NSCTbased multimodal medical image fusion using pulsecoupled neural network and modified spatial frequency,” Medical and Biological Engineering and Computing, vol. 50, no. 10, pp. 1105–1114, 2012. View at: Publisher Site  Google Scholar
 T. Z. Xiang, R. R. Gao, and L. Yan, “MultiScale fusion of infrared and visible images based on PCNN,” Journal of Geomatics Science and Technology, vol. 3, pp. 273–278, 2016. View at: Google Scholar
 M.L. Li, Y.J. Li, H.M. Wang, and K. Zhang, “Fusion algorithm of infrared and visible images based on NSCT and PCNN,” OptoElectronic Engineering, vol. 37, no. 6, pp. 90–95, 2010. View at: Publisher Site  Google Scholar
 T. Xiang, L. Yan, and R. Gao, “A fusion algorithm for infrared and visible images based on adaptive dualchannel unitlinking PCNN in NSCT domain,” Infrared Physics and Technology, vol. 69, pp. 53–61, 2015. View at: Publisher Site  Google Scholar
 L. Yan and T.Z. Xiang, “Fusion of infrared and visible images based on edge feature and adaptive PCNN in NSCT domain,” Acta Electronica Sinica, vol. 44, no. 4, pp. 761–766, 2016. View at: Publisher Site  Google Scholar
 J. Wang, Q. Li, Z. Jia, N. Kasabov, and J. Yang, “A novel multifocus image fusion method using PCNN in nonsubsampled contourlet transform domain,” Optik, vol. 126, no. 20, pp. 2508–2511, 2015. View at: Publisher Site  Google Scholar
 Y. Chai, H. F. Li, and M. Y. Guo, “Multifocus image fusion scheme based on features of multiscale products and PCNN in lifting stationary wavelet domain,” Optics Communications, vol. 284, no. 5, pp. 1146–1158, 2011. View at: Google Scholar
 A. L. da Cunha, J. Zhou, and M. N. Do, “The nonsubsampled contourlet transform: theory, design, and applications,” IEEE Transactions on Image Processing, vol. 15, no. 10, pp. 3089–3101, 2006. View at: Publisher Site  Google Scholar
 M. N. Do and M. Vetterli, “The contourlet transform: an efficient directional multiresolution image representation,” IEEE Transactions on Image Processing, vol. 14, no. 12, pp. 2091–2106, 2005. View at: Publisher Site  Google Scholar
 Z. Wang, Y. Ma, F. Cheng, and L. Yang, “Review of pulsecoupled neural networks,” Image and Vision Computing, vol. 28, no. 1, pp. 5–13, 2010. View at: Publisher Site  Google Scholar
 C. Xi, Z. Wei, and G. Fei, “Image fusion algorithm based on adaptive pulse coupled neural networks in curvelet domain,” in Proceedings of the IEEE 10th International Conference on Signal Processing (ICSP '10), pp. 845–848, October 2010. View at: Publisher Site  Google Scholar
 Z. Song, H. Jiang, Q. Yang, Z. Wang, and G. Zhang, “A registration method based on contour point cloud for 3D wholebody PET and CT images,” BioMed Research International, vol. 2017, Article ID 5380742, 11 pages, 2017. View at: Publisher Site  Google Scholar
 M. B. A. Haghighat, A. Aghagolzadeh, and H. Seyedarabi, “A nonreference image fusion metric based on mutual information of image features,” Computers and Electrical Engineering, vol. 37, no. 5, pp. 744–756, 2011. View at: Publisher Site  Google Scholar  Zentralblatt MATH
 S. Li, B. Yang, and J. Hu, “Performance comparison of different multiresolution transforms for image fusion,” Information Fusion, vol. 12, no. 2, pp. 74–84, 2011. View at: Publisher Site  Google Scholar
 X.B. Qu, J.W. Yan, H.Z. Xiao, and Z.Q. Zhu, “Image fusion algorithm based on spatial frequencymotivated pulse coupled neural networks in nonsubsampled contourlet transform domain,” Acta Automatica Sinica, vol. 34, no. 12, pp. 1508–1514, 2008. View at: Publisher Site  Google Scholar
Copyright
Copyright © 2017 Zhiying Song et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.