Table of Contents Author Guidelines Submit a Manuscript
Computational and Mathematical Methods in Medicine
Volume 2017, Article ID 8407019, 9 pages
https://doi.org/10.1155/2017/8407019
Research Article

A Novel Fusion Framework Based on Adaptive PCNN in NSCT Domain for Whole-Body PET and CT Images

Software College, Northeastern University, Shenyang 110819, China

Correspondence should be addressed to Huiyan Jiang; nc.ude.uen.liam@gnaijyh

Received 9 January 2017; Revised 12 March 2017; Accepted 28 March 2017; Published 3 April 2017

Academic Editor: Michele Migliore

Copyright © 2017 Zhiying Song et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

The PET and CT fusion images, combining the anatomical and functional information, have important clinical meaning. This paper proposes a novel fusion framework based on adaptive pulse-coupled neural networks (PCNNs) in nonsubsampled contourlet transform (NSCT) domain for fusing whole-body PET and CT images. Firstly, the gradient average of each pixel is chosen as the linking strength of PCNN model to implement self-adaptability. Secondly, to improve the fusion performance, the novel sum-modified Laplacian (NSML) and energy of edge (EOE) are extracted as the external inputs of the PCNN models for low- and high-pass subbands, respectively. Lastly, the rule of max region energy is adopted as the fusion rule and different energy templates are employed in the low- and high-pass subbands. The experimental results on whole-body PET and CT data (239 slices contained by each modality) show that the proposed framework outperforms the other six methods in terms of the seven commonly used fusion performance metrics.

1. Introduction

Medical images are very significant to clinical diagnosis and treatment. However, only one modality of image could not provide sufficient clinical information. For example, the Positron Emission Tomography (PET) images only reflect functional information and the Computed Tomography (CT) images only reflect the anatomical information [1]. Therefore, it is necessary to use different modalities of images to provide complementary information to physicians for better diagnosis. Medical image fusion is the process of collaboratively combining the complementary information from multimodal source medical images into one single fused image for further process. The fused image is suitable for visual perception, analysis, and diagnosis which is of great clinical meaning [2, 3].

According to different image domains, image fusion methods are generally classified into two categories: spatial domain-based and transform domain-based methods [4]. The spatial domain-based methods perform on the pixels of the original image directly. They are simple but with poor performance as they would lead to reduction of the contrast and distortion of the spectral characteristics [2]. Currently, the image fusion methods are mainly based on the transform domain, such as discrete wavelet transform (DWT), nonsubsampled shearlet transform (NSST), contourlet transform, and nonsubsampled contourlet transform (NSCT) [1, 4]. By transforming, source images could be decomposed into high- and low-pass subbands. After that, different fusion rules could be designed for different subbands. Finally, the inverse transform would be adopted to obtain the final fusion result.

It is known that the fusion result could be improved by making better use of image information. NSCT is a multiscale, multidirectional, and fully shift-invariant transform which could well suppress pseudo-Gibbs phenomena [5, 6]. Besides, more information could be obtained by NSCT leading to better fusion performance. Thus, NSCT is suitable for image fusion [710]. For example, [5] proposed a fusion method based on NSCT which outperformed the DWT-based methods. However, the method did not make enough use of the feature information (e.g., edges, textures) contained by source images as only the gray value of pixel was considered in the low-pass subband. Besides, the fusion rule of low-pass subband, averaging method, could lead to loss of contrast and hence possible cancellation of few patterns in source images. Reference [8] aimed to improve the fusion performance by using maximum entropy of squares of the coefficients in a local window for the low-pass subband and maximum weighted sum-modified Laplacian for the high-pass subbands. Quantitative evaluation results demonstrated that the method was superior to the compared methods. In [10], the average weighting and maximum regional energy rules were adopted as the fusion rules for low- and high-pass subbands, respectively. Results showed the effectiveness of the method on the fusion of images of brains.

However, there is still room to improve for the fusion performance of NSCT-based methods. PCNN, a visual cortex-inspired neural network, involves the global features of the source images that could extract and contain the detailed information effectively [11, 12]. Therefore, many researchers are devoted to combine PCNN with NSCT for a better fusion performance [1316]. For instance, [13] adopted PCNN in the high-pass subbands of NSCT domain and achieved better fusion results than methods based on Laplacian, DWT, and NSCT. Nevertheless, this PCNN model had lack of adaptability as there were many parameters needed to be adjusted manually. To improve the adaptability, [14] developed a fusion method based on adaptive dual-channel unit-linking PCNN in NSCT domain to decrease the number of parameters and achieved satisfactory fusion performance. Human eyes are sensitive to edge, direction, and texture information rather than single pixels; however, methods in [13, 14] only considered the normalized gray level of image pixels as the external input of PCNN, which could affect the fusion effectiveness to some extent. Reference [15] employed an adaptive PCNN-NSCT based fusion method where the orientation information was utilized as the linking strength in PCNN model. To make better use of edge information, [15] adopted the modified spatial frequency (SF) as the external input of PCNN, which achieved a preferable fusion performance. However, in [15], the PCNN was only adopted in the high-pass subbands where the feature information contained by low-pass subbands was not considered enough.

In conclusion, it is observed that although the PCNN-NSCT based methods could achieve good results, there are still some critical factors which could affect the fusion performance. One is that many parameters need to be set manually in PCNN, that is, lack of adaptability. Another is how to design the external input of PCNN and the fusion rules for more ideal fusion results. To address these issues, in this paper, a modified fusion framework based on PCNN-NSCT is proposed for fusing whole-body PET and CT images. There are mainly three contributions. Firstly, for the sake of implementing the adaptability of PCNN, the average gradient of each pixel is utilized as the linking strength. Secondly, to take better advantages of feature information contained by input images, we use PCNN model in both the low- and high-pass subbands, where the novel sum-modified Laplacian (NSML) and the energy of edge (EOE) are selected as the external input of the PCNN model in low- and high-pass subbands, respectively. This is because NSML [14] could well reflect the edge details of low-pass subband and the EOE [16] could well retain the details of source input as it could denote the edge features in horizontal, vertical, and diagonal directions. Lastly, the region energy is chosen as the fusion rule and two different energy templates are used for the low- and high-pass subbands, respectively. Experiments on whole-body PET and CT images showed the effectiveness of our framework.

The remaining sections of this paper are organized as follows. Section 2 reviews the theory of NSCT and PCNN. Section 3 introduces the proposed framework in detail. Section 4 presents the experimental results and discussion. Finally, the conclusions are given in Section 5.

2. Background Knowledge

2.1. Nonsubsampled Contourlet Transform (NSCT)

NSCT is a multiscale, multidirectional, and translation invariant transform [17, 18]. Different from contourlet transform [19], NSCT does not employ down-samplers/up-samplers so that it could ensure the translation invariance and could effectively represent the edge and contour information. As a result, the pseudo-Gibbs phenomena could be well overcome which could improve the performance of image fusion. The structure of a NSCT with two-stage decomposition is shown in Figure 1.

Figure 1: Two-stage decomposition framework of NSCT.

As shown in Figure 1, NSCT is composed by nonsubsampled pyramid (NSP) decomposition and nonsubsampled directional filter banks (NSDFB). These two parts could ensure the multiscale property and the multidirectional property, respectively. Firstly, the NSP is performed on the source image to achieve multiscale decomposition. Through NSP decomposition, one low-pass subband and one high-pass subband could be obtained at each NSCT decomposition stage. Then, the NSDFB is employed on the high-pass subbands at each stage to produce high-pass directional coefficients. Through this, the more detailed directional information which is important for fusion could be extracted, with, finally, iterating the former steps on the low-pass subband until the defined decomposition levels are reached. As a result, one low-pass subband coefficient and several high-pass subband coefficients with the same size as the source image could be obtained.

2.2. Pulse Couple Neuron Network (PCNN)

PCNN is a single layered, two-dimensional array of laterally connected network of integrate-and-fire neurons [20]. Each particular neuron corresponds to one particular pixel, which would be also affected by the surrounding neurons. As the traditional PCNN model is complicated, a simplified PCNN model [21] is used in this paper. As shown in Figure 2, a PCNN neuron contains three parts: receptive field, linking modulation field, and pulse generator. Mathematically, the PCNN model could be descripted as

Figure 2: The model of PCNN.

Firstly, the input of the neuron consists of two parts: the external input (external stimulus) and the pulse output of its neighboring neurons . Traditionally, is the normalized gray value of the corresponding pixel . Subsequently, the nonlinear modulation is performed in the linking modulation filed and then the internal activity could be obtained. Finally, is compared with the dynamic threshold . If is larger than , then the neuron would be ignited. In addition, is the total iteration times. and are the time constants and normalizing constants for the linked input and dynamic threshold, respectively. The linking strength reflects the weight of linking field that plays a key role in fusion. Traditionally, is chosen according to experiences which is lack of self-adaptability.

3. Our Proposed Fusion Framework

The linking strength plays a key role in PCNN which determines the lifting range and exciting character. It is generally determined manually according to experience which is lack of self-adaptability. In addition, the external input of the PCNN is usually the pixel coefficient in spatial domain or in the transform domain which does not make full use of the edge information of the source image. To address these shortcomings, this paper proposes a novel PCNN-NSCT based fusion framework where the regional average gradient is as the linking strength to achieve self-adaptability. Moreover, the novel sum-modified Laplacian (NSML) and energy of edge (EOE) are calculated as the external inputs for the low- and high-pass PCNNs, respectively. Besides, it is known that the fusion rules determine the fusion effectiveness. Generally, the maximum or averaging rules are used as the fusion rules which will lead to loss of contrast or information. Thus, the rule of max region energy is employed as the fusion rule in this paper. To further improve the fusion effectiveness, different energy templates are used for low- and high-pass subbands.

Since devices that capture CT and PET images are different from each other, it is necessary to correct spatial displacements such as offset, scale, and geometric distortion in advance. In image fusion, the alignment or registration process is important. In this paper, we adopt the registration framework that we proposed in [22] including preprocessing, feature extraction, and registration, which would not be introduced in this paper.

Figure 3 illustrates the process flow of our proposed fusion framework. The detailed fusion process consists of the following steps:(1)The NSCT decomposition is performed on both the source PET and CT images. Then one low-pass subband image and a series of high-pass subband images could be obtained.(2)The NSML feature and the EOE feature are computed for the low- and high-pass subbands, respectively. These features would be used as the external inputs of the subsequent PCNN models. After that, both the input coefficients of the PCNN models of the low- and high-pass subbands are normalized to .(3)Through PCNN, the firing maps could be produced for these subbands.(4)The firing maps of the low- and high-pass subbands could be fused according to the rule of max region energy. Note that different energy templates are used in low- and high-pass subbands.(5)After getting the fused subbands, the inverse NSCT is applied and then the final fused image could be produced.

Figure 3: The process flow of the proposed image fusion framework.
3.1. Novel Sum-Modified Laplacian (NSML)

Laplacian energy could well reflect the edge features of the low-pass subband images, so the NSML is employed in this paper as the external input of PCNN to improve the fusion performance. The definition of NSML is as where means the coefficients of the low-pass subband image at . and denote the sizes of neighbor window which are generally , , and . In this paper, is chosen for computing NSML. In addition, means the weighted template, which emphasizes the coefficient at the center of window.

3.2. Energy of Edge (EOE)

Traditionally, the gray value of the original pixel is utilized as the input of PCNN which does not consider the effect of the neighborhoods. Moreover, it is known that human eyes are more sensitive to edge and directional and texture information rather than single pixel information. In order to make better use of the edge information, the EOE is selected as the external stimulation for PCNN model for high-pass subbands. The definition of EOE is shown as where is the weighted template. denotes the neighborhood of . means the coefficients of the high-pass subband image at . , , and mean the directional filtering operators. The EOE could well reflect the edge information in horizontal, vertical, and diagonal directions.

3.3. Region Energy

The rules of weighted average and larger absolute value are commonly used as the fusion rules to calculate the fused coefficients. However, these methods might lead to losing part of details of the source images and reducing the contrast. To improve the fusion performance, the relationships of the neighboring regions should be considered. In this paper, the rule of max region energy is adopted as the fusion rule. The definition of region energy is shown in (4) and (5): where means the coefficients of the firing map image at . is the energy template. Different energy templates are used for low- and high-pass subbands. is the energy template used for low-pass subband images. And is used for high-pass subbands to strengthen the coefficients at the window center. In addition, the rule of max region energy is defined as (6), where is the fused coefficient and and are the subband images which are obtained by employing NSCT on the source images and . and are the region energy for the corresponding firing maps of and , respectively.

3.4. Improved PCNN

The linking strength , reflecting the variance of the coefficients of the subband images, plays a key role in fusion process. In traditional PCNN based fusion process, is usually assigned manually according to experiences and all neurons in PCNN are set with the same linking strength value. However, according to [17], the values of should not be all the same in different neurons. Besides, it has been verified that is relevant to the image features of the corresponding pixels of the input images. If the external input coefficient is larger, then a higher value should be assigned to . As a result, considering the edge information of the source image, the region average gradient is selected as the linking strength to improve its self-adaptability in this paper. The higher value of the average gradient, the higher clarity of the image. The definition of region average gradient is shown as where means the coefficients of the input image at . is the average gradient at . The larger value of means that the image has a higher clarity. Moreover, the higher the value of is, the earlier the ignition of the correspondent neuron would be which will result in better use of the detailed information of the input image and increase the fusion effectiveness.

4. Results and Discussions

4.1. Experimental Data and Platform

To verify the effectiveness of the proposed fusion framework, we conduct experiments on whole-body PET/CT data which are provided by the General Hospital of Shenyang Military Area Command, Shenyang, China. All of our data is in conformity with laws and ethical standards. Each model of data consists of 239 slices. The original PET and CT images captured from devices are different from each other, such as the images sizes (the sizes of PET and CT images are 128 × 128 and 512 × 512, resp.) and the scale and geometric distortion, which would affect the effectiveness of fusion. Thus, it is necessary to do an interpolation and a registration process before fusion so that all PET images would be resized to 512 × 512 and the geometric displacement would be corrected. In this paper, we use the registration framework that we proposed in [22] to preprocess the PET and CT images before fusion. This framework could achieve a good performance on the registration of whole-body PET and CT images. The experimental platform is Intel® Core TM i7-2600 CPU @ 3.40 GHz, 8 G RAM, 1 T hard disk, Windows 7 OS. The integrated development environment is the MATLAB 2015b. Besides, the ITK-SNAP is utilized to view images in this paper.

4.2. Performance Measures

To quantitatively evaluate fusion performance of the proposed framework and the other compared methods, seven commonly used metrics are applied in this paper including average gradient (AG), Shannon entropy (EN), joint entropy (JE), cross entropy (CE), image quality index (IQI), , and [2, 10, 23].

The average gradient reflects the variance of the gray value which could be used to evaluate the clarity of one image. The higher value of gradient value denotes that the image is clearer and the fusion performance is better. The definition is shown as where means the gray value of the input image at . and mean the size of the input image.

The Shannon entropy measures the information content in the source image. The higher value of Shannon entropy of the fused image, the more information it contained, the better fusion performance it has. The definition of Shannon entropy is shown as where means the probability of the gray level in the source image.

The joint entropy denotes the similarity of the fused image with the source images. The definition is as where means the joint probability of the gray level in image and the gray level in image . The larger joint entropy denotes better performance.

The cross entropy measures the difference between the fused image and the source images. The cross entropy image fusion metric is defined as the average of the relative entropies between the source images and and the fused image ; see (11)where means relative entropy of two probability distribution functions and . and are the normalized histograms of the source images and , respectively. The lower cross entropy means the better fusion performance.

The IQI between the fused image and the source images and are defined as where ,  , and are the mean of , variance of , and the covariance between and . If the IQI achieves higher values, it means the better fusion quality of the fused image.

is an edge-dependent fusion quality metric based on the IQI. The definition of is introduced as where the definition of IQI is given in (13). expresses the contribution of the edge images compared with the original images (usually ). and denote the structural similarity measures between the source images and the fused image in a sliding local window . is defined as , where the overall saliency of a window is defined as . and denote saliency of image and image in window . The larger means better fusion performance.

is a fusion performance metric based on the edge information to measure the similarity among images. The definition of is defined as where and denote the edge strength and orientation preservation values, respectively. The larger value means the better fusion performance.

4.3. Results Evaluation and Discussions

To evaluate the performance of the proposed PET and CT fusion framework, we compare the fusion results of the proposed framework with the following six methods on the same dataset.

Method 1. It is DWT-based method where the fused coefficients are selected according to the rule of maximum absolute values.

Method 2. It is NSCT-based method (NSCT) where the maximum selection rule is used for both the low- and high-pass subbands [24].

Method 3. It is NSCT-PCNN based method (NSCT_PCNN_1) where the same rules with Method 2 are used here.

Method 4. It is image fusion method in the NSCT domain using spatial frequency-based PCNN method (NSCT-SF-PCNN) where the maximum selection rule is used for both the low- and high-pass subbands [25].

Method 5. It is an improved NSCT-PCNN based fusion method [10, 12, 14] named as NSCT_PCNN_2, where the NSML and modified spatial frequency (MSF) are as the external input of PCNN for low- and high-pass subbands, respectively.

Method 6. It is an improved NSCT-PCNN based fusion method [16] named as NSCT_PCNN_3, where EOE is as the external input of the PCNN.

Figure 4 shows three examples of the fusion results obtained by these methods. (a) and (b) are the source CT images and PET images to be fused. (c)–(i) are the fusion results obtained by different compared methods. (c) shows the fusion results based on DWT. (d) presents the fusion results based on NSCT. (e) exhibits the fusion results produced by NSCT_PCNN_1. (f) shows the fusion results produced by NSCT_SF_PCNN. (g) and (h) are the fusion results produced by NSCT_PCNN_2 and NSCT_PCNN_3, respectively. (i) exhibits the fusion results produced by the proposed fusion framework. From the visual analysis of Figure 4, it is observed that the proposed framework could successfully preserve both the feature information of the CT images (e.g., the bony structures) and the PET images (e.g., high metabolic areas). Specifically, the fusion results produced by NSCT have the worst contrast and there are many artifacts introduced in the fusion results produced by DWT. Besides, the rest of the compared methods have similar performance with the proposed framework and it is difficult to determine which one is better by eyes.

Figure 4: Three examples of the fusion results obtained by different methods. (a) CT images. (b) PET images. The (c)–(i) columns are fusion results produced by (c) DWT, (d) NSCT, (e) NSCT_PCNN_1, (f) NSCT_SF_PCNN, (g) NSCT_PCNN_2, (h) NSCT_PCNN_3, and (i) the proposed framework, respectively.

In order to evaluate the fusion results more intuitively, the 3D reconstruction is performed on the source data and the fusion result sequences (see Figure 5). In addition, the pseudo-color coding is also employed on these reconstructions for visualizing the results better. In Figure 5, the first row represents transverse plane of the reconstruction. The second row and third row represent the sagittal plane and the coronal plane, respectively. (a) and (b) are the reconstructions of the source CT and PET image sequences, respectively. (c)–(h) are the reconstructions of the fusion result sequences by DWT, NSCT, NSCT_PCNN_1, NSCT_SF_PCNN, NSCT_PCNN_2, and NSCT_PCNN_3, respectively. (i) is the reconstruction of the fusion results obtained by the proposed framework. From Figure 5, the fusion results obtained by the proposed framework could preserve both the structural feature information of CT and the functional feature information of PET. Besides, there are many artifacts in the results produced by DWT. Similar with Figure 4, the fusion results of NSCT are still the worst with the lowest contrast.

Figure 5: The 3D reconstructions with pseudo-color of the fusion result sequences. The first row represents transverse plane. The second row represents the sagittal plane. The last row represents the coronal plane. (a) and (b) are the 3D reconstructions of the source CT and PET image sequences, respectively. (c)–(i) are the reconstructions of the fusion result sequences that are produced by (c) DWT, (d) NSCT, (e) NSCT_PCNN_1, (f) NSCT_SF_PCNN, (g) NSCT_PCNN_2, (h) NSCT_PCNN_3, and (i) the proposed framework, respectively.

According to Figures 4 and 5, it is difficult to distinguish which method achieves the best performance directly. In order to evaluate the fusion results between different algorithms objectively and quantitatively, this paper adopts seven metrics for performance evaluation which have been introduced in Section 4.2. Table 1 shows the average values of these metrics on 239 pairs of PET and CT images. Figures 6 and 7 are the bar charts corresponding to Table 1. Note that the lower the value of CE, the better the fusion performance. The higher the values of the rest of metrics, the better the fusion performance. It can be seen that the performance of our proposed framework on these metrics is always on the top two among all algorithms which outperforms other methods. Although the DWT performs best on AG, CE, and , its values of EN, HE, and are worse than the others. In addition, the NSCT gives poorer results than other NSCT-PCNN based algorithms because PCNN could make use of the global feature information of the image which would improve the fusion performance. Following the analysis and discussion, a conclusion can be drawn that the proposed fusion framework outperforms the other image fusion methods for the fusion of whole-body PET and CT images. The fusion results of our framework could combine more information which is useful for diagnoses.

Table 1: The average fusion performances of different methods on PET/CT data.
Figure 6: The average values of AG, EN, and JE of the compared methods.
Figure 7: The average values of CE, IQI, , and of the compared methods.

5. Conclusions

In this paper, a novel fusion framework based on adaptive PCNN in NSCT domain is proposed for fusing the whole-body PET and CT images. Our framework utilizes the average gratitude of each pixel as the linking strength in PCNN model to make it more adaptive. Besides, in order to make full use of the feature information contained by the input images, the NSML and EOE are as the linked input of the PCNN model in low-pass and high-pass subbands, respectively, to improve the fusion performance. Moreover, the rule of max region energy is adopted as the fusion rule and different energy templates are used for low- and high-pass subbands. The experiments on whole-body PET and CT images with well alignments demonstrate the good performance of the proposed framework. An evaluation on seven metrics including AG, EN, JE, CE, IQI, , and illustrates objectively that the proposed framework outperforms the other six methods.

Conflicts of Interest

The authors declare that there are no conflicts of interest regarding the publication of this paper.

Acknowledgments

This work was supported by the National Natural Science Foundation of China (no. 61472073, no. 61272176).

References

  1. V. Bhateja, A. Srivastava, A. Moin, and A. Lay-Ekuakille, “NSCT based multispectral medical image fusion model,” in Proceedings of the IEEE International Symposium on Medical Measurements and Applications (MeMeA '16), pp. 1–5, Benevento, Italy, May 2016. View at Publisher · View at Google Scholar
  2. S. Singh, D. Gupta, R. S. Anand, and V. Kumar, “Nonsubsampled shearlet based CT and MR medical image fusion using biologically inspired spiking neural network,” Biomedical Signal Processing and Control, vol. 18, pp. 91–101, 2015. View at Publisher · View at Google Scholar · View at Scopus
  3. X. Xu, D. Shan, G. Wang, and X. Jiang, “Multimodal medical image fusion using PCNN optimized by the QPSO algorithm,” Applied Soft Computing Journal, vol. 46, pp. 588–595, 2016. View at Publisher · View at Google Scholar · View at Scopus
  4. Z. Fu, X. Wang, J. Xu, N. Zhou, and Y. Zhao, “Infrared and visible images fusion based on RPCA and NSCT,” Infrared Physics and Technology, vol. 77, pp. 114–123, 2016. View at Publisher · View at Google Scholar · View at Scopus
  5. Q. Zhang and B.-L. Guo, “Multifocus image fusion using the nonsubsampled contourlet transform,” Signal Processing, vol. 89, no. 7, pp. 1334–1346, 2009. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at Scopus
  6. Y.-Q. Wu, C. Wu, and S.-H. Wu, “Fusion of remote sensing images based on nonsubsampled contourlet transform and region segmentation,” Journal of Shanghai Jiaotong University (Science), vol. 16, no. 6, pp. 722–727, 2011. View at Publisher · View at Google Scholar · View at Scopus
  7. G. Bhatnagar, Q. M. J. Wu, and Z. Liu, “A new contrast based multimodal medical image fusion framework,” Neurocomputing, vol. 157, pp. 143–152, 2015. View at Publisher · View at Google Scholar · View at Scopus
  8. P. Ganasala and V. Kumar, “CT and MR image fusion scheme in nonsubsampled contourlet transform domain,” Journal of Digital Imaging, vol. 27, no. 3, pp. 407–418, 2014. View at Publisher · View at Google Scholar · View at Scopus
  9. N. Amini, E. Fatemizadeh, and H. Behnam, “MRI-PET image fusion based on NSCT transform using local energy and local variance fusion rules,” Journal of Medical Engineering and Technology, vol. 38, no. 4, pp. 211–219, 2014. View at Publisher · View at Google Scholar · View at Scopus
  10. Z. Y. Jin and Y. J. Wang, “Multi-modality medical image fusion method based on non-subsampled contourlet transform,” Chinese Journal of Medical Physics, vol. 33, no. 5, pp. 445–450, 2016. View at Google Scholar
  11. S. Das and M. K. Kundu, “NSCT-based multimodal medical image fusion using pulse-coupled neural network and modified spatial frequency,” Medical and Biological Engineering and Computing, vol. 50, no. 10, pp. 1105–1114, 2012. View at Publisher · View at Google Scholar · View at Scopus
  12. T. Z. Xiang, R. R. Gao, and L. Yan, “Multi-Scale fusion of infrared and visible images based on PCNN,” Journal of Geomatics Science and Technology, vol. 3, pp. 273–278, 2016. View at Google Scholar
  13. M.-L. Li, Y.-J. Li, H.-M. Wang, and K. Zhang, “Fusion algorithm of infrared and visible images based on NSCT and PCNN,” Opto-Electronic Engineering, vol. 37, no. 6, pp. 90–95, 2010. View at Publisher · View at Google Scholar · View at Scopus
  14. T. Xiang, L. Yan, and R. Gao, “A fusion algorithm for infrared and visible images based on adaptive dual-channel unit-linking PCNN in NSCT domain,” Infrared Physics and Technology, vol. 69, pp. 53–61, 2015. View at Publisher · View at Google Scholar · View at Scopus
  15. L. Yan and T.-Z. Xiang, “Fusion of infrared and visible images based on edge feature and adaptive PCNN in NSCT domain,” Acta Electronica Sinica, vol. 44, no. 4, pp. 761–766, 2016. View at Publisher · View at Google Scholar · View at Scopus
  16. J. Wang, Q. Li, Z. Jia, N. Kasabov, and J. Yang, “A novel multi-focus image fusion method using PCNN in nonsubsampled contourlet transform domain,” Optik, vol. 126, no. 20, pp. 2508–2511, 2015. View at Publisher · View at Google Scholar · View at Scopus
  17. Y. Chai, H. F. Li, and M. Y. Guo, “Multifocus image fusion scheme based on features of multiscale products and PCNN in lifting stationary wavelet domain,” Optics Communications, vol. 284, no. 5, pp. 1146–1158, 2011. View at Google Scholar
  18. A. L. da Cunha, J. Zhou, and M. N. Do, “The nonsubsampled contourlet transform: theory, design, and applications,” IEEE Transactions on Image Processing, vol. 15, no. 10, pp. 3089–3101, 2006. View at Publisher · View at Google Scholar · View at Scopus
  19. M. N. Do and M. Vetterli, “The contourlet transform: an efficient directional multiresolution image representation,” IEEE Transactions on Image Processing, vol. 14, no. 12, pp. 2091–2106, 2005. View at Publisher · View at Google Scholar · View at Scopus
  20. Z. Wang, Y. Ma, F. Cheng, and L. Yang, “Review of pulse-coupled neural networks,” Image and Vision Computing, vol. 28, no. 1, pp. 5–13, 2010. View at Publisher · View at Google Scholar · View at Scopus
  21. C. Xi, Z. Wei, and G. Fei, “Image fusion algorithm based on adaptive pulse coupled neural networks in curvelet domain,” in Proceedings of the IEEE 10th International Conference on Signal Processing (ICSP '10), pp. 845–848, October 2010. View at Publisher · View at Google Scholar · View at Scopus
  22. Z. Song, H. Jiang, Q. Yang, Z. Wang, and G. Zhang, “A registration method based on contour point cloud for 3D whole-body PET and CT images,” BioMed Research International, vol. 2017, Article ID 5380742, 11 pages, 2017. View at Publisher · View at Google Scholar
  23. M. B. A. Haghighat, A. Aghagolzadeh, and H. Seyedarabi, “A non-reference image fusion metric based on mutual information of image features,” Computers and Electrical Engineering, vol. 37, no. 5, pp. 744–756, 2011. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at Scopus
  24. S. Li, B. Yang, and J. Hu, “Performance comparison of different multi-resolution transforms for image fusion,” Information Fusion, vol. 12, no. 2, pp. 74–84, 2011. View at Publisher · View at Google Scholar · View at Scopus
  25. X.-B. Qu, J.-W. Yan, H.-Z. Xiao, and Z.-Q. Zhu, “Image fusion algorithm based on spatial frequency-motivated pulse coupled neural networks in nonsubsampled contourlet transform domain,” Acta Automatica Sinica, vol. 34, no. 12, pp. 1508–1514, 2008. View at Publisher · View at Google Scholar · View at Scopus