Research Article  Open Access
Chenhui Qiu, Yuanyuan Wang, Huan Zhang, Shunren Xia, "Image Fusion of CT and MR with Sparse Representation in NSST Domain", Computational and Mathematical Methods in Medicine, vol. 2017, Article ID 9308745, 13 pages, 2017. https://doi.org/10.1155/2017/9308745
Image Fusion of CT and MR with Sparse Representation in NSST Domain
Abstract
Multimodal image fusion techniques can integrate the information from different medical images to get an informative image that is more suitable for joint diagnosis, preoperative planning, intraoperative guidance, and interventional treatment. Fusing images of CT and different MR modalities are studied in this paper. Firstly, the CT and MR images are both transformed to nonsubsampled shearlet transform (NSST) domain. So the lowfrequency components and highfrequency components are obtained. Then the highfrequency components are merged using the absolutemaximum rule, while the lowfrequency components are merged by a sparse representation (SR) based approach. And the dynamic group sparsity recovery (DGSR) algorithm is proposed to improve the performance of the SRbased approach. Finally, the fused image is obtained by performing the inverse NSST on the merged components. The proposed fusion method is tested on a number of clinical CT and MR images and compared with several popular image fusion methods. The experimental results demonstrate that the proposed fusion method can provide better fusion results in terms of subjective quality and objective evaluation.
1. Introduction
Image fusion techniques combine multiple input images of the same scene or object to get a composite image that is expected to be more informative and suitable for human visual perception or further image processing compared to any of the input images [1–3]. Multimodal medical image fusion, one of the major image fusion applications, draws increasing attention from medical academia [4, 5] and has been used for joint diagnosis, preoperative planning, intraoperative guidance, interventional treatment, and so on [6, 7]. CT and MRI techniques have been widely used in clinical diagnosis and treatment; nevertheless the accurate differences between lesion, bone, and peritumoral soft tissue sometimes are hardly obtained by the single CT or MR images [7]. Although CT images provide electron density map required for accurate radiation dose estimation and superior cortical bone contrast, they are limited in soft tissue contrast. MRI provides good soft tissue contrast that permits better visualization of tumors or tissue abnormalities. However, it lacks signal from cortical bone and has image intensity values that have no relation to electron density [8]. Hence, for precise joint diagnoses and effective treatment procedures, the superior method for fusing CT and MR images should be explored urgently.
Multiscale transform (MST) theories are the most widely used tools in various image fusion situations [2, 3]. Classical MST tools include pyramid ones such as Laplacian pyramid (LP) [9] and gradient pyramid (GP) [10] and wavelet ones such as discrete wavelet transform (DWT) [11] and dualtree complex wavelet transform (DTCWT) [12]. Since the pyramid and wavelet tools are originally proposed to process onedimensional signals, even their improved versions cannot capture geometric structure information of the image very well. Current MST tools are multiscale geometric analysis (MGA) ones. MGA tools are able to extract geometric structure information including edges, curves, and textures of images much more effectively, since they are the “true” 2D transform. Curvelet transform (CVT) [13], contourlet transform (COT) [14], nonsubsampled contourlet transform (NSCT) [15], shearlet transform (ST) [16], and nonsubsampled shearlet transform (NSST) [17] are the most representative tools of MGA. CVT is proposed to represent a curve as the superposition of bases of various lengths and widths obeying a scaling law. NSCT is the shiftinvariant version of COT essentially, which is built upon nonsubsampled multiscale pyramids and nonsubsampled directional filter banks. ST is much more efficient than COT in computational consumption. Besides, there are no restrictions on the number of directions and the size of support for the shearing [17]. NSST is the shiftinvariant version of ST essentially.
The geometric structure information in different scales of the image can be extracted effectively by these MGA tools. The tools above are widely studied to fuse general multimodal images [8, 18–21].
Sparse representation (SR) is another popular and powerful theory used to solve the image fusion problem. SR addresses the signals’ natural sparsity by simulating the sparse coding mechanism of human vision systems [22]. Yang and Li first apply SR theory to multifocus image fusion, and they use orthogonal matching pursuit (OMP) to implement sparse coding [23]. They also study the influence of dictionary and simultaneous OMP algorithm on the fusion of multimodal images [24].
Although MSTbased and SRbased image fusion methods have achieved giant success, they have their special defects, respectively. We will discuss them in Section 4. In order to combine the advantages of MST and SR and avoid their defects, Wang et al. propose a fusion method that combines NSCT and SR, and OMP algorithm is used in their method to calculate the SR coefficients of the lowfrequency components [25]. A general framework for image fusion combining MST (NSST is not included) and SR is proposed and discussed in detail by Liu et al. in [26]. They make comparisons and find that the method with combination of NSCT and SR which utilizes OMP algorithm for sparse coding obtains excellent fusion performance.
In this paper, we propose a novel fusion method to fuse clinical CT and MR images by combining NSST and SR. Firstly, the source images are both transformed to NSST domain. Then the highfrequency components are merged by the absolutemaximum rule, while the lowfrequency components are merged with the SRbased approach. Wang et al. and Liu et al. utilize OMP algorithm to implement sparse coding, while the dynamic group sparsity recovery (DGSR) algorithm is first explored in our proposed method to improve the performance of SRbased approach. Finally, the fused image is obtained by performing the inverse NSST on the merged lowfrequency and highfrequency components.
Our proposed fusion method is tested on forty pairs of clinical brain CT and MR anatomical images and compared with seven popular image fusion methods. MR imaging techniques can provide four kinds of anatomical images which are all considered in our research. The experimental results demonstrate that the proposed fusion method can provide superior fusion results in terms of subjective quality and objective evaluation.
The remainder of this paper is organized as follows. The related work including NSST and SR is presented in Section 2. The proposed fusion method is described in Section 3. Experiments and discussions are shown in Section 4. Finally, we draw the conclusions and depict some future work in Section 5.
2. Related Work
2.1. Nonsubsampled Shearlet Transform (NSST)
Shearlet has been regarded as the best sparse directional image representation frame among MST theories so far [5]. In dimension , the affine systems with composite dilations are described as follows:where , refers to anisotropy matrix which is associated with scale transformations, and refers to shear matrix which is associated with areapreserving geometrical transformations, such as rotations and shear. , , and are scale, direction, and shift parameter, respectively. For each and , the matrices of and which play vitally important roles in the process of ST are given as follows [17]:Assume that and ; then we obtainLet and , for any and ; also let and be given bywhere , , and .
Each element of is supported on a pair of trapezoids, of approximate size , oriented along lines of slope .
Then the ST function is obtained:where , , and .
ST has the following properties: well spatial and frequent localization, strong selectivity of anisotropic directionality, well parabolic scaling, and approximately sparse representation. However, ST engenders the Gibbs phenomena because of the lack of shift invariance. NSST is the shiftinvariant version of ST. It utilizes nonsubsampled Laplacian pyramid (NSLP) filters as the substitutes for the LP filters which are used in the ST mechanism. Multiscale decomposition is implemented by NSLP. One lowfrequency component and one highfrequency component can be produced at each NSLP decomposition level. And the next NSLP decomposition is performed on the last (previous) lowfrequency component to capture the singularities of the source image iteratively. When the decomposition level is set to , the source image is decomposed into components with the same size of the source image, in which one is the lowfrequency component.
Shearing filter (SF) is performed on highfrequency components of each NSLP decomposition level without subsampling to achieve the multidirectional factorization, which ensures the shift invariance property of NSST. Assume that SF performs the directional decomposition with stages on a highfrequency component decomposed by NSLP; then directional subbands with the same size as the source image are produced [17]. Twolevel decomposition of the NSST is shown in Figure 1. This schematic diagram illustrates the NLSP decomposition and the corresponding directional decomposition by SF.
2.2. Sparse Representation (SR)
SR theory is based on the assumption that the natural signals can be represented or approximately represented by a linear combination of a “few” atoms from dictionary [22]. For signals , SR theory indicates the existence of a dictionary , which contains prototype signals that are referred to atoms. Then, for any signal , there is a linear combination of atoms from which can approximate to very well. The signal can be expressed as , where is the unknown sparse coefficient vector. It usually assumes that , which implies that is the overcompleted and redundant dictionary. So there are numerous feasible solutions of the underdetermined system . Finding the sparsest that contains the fewest nonzero entries involves solving the following optimization problem:where denotes the number of nonzero entries in ; is an error tolerance.
The above optimization is an NPhard problem. The main idea to solve this problem is obtaining an approximate solution instead of the sparsest one. OMP is the representative algorithm to get the approximate solution.
3. Proposed Fusion Method
The proposed fusion method of CT and MR images mainly contains four phases: NSST decomposition, highfrequency fusion, lowfrequency fusion, and NSST reconstruction. The schematic diagram of the proposed method is shown in Figure 2.
3.1. NSST Decomposition
NSST decomposition is performed on CT and MR images to obtain highfrequency components that are denoted as and lowfrequency components that are denoted as .
3.2. HighFrequency Fusion
The highfrequency components at different levels contain much salient information of regional boundary, edge, and line features. The larger absolute values of the coefficients in each of the highfrequency components indicate more important salient information. Therefore, the popular absolutemaximum rule is appropriate to merge the highfrequency components . The coefficients of highfrequency components and and the fused image are denoted as and and , respectively. The rule is expressed as follows:
3.3. LowFrequency Fusion
and are merged with SRbased fusion approach in our research. Image fusion methods always depend on the local information of source images. Besides, the stability and efficiency of sparse coding should be taken into account. Hence, the source images need to be divided into image patches.
The sliding window technique is adopted to divide and into image patches of size from upper left to lower with a suitable step length, respectively [23]. Then all patches of are transformed into vectors via lexicographic ordering and normalization operation. And all these vectors constitute the matrix , in which each column corresponds to one patch of . The size of is . can be obtained in the same way.
For the th column vector of , its sparse representation is calculated by the DGSR algorithm. According to (6), the sparse coding vector for is obtained as follows:
Similarly, we can get the sparse coding vector for which is the th column vector of .where is the offline learned dictionary that is trained by means singular value decomposition (SVD) algorithm [27].
Then and are merged by maximumnorm 1 rule to produce the fused sparse vectors . The fusion rule is expressed as follows:
The th column vector of the matrix is calculated by
Then can be constituted by all these column vectors.
Finally, the lowfrequency component of the fused image is reconstructed by . This reconstruction is the inverse process of normalization and sliding window technique.
In the fusion application of CT and MR images, the nonzero sparse coefficients of source images are not randomly distributed but are groupclustered. They tend to be clustered into groups, although the clustering group structures are dynamic and unpredictable. Therefore, DGSR algorithm that prunes data residues in the iterative process according to both sparsity and the group clustering trend is appropriate for our research. The experimental results in [28] demonstrate that the DGSR algorithm achieves better performance than OMP algorithm. In this paper, the main work is to explore the image fusion method for CT and MR; the details of DGSR algorithm can be found in [28]. The MATLAB codes of DGSR algorithm and relevant experiments can be downloaded at http://ranger.uta.edu/~huang/Downloads.htm.
3.4. NSST Reconstruction
The inverse NSST is performed on both low frequency and high frequency to reconstruct the final fused image .
4. Experiments and Discussions
4.1. CT and MR Images
CT imaging techniques provide anatomical images of tissues or lesions. MRT1W, MRT2W, and MRPDW are derived from different settings of longitudinal relaxation time T1, transversal relaxation time T2, echo pulse sequence, and proton density of tissues. T1W means that the contrasts of tissues in the MR image mainly depend on T1 weight. T2W and PDW are of the similar meaning. MRCE image is acquired by adding gadoliniumcontaining contrast medium, which is mainly used for the definite diagnosis of cerebral tumor. They are all MR anatomical images.
Because of the limited space of paper, only one representative group of clinical brain CT and MR images are presented as the source images which are shown in Figure 3. The images of Figure 3 are the 12th slices of a Ewing’s sarcoma patient at the age of 22. He had a right homonomous hemianopia, a left inferior quadrantanopia, right lower extremity hyperreflexia, and right extensor plantar response. The images are of the same slice of the same patient so that we can find the difference of different imaging modalities obviously. They can be downloaded from http://www.med.harvard.edu/aanlib/.
(a)
(b)
(c)
(d)
(e)
4.2. Experimental Setting
The clinical brain CT and MR images above have the same size, , and they are preregistered in pairs. All the experiments are implemented in MATLAB 7.14 and on a PC with 2.66GB CPU, 4 GB RAM, and 32bit OS.
To assess the performance of the proposed method, the GP [10], DTCWT [12], CVT [18], NSCT [19], NSST [20], SROMP [23], and NSCTSROMP [25, 26] based fusion methods are implemented for comparison.
The fusion rules for lowfrequency components of these MSTbased methods are all chosen as “averaging” rule. The fusion rules for highfrequency components of these MSTbased methods are all chosen as “maxabsolute” rule [2]. For SROMP based method, the fusion rule is chosen as “maxnorm 1” rule [23]. For NSCTSROMP based method, the fusion rules for lowfrequency and highfrequency components are the same as the proposed method’s fusion rules [25, 26]. For these MSTbased methods, NSCTSROMP based method, and the proposed method, the decomposition levels are all set to 4, which are most appropriate according to our experiments and the analysis in [2, 26]. For DTCWTbased method, the filters of first level and other levels are selected as “53” and “q_a,” respectively. For NSCTbased and NSCTSROMP based methods, the pyramidal filters and directional filters are selected as “pyrexc” and “79,” respectively. And the directional numbers form coarser to finer scales are chosen as , , , and , respectively. For NSSTbased method and the proposed method, NSLP filters are selected as “pyrexc.” The directional numbers are the same as aforementioned. And the Meyer wavelet window is used for SF. The sizes of SFs which are ordered from coarser to finer scales are chosen as 32, 32, 16, and 16, respectively [17]. For SROMP and NSCTSROMP based methods and the proposed method, the size of image patches is set to , the step length is set to 2, the error tolerance is set to 0.1, and the dictionary size is set to [2, 23].
4.3. Objective Evaluation Metrics
Generally, image fusion results can be evaluated in subjective and objective ways. Subjective ways are difficult to execute because they are based on psychovisual testing and cannot be applied in an automatic system. Besides, there is little visual difference among the fusion results in most cases, so subjective ways are difficult to correctly evaluate the fusion results. Therefore, many objective evaluation metrics have been developed [2, 3]. In this paper, standard deviation (STD) [29], mean structural similarity (MSSIM) [30], mutual information (MI) [31], sum of the correlations of differences (SCD) [32], [33], [34], [34], [35], and visual information fidelity for fusion (VIFF) [36] are used to evaluate the fusion results objectively. The larger values of these metrics imply the better fusion results. It is worthwhile to note that if the fused images are used for preoperative planning, intraoperative guidance, and interventional treatment, which depend on further image processing, the superiority of objective evaluation is important and valuable. We also give simply subjective assessments on the fusion results.
4.4. Results and Discussions
Figure 4 shows the fusion results of Figures 3(a) and 3(b) with different fusion methods. In the right halves of Figure 3(b) and Figures 4(a)–4(h), there are two lesions labeled by the arrows. Actually they are the one lesion. And the small one is the diffusion of the large one. The small lesion in right half and the lesion in left half of Figures 4(a), 4(b), 4(c), and 4(f) is blurred to some extent. Although Figures 4(d) and 4(e) exhibit the small lesion in right half well, they cannot exhibit the lesion in left half with good visual quality. The cerebral bony structure is the basic sclerous tissue in brain anatomical image. Only Figures 4(g) and 4(h) show the cerebral bony structure as well as the expected effect. On the whole, the visual information of tissues and lesions of CT and MRT1W images is most perfectly preserved in Figures 4(g) and 4(h). Besides, the CVT and NSCTbased methods produce artifacts along the cerebral bony structure. Although SROMP based method can capture the salient information, it cannot extract the blurry features (such as the lesion in left half) effectively.
(a)
(b)
(c)
(d)
(e)
(f)
(g)
(h)
Table 1 gives the objective evaluation metrics of the fusion results of Figures 3(a) and 3(b) with different fusion methods. The best result for each metric is labeled in bold. The second best result for each metric is labeled with italic font. The proposed method provides the best results for six metrics and the second best result for one metric. Although the fusion results of NSCTSROMP based method and the proposed method are very similar in subjective quality, the proposed method is superior to NSCTSROMP based method in all of the objective evaluation metrics. In addition, the computational efficiency of the proposed method is much higher than that of NSCTSROMP based method because the decomposition and reconstruction of NSST are much faster than those of NSCT [17].

The zoomedin version of the bottomright part of Figures 4(g) and 4(h) is shown in Figure 5. We can find that the contrast of the larger lesion and its limbic region in Figure 5(b) is higher than the corresponding part in Figure 5(a). This verdict coincides with the STD metric evaluation. Besides, the visual quality of Figure 5(b) is higher than that of Figure 5(a), which is consistent with the VIFF metric. (Note that medical display system provides better visual effect than common display system; it distinguishes the difference between Figures 5(a) and 5(b) more clearly.)
(a)
(b)
Figure 6 shows the fusion results of Figures 3(a) and 3(c) with different fusion methods. The subjective analysis of Figure 6 is similar to the subjective assessment of Figure 4. The proposed method provides the best visual quality of the lesion labeled by the arrow, sulcus, and bony structure. NSCTSROMP based method provides similar but little worse subjective visual quality with careful inspection.
(a)
(b)
(c)
(d)
(e)
(f)
(g)
(h)
Table 2 gives the objective evaluation metrics of the fusion results of Figures 3(a) and 3(c) with different fusion methods. The proposed method provides the best results for six metrics and the second best result for one metric.

Figures 7 and 8 show the fusion results of Figures 3(a) and 3(d) and Figures 3(a) and 3(e) with different fusion methods, respectively. The proposed method provides the best visual fusion performances in the two cases.
(a)
(b)
(c)
(d)
(e)
(f)
(g)
(h)
(a)
(b)
(c)
(d)
(e)
(f)
(g)
(h)
Tables 3 and 4 give the objective evaluation metrics of the fusion results of Figures 3(a) and 3(d) and Figures 3(a) and 3(e) with different fusion methods, respectively. The proposed method provides the best results for six metrics and the second best result for one metric in Table 3. It also provides the best results for five metrics and the second best result for one metric in Table 4.


Table 5 gives the objective evaluation metrics of the fusion results of forty pairs of clinical brain CT and different MR modalities with different fusion methods. The means and the standard deviations of different metrics of different methods are shown in Table 5. We can find that the proposed method provides the best results for six metrics.

NSST is the most representative MST tool, which can extract the geometric structure information of source images more effectively than NSCT, CVT, and classical MST tools. The salient information such as edges, curves, and textures can be captured by NSST very well and can be preserved in its highfrequency components. Therefore, NSSTbased method obtains the best results of or metrics in some cases. However, NSST cannot express the lowfrequency components sparsely. If ordinary fusion rule is utilized to merge the lowfrequency components, the fusion performance may be degraded as the lowfrequency components contain a lot of energy of source images.
SR theory can represent an image in the approximately sparsest sense. It is the best approach of twodimensional signal in terms of information theory, so SRbased method obtains the best results of MI metric form Tables 1–5. While SRbased fusion approach can represent the underlying salient information of source images very efficiently, it is difficult to reconstruct the smallscale details of source images and the blurred features.
The proposed method combines NSST and SR in order to preserve their advantages and avoid their defects. DGSR algorithm is first explored in the proposed method to ensure the satisfactory performance when SRbased fusion approach is implemented on lowfrequency components. It is worthwhile to note that NSST plays a more important role than SR in the proposed fusion method, since the fusion framework is mainly based on NSST. The geometric structure information including edges, curves, and textures of images is preserved in highfrequency components of NSST, while a minority of salient features and a majority of energy are preserved in the lowfrequency components of NSST. The SRbased approach is used to fuse the lowfrequency components and DGSR algorithm is explored to calculate sparse representation coefficients of the lowfrequency components. We can find that the proposed method has the superiority in STD, MSSIM, SCD, , , and VIFF metrics from the tables.
5. Conclusions and Future Work
In this paper, we have proposed a novel image fusion method for fusing CT and MR images, which combines the advantages of NSST and SR. Moreover, DGSR algorithm is utilized when the SRbased fusion approach is performed on the lowfrequency components of NSST decompositions. To assess the performance of the proposed method, the GP, DTCWT, CVT, NSCT, NSST, SROMP, and NSCTSROMP based fusion methods are implemented for comparison. The experimental results show that the proposed method provides superior fusion results in terms of subjective quality and objective evaluation metrics.
However, fusing MR and lowdose CT images is not considered in our paper because the lowdose CT image contains a lot of noise, which is complicated and cannot be modeled as Gaussian distribution or Rician distribution. We may concentrate on this fusion topic in the future.
Besides, the automatic or guided target delineation in radiotherapy planning which is based on the fused CT and MR images will be studied in the future.
Conflicts of Interest
The authors declare that they have no conflicts of interest.
Acknowledgments
The authors would like to sincerely thank Dr. Longyan Ni and Dr. Hongxing Liu from Nanjing Brain Hospital affiliated to Nanjing Medical University (China), Professor Shutao Li from Hunan University (China), and Dr. Haitao Yin from Nanjing University of Posts and Telecommunications (China). This work is supported by National Key Research and Development Program of China (no. 2016YFC1306600).
References
 A. A. Goshtasby and S. Nikolov, “Image fusion: advances in the state of the art,” Information Fusion, vol. 8, no. 2, pp. 114–118, 2007. View at: Publisher Site  Google Scholar
 S. Li, B. Yang, and J. Hu, “Performance comparison of different multiresolution transforms for image fusion,” Information Fusion, vol. 12, no. 2, pp. 74–84, 2011. View at: Publisher Site  Google Scholar
 S. Li, X. Kang, L. Fang, J. Hu, and H. Yin, “Pixellevel image fusion: a survey of the state of the art,” Information Fusion, vol. 33, pp. 100–112, 2017. View at: Publisher Site  Google Scholar
 A. P. James and B. V. Dasarathy, “Medical image fusion: a survey of the state of the art,” Information Fusion, vol. 19, no. 1, pp. 4–19, 2014. View at: Publisher Site  Google Scholar
 J. Du, W. Li, K. Lu, and B. Xiao, “An overview of multimodal medical image fusion,” Neurocomputing, vol. 215, pp. 3–20, 2016. View at: Publisher Site  Google Scholar
 G. J. Webster, J. E. Kilgallon, K. F. Ho, C. G. Rowbottom, N. J. Slevin, and R. I. Mackay, “A novel imaging technique for fusion of highquality immobilised MR images of the head and neck with CT scans for radiotherapy target delineation,” British Journal of Radiology, vol. 82, no. 978, pp. 497–503, 2009. View at: Publisher Site  Google Scholar
 J. Dai, X. Wang, Y. Dong, H. Yu, D. Yang, and G. Shen, “Two and threedimensional models for the visualization of jaw tumors based on CTMRI image fusion,” The Journal of Craniofacial Surgery, vol. 23, no. 2, pp. 502–508, 2012. View at: Publisher Site  Google Scholar
 P. Ganasala and V. Kumar, “CT and MR image fusion scheme in nonsubsampled contourlet transform domain,” Journal of Digital Imaging, vol. 27, no. 3, pp. 407–418, 2014. View at: Publisher Site  Google Scholar
 P. J. Burt and E. H. Adelson, “The laplacian pyramid as a compact image code,” IEEE Transactions on Communications, vol. 31, no. 4, pp. 532–540, 1983. View at: Publisher Site  Google Scholar
 V. S. Petrović and C. S. Xydeas, “Gradientbased multiresolution image fusion,” IEEE Transactions on Image Processing, vol. 13, no. 2, pp. 228–237, 2004. View at: Publisher Site  Google Scholar
 H. Li, B. S. Manjunath, and S. K. Mitra, “Multisensor image fusion using the wavelet transform,” CVGIP: Graphical Models and Image Processing, vol. 57, no. 3, pp. 235–245, 1995. View at: Publisher Site  Google Scholar
 J. J. Lewis, R. J. O'Callaghan, S. G. Nikolov, D. R. Bull, and N. Canagarajah, “Pixel and regionbased image fusion with complex wavelets,” Information Fusion, vol. 8, no. 2, pp. 119–130, 2007. View at: Publisher Site  Google Scholar
 E. J. Candès and D. L. Donoho, “Curvelets and curvilinear integrals,” Journal of Approximation Theory, vol. 113, no. 1, pp. 59–90, 2001. View at: Publisher Site  Google Scholar  MathSciNet
 M. N. Do and M. Vetterli, “The contourlet transform: an efficient directional multiresolution image representation,” IEEE Transactions on Image Processing, vol. 14, no. 12, pp. 2091–2106, 2005. View at: Publisher Site  Google Scholar
 A. L. da Cunha, J. Zhou, and M. N. Do, “The nonsubsampled contourlet transform: theory, design, and applications,” IEEE Transactions on Image Processing, vol. 15, no. 10, pp. 3089–3101, 2006. View at: Publisher Site  Google Scholar
 D. Labate, W.Q. Lim, G. Kutyniok, and G. Weiss, “Sparse multidimensional representation using shearlets,” in Proceedings of the Wavelets XI, pp. 1–9, usa, August 2005. View at: Publisher Site  Google Scholar
 G. Easley, D. Labate, and W.Q. Lim, “Sparse directional image representations using the discrete shearlet transform,” Applied and Computational Harmonic Analysis , vol. 25, no. 1, pp. 25–46, 2008. View at: Publisher Site  Google Scholar  MathSciNet
 S. Mehta and B. Marakarkandy, “CT and MRI image fusion using curvelet transform,” Journal of Information, Knowledge and Research in Electronics and Communication Engineering, vol. 2, no. 2, pp. 848–852, 2013. View at: Google Scholar
 G. Bhatnagar, Q. M. J. Wu, and Z. Liu, “Directive contrast based multimodal medical image fusion in NSCT domain,” IEEE Transactions on Multimedia, vol. 15, no. 5, pp. 1014–1024, 2013. View at: Publisher Site  Google Scholar
 S. Singh, D. Gupta, R. S. Anand, and V. Kumar, “Nonsubsampled shearlet based CT and MR medical image fusion using biologically inspired spiking neural network,” Biomedical Signal Processing and Control, vol. 18, pp. 91–101, 2015. View at: Publisher Site  Google Scholar
 X. Liu, W. Mei, H. Du, and J. Bei, “A novel image fusion algorithm based on nonsubsampled shearlet transform and morphological component analysis,” Signal, Image and Video Processing, vol. 10, no. 5, pp. 959–966, 2016. View at: Publisher Site  Google Scholar
 B. A. Olshausen and D. J. Field, “Emergence of simplecell receptive field properties by learning a sparse code for natural images,” Nature, vol. 381, no. 6583, pp. 607–609, 1996. View at: Publisher Site  Google Scholar
 B. Yang and S. Li, “Multifocus image fusion and restoration with sparse representation,” IEEE Transactions on Instrumentation and Measurement, vol. 59, no. 4, pp. 884–892, 2010. View at: Publisher Site  Google Scholar
 B. Yang and S. Li, “Pixellevel image fusion with simultaneous orthogonal matching pursuit,” Information Fusion, vol. 13, no. 1, pp. 10–19, 2012. View at: Publisher Site  Google Scholar
 J. Wang, J. Peng, X. Feng, G. He, J. Wu, and K. Yan, “Image fusion with nonsubsampled contourlet transform and sparse representation,” Journal of Electronic Imaging, vol. 22, article 043019, no. 4, 2013. View at: Publisher Site  Google Scholar
 Y. Liu, S. Liu, and Z. Wang, “A general framework for image fusion based on multiscale transform and sparse representation,” Information Fusion, vol. 24, pp. 147–164, 2015. View at: Publisher Site  Google Scholar
 M. Aharon, M. Elad, and A. Bruckstein, “KSVD: an algorithm for designing overcomplete dictionaries for sparse representation,” IEEE Transactions on Signal Processing, vol. 54, no. 11, pp. 4311–4322, 2006. View at: Publisher Site  Google Scholar
 J. Huang, X. Huang, and D. Metaxas, “Learning with dynamic group sparsity,” in Proceedings of the 12th International Conference on Computer Vision, ICCV 2009, pp. 64–71, jpn, October 2009. View at: Publisher Site  Google Scholar
 M. Deshmukh and U. Bhosale, “Image fusion and image quality assessment of fused images,” International Journal of Image Processing, vol. 4, no. 5, pp. 484–508, 2010. View at: Google Scholar
 Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, “Image quality assessment: from error visibility to structural similarity,” IEEE Transactions on Image Processing, vol. 13, no. 4, pp. 600–612, 2004. View at: Publisher Site  Google Scholar
 G. H. Qu, D. L. Zhang, and P. F. Yan, “Information measure for performance of image fusion,” IEEE Electronics Letters, vol. 38, no. 7, pp. 313–315, 2002. View at: Publisher Site  Google Scholar
 V. Aslantas and E. Bendes, “A new image quality metric for image fusion: the sum of the correlations of differences,” AEÜ  International Journal of Electronics and Communications, vol. 69, no. 12, pp. 1890–1896, 2015. View at: Publisher Site  Google Scholar
 Z. Wang and A. C. Bovik, “A universal image quality index,” IEEE Signal Processing Letters, vol. 9, no. 3, pp. 81–84, 2002. View at: Publisher Site  Google Scholar
 G. Piella and H. Heijmans, “A new quality metric for image fusion,” in Proceedings of the International Conference on Image Processing (ICIP '03), vol. 3, pp. 173–176, September 2003. View at: Publisher Site  Google Scholar
 C. S. Xydeas and V. Petrović, “Objective image fusion performance measure,” IEEE Electronics Letters, vol. 36, no. 4, pp. 308309, 2000. View at: Publisher Site  Google Scholar
 Y. Han, Y. Cai, Y. Cao, and X. Xu, “A new image fusion performance metric based on visual information fidelity,” Information Fusion, vol. 14, no. 2, pp. 127–135, 2013. View at: Publisher Site  Google Scholar
Copyright
Copyright © 2017 Chenhui Qiu et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.