Abstract

Wavelet transforms have emerged as a powerful tool in image fusion. However, the study and analysis of medical image fusion is still a challenging area of research. Therefore, in this paper, we propose a multiscale fusion of multimodal medical images in wavelet domain. Fusion of medical images has been performed at multiple scales varying from minimum to maximum level using maximum selection rule which provides more flexibility and choice to select the relevant fused images. The experimental analysis of the proposed method has been performed with several sets of medical images. Fusion results have been evaluated subjectively and objectively with existing state-of-the-art fusion methods which include several pyramid- and wavelet-transform-based fusion methods and principal component analysis (PCA) fusion method. The comparative analysis of the fusion results has been performed with edge strength (), mutual information (MI), entropy (), standard deviation (SD), blind structural similarity index metric (BSSIM), spatial frequency (SF), and average gradient (AG) metrics. The combined subjective and objective evaluations of the proposed fusion method at multiple scales showed the effectiveness and goodness of the proposed approach.

1. Introduction

The development of multimodality medical imaging sensors for extracting clinical information has influenced to explore the possibility of data reduction and having better visual representation. X-ray, ultrasound, magnetic resonance imaging (MRI), and computed tomography (CT) are a few examples of biomedical sensors. These sensors are used for extracting clinical information, which is generally complementary in nature. For example, X-ray is widely used in detecting fractures and abnormalities in bone position, CT is used in tumor and anatomical detection, and MRI is used to obtain information about tissues. Thus, none of these modalities is able to carry all complementary and relevant information in a single image. Medical image fusion [1, 2] is the only possible way to combine and merge all relevant and complementary information from multiple source images into single composite image which facilitates more precise diagnosis and better treatment.

The basic requirements for image fusion [3] are as follows: first, fused image should possess all possible relevant information contained in the source images; second, fusion process should not introduce any artifact or unexpected feature in the fused image.

Image fusion can be classified into three categories: pixel level fusion, feature level fusion, and decision or symbol level fusion [4]. Pixel level fusion [5] deals with information associated with each pixel and fused image can be obtained from the corresponding pixel values of source images. In feature level fusion [6], source images are segmented into regions and features like pixel intensities, edges, and textures, are used for fusion. Decision or symbol level fusion [7] is a high-level fusion which is based on statistics, voting, fuzzy logic, prediction and heuristics, and so forth. For present work, we have considered pixel level image fusion due to its simple computation and understanding.

Spatial and transform domains [8] are the two fundamental approaches for image fusion. In spatial domain fusion, the fusion rule is directly applied to the intensity values of the source images. Averaging, weighted averaging, principal component analysis (PCA) [9], linear fusion [10], and sharp fusion [11] are a few examples of spatial domain fusion scheme. One of the major disadvantages of spatial domain fusion method is that it introduces spatial distortions in the resultant fused image and does not provide any spectral information. These spatial distortions have been observed in sharp fusion [11] method for medical images and reported in [12]. Since medical images are generally of poor contrast, the spatial information should be preserved in the medical images without introducing any distortion or noise. These requirements of medical images are better preserved in transform domain fusion.

Therefore, transform domain fusion techniques have been used for fusion to overcome the limitations of spatial domain fusion methods. Pyramid and wavelet transforms [13] have been used for multiscale fusion under category of transform domain methods. Transform domain fusion is performed by decomposing source images into transformed representations followed by application of fusion rules to the transformed representation. Finally, fused image is obtained by inverse transformation. Pyramid and wavelet transforms are the mostly used transforms for image fusion. Several pyramid transforms like Laplacian pyramid [14], gradient pyramid [15], contrast pyramid [16], ratio of low pass pyramid [17], morphological pyramid [18], and FSD pyramid [19] have been used for image fusion. However, pyramid transform based fusion methods suffered from blocking effect [20] in the regions where the input images are significantly different. Further, pyramid-transform-based fusion methods do not provide any directional information and have poor signal-to-noise ratio. In contrast to pyramid transforms, wavelet transforms have better representation of detailed features of image; hence, wavelet domain fusion methods provide better results than pyramid-based fusion methods.

The discrete wavelet transform (DWT) is the most commonly used wavelet transform for medical image fusion. A simple DWT-based medical image fusion, which follows weighted fusion rule, has been introduced by Cheng et al. [21]. Another pixel- and region-based multiresolution image fusion for MRI and CT image is discussed in [22]. Several literatures on medical image fusion using DWT can be easily found in [2330] which use different fusion rules for merging multimodality medical images. Some advanced wavelet families such as contourlet transform, curvelet transform, and nonsubsampled contourlet transform [31] have been used for medical image fusion and it has been stated that these advanced wavelet families have better performance than wavelet transforms. However, these are computationally costly and require huge memory. Further, the study and analysis of DWT for medical image fusion has not been studied well and it still needs attention of researchers.

Since estimation of decomposition levels for a wavelet transform [32] has always been challenging and literatures on multilevel medical image fusion [33] have been a motivation for us to explore DWT for multiscale image fusion, therefore, in this work, we have used DWT for multimodality medical image fusion and presented a new multiscale fusion approach using maximum selection rule. The multiscale fusion approach provides us flexibility to select appropriate fused medical image. The experiments have been performed over several multimodality medical images at multiple scales. The fusion results have been compared with other state-of-the-art fusion methods which include several pyramid- and wavelet-transform-based fusion methods and PCA fusion method. The quantitative analysis of the fusion results has been performed with edge strength (), mutual information (MI), entropy (), standard deviation (SD), blind structural similarity index metric (BSSIM), spatial frequency (SF), and average gradient (AG) metrics.

The rest of the paper is organized as follows. Section 2 explains the basics of DWT and its usefulness in image fusion. The proposed fusion method is explained in Section 3. Fusion results and evaluations are given in Section 4. Finally, conclusions of the work are given in Section 5.

2. Wavelet Transform and Image Fusion

Recently, wavelet transforms have emerged as a powerful signal and image processing tool which provides an efficient way of fusion using multiresolution analysis [34]. This section explains the basics of DWT and its usefulness in image fusion.

The DWT of a given signal is performed by analysis and synthesis of signal using scaling function and wavelet function [35, 36]. The basic equation of multiresolution theory is the scaling equation where ’s are the approximation or low-pass coefficients and maintains the norm of the scaling factor by a factor of two.

The wavelet function which is responsible for computing high-frequency or detailed coefficients is given by where ’s are the high frequency or detailed wavelet coefficients.

Signal decomposition is performed using the scaling coefficients and the wavelet coefficients . Forward wavelet analysis of signal at any scale is denoted by where and are scaling and wavelet coefficients at scale and can be computed by following relation:

Reconstruction of signal can be made by combining scaling and wavelet coefficients, and mathematically, it is represented by

The forward and backward analysis of signals provides us the facility to have multiscale signal representations at varying scales. Further, DWT analysis is capable of providing three spatial orientations, namely, horizontal, diagonal, and vertical. This can be denoted by the following combination of scaling and wavelet functions:

The two-dimensional decomposition process using DWT is shown in Figure 1. This can be easily seen that a 2D-DWT provides multiscale representation at different levels. Figure 1(a) shows the DWT decompositions up to level 3, and Figure 1(b) shows the two-level decomposition of Lena image.

This wavelet decomposition is exploited for image fusion and could be easily understood from Figure 2. The DWT provides an efficient way for performing image fusion at multiple scales with several advantages. These are as follows.

Locality. The information of an image is represented by wavelet coefficients, which is local in space and frequency. Thus, for fusion, we can apply fusion rule locally and that would not affect the other portions of the image.

Multiresolution Analysis. The image can be represented at different scales, and this allows producing fused images at multiple levels [33].

Edge Detection. Wavelet transform can act as local edge detectors. The edges in the image are represented by large wavelet coefficients at the corresponding locations, while noise is generally represented by smaller values. Wavelet transform represents three directional edges: vertical, horizontal, and diagonal. This property helps in preserving the edges and implementation of edge-sensitive fusion methods.

Decorrelation. Most of the wavelet coefficients of an image tend to be approximately decorrelated; that is, dependencies between wavelet coefficients are predominantly local. Thus, during the fusion process if there is some change in wavelet coefficients, then generally this would not affect the other portions of image. This allows applying fusion rule on selected wavelet coefficients without affecting other parts of the image.

Energy Compaction. In wavelet domain, the most essential information of the image is compressed into relatively few large coefficients, which coincides with the area of major spatial activity (edges, corners, etc.). This property facilitates the implementation of energy-based fusion rules which preserve the salient features of images.

3. The Proposed Fusion Approach

The usefulness of DWT made it suitable for medical image fusion, where one wishes to capture all relevant information from a single fused image with reduced cost and storage overhead. The proposed fusion approach follows the framework shown in Figure 2; decomposition was followed by application of fusion rule and reconstruction. We have exploited the concept of multiresolution analysis with multiscale fusion approach. The higher the scale is the more the detailed information is captured from source images to fused image. Since medical images are of poor contrast, more detailed and relevant information should be preserved. Thus, by varying scale, we have flexibility to select appropriate fused image for further operations. For the proposed fusion scheme, we vary the scale from minimum to maximum levels. One of the important issues is the selection of wavelet for decomposition. Regularity, number of vanishing moment, and finite support are the few important criteria for selecting mother wavelet [37]. However, it was shown in [31] that short filter banks for wavelet decomposition are useful and works well for fusion. Also, [38] shows the effectiveness of selecting short length wavelet with a fixed criterion. In both cases, “db3” wavelet has been found suitable for decomposition; therefore, we used “db3” wavelet for the proposed fusion scheme.

The proposed fusion approach is based on the maximum selection scheme as high-valued wavelet coefficients carry salient information such as edges, boundaries, and contours. Therefore, the absolute values of wavelet coefficients have been used for deciding fused wavelet coefficients. For two-source medical images and , the steps of the proposed fusion scheme are as follows.

(i) Decompose source images using DWT: where and are the wavelet coefficients of source images and at scale .

(ii) Calculate fused wavelet coefficients at scale by following the expression:

(iii) Reconstruct fused image at scale using inverse DWT:

4. Fusion Results and Discussions

In this section, we have shown fusion results for the proposed method. The fusion results have been shown for three sets of medical image pairs of size 256 × 256 shown in Figures 3(a), 3(b), 4(a), 4(b), 5(a), and 5(b), respectively. The proposed method has been experimented at multiple scales varying from level 2 to level 8 (maximum level of scale) for these medical images. We have performed subjective and objective comparisons to evaluate fusion results obtained by the proposed method. To perform subjective comparison, we have selected gradient pyramid (GP), contrast pyramid (CP), ratio pyramid (RP), PCA, DWT with DBSS, and SIDWT with Haar fusion methods which are available on http://www.metapix.de/ and provided by Rockinger [39].

For objective evaluation of proposed fusion approach with other state-of-the-art fusion methods, nonreference metrics are required as no ground truth image is available for comparison. Therefore, we have used nonreference metrics, namely, edge strength (), mutual information (MI), entropy (), standard deviation (SD), blind structural similarity index metric (BSSIM), spatial frequency (SF), and average gradient (AG) for objective evaluation of our work.

The illustration of fusion results is separately given in Sections 4.1 and 4.2 for subjective and objective evaluations, respectively.

4.1. Subjective Evaluation

The first set of medical images is brain CT and MRI, shown in Figures 3(a) and 3(b). It can be easily seen that the CT image shows the edgy structure while MRI provides information about soft tissues. The results for proposed multiscale fusion method have been shown in Figures 3(c)3(i) from level 2 to level 8. These results show the variations in contrast of fused image as level progresses. On comparing the obtained fused images from level 2 to level 8 (shown in Figures 3(c)3(i)) with GP, CP, RP, and PCA fused images, which are shown in Figures 3(j)3(m), it can be easily concluded that the proposed method outperforms these fusion methods and has good visual representation of fused image. The fused images with GP, CP, RP, and PCA methods are not able to capture the information from CT and MRI pairs. Further, the proposed method has the better quality than DWT with DBSS and nearly same with SIDWT with Haar fusion methods.

The second set of medical images is magnetic resonance angiogram (MRA) and T1-MR image which is shown in Figures 4(a) and 4(b). The comparison of proposed fusion results with GP, CP, RP, PCA, DWT with DBSS, and SIDWT with Haar fusion methods, shown in Figures 4(c)4(o), clearly implies that the fused images with proposed method have better quality and contrast in comparison to other fusion methods.

Similarly, on observing the third set of medical images (CT and MRI) and fusion results for these images which are shown in Figures 5(a)5(o), one can easily verify the fact that again the proposed method has been found superior in terms of visual representation over GP, CP, RP, PCA, DWT with DBSS, and SIDWT with Haar fusion methods.

4.2. Objective Evaluation

For objective evaluation of the fusion results, shown from Figures 35, we have used seven nonreference fusion metrics: edge strength () [40], mutual information (MI) [41], entropy () [9, 12, 27], standard deviation (SD) [12, 27], blind structural similarity index metric (BSSIM) [33], spatial frequency (SF) [9, 33], and average gradient (AG) [27]. These metrics are well defined in the literature and are used excessively for objective evaluation of fusion results. Higher values of these metrics imply better fused result. We have computed the values of fusion results and tabulated them in Tables 13 for fusion results shown in Figures 35, respectively.

On observing Table 1, one can easily observe that the fusion measures for proposed multiscale fusion method from level 2 to level 8 have higher values of fusion measures than any of the GP, CP, RP, PCA, DWT with DBSS, and SIDWT with Haar fusion methods. However, the proposed fusion method from level 2 to level 8 has lesser values of SF than CP and RP fusion methods. Also, the proposed method has lesser values of BSSIM than PCA fused image. For these cases, we have performed an overall comparison in Table 1 and it simply states that the proposed multiscale fusion method has better performance for the first set of medical images.

Similarly, observation of Table 2 yields that the proposed fusion method form level 2 to level 8 has higher values of fusion measures than other fusion methods except values of for PCA and SIDWT with Haar fusion methods and value of BSSIM for PCA fusion method. However, an overall comparison again shows the superiority of the proposed multiscale fusion scheme for the second set of medical images.

Moreover, Table 3 shows the goodness of the proposed fusion method for the third set of medical images except and BSSIM fusion measures. But again by the same criteria chosen for Tables 1 and 2, the proposed multiscale fusion from level 2 to level 8 has better performance than GP, CP, RP, PCA, DWT with DBSS, and SIDWT with Haar fusion methods.

4.3. Combined Evaluation

Since the subjective and objective evaluations separately are not able to examine fusion results, we have combined both, subjective and objective evaluations. The values of fusion measures for fusion results of the first set of medical images (Figure 3) are shown in Table 1. The observations from Table 1 show the variations in the values of SF for CP and RP fusion methods and BSSIM for PCA fusion method. The proposed method from level 2 to level 8 has lesser value of these measures; however, qualitative evaluation of the proposed fusion method from Figure 3 clearly proves the superiority of the proposed method over CP, RP, and PCA fusion methods as these are not able to merge edge and tissue information from source CT and MRI images.

Again, Table 2 shows the higher values of for PCA and SIDWT with Haar fusion method and value of BSSIM for PCA fusion method, than the proposed fusion scheme for the second set of medical images. However, from Figure 4, it can be easily seen that the proposed method provides better visual representation than any of these fusion methods. Hence, combined evaluation shows the goodness of the proposed multiscale fusion approach.

The fusion measures of Figure 5 are given in Table 3 for the third set of medical images. The measures show the variations in the value of and BSSIM for the proposed method, and the proposed method has lesser values of than GP, CP, RP, and SIDWT with Haar fusion methods and lesser values of BSSIM than RP, PCA, and SIDWT with Haar fusion methods. However, qualitative analysis of fusion results shown in Figure 5 clearly shows that GP, CP, RP, and SIDWT with Haar fusion methods have failed to incorporate the features of the source CT and MRI images into one image.

Thus, this combined evaluation for fusion results shown in Figures 35 with fusion measures tabulated in Tables 13 clearly proves the superiority of the proposed multiscale fusion approach over GP, CP, RP, PCA, DWT with DBSS, and SIDWT with Haar fusion methods.

5. Conclusions

In this work, we have proposed a new multiscale image fusion approach for multimodal medical images in wavelet domain and used DWT for proposed fusion method. The multimodal medical images are fused at multiple scales from level 2 (minimum) to level 8 (maximum) scales with maximum fusion rule. The multiscale image fusion method enables the selection of appropriate fused image with better flexibility. To show the effectiveness of the proposed work, we have performed subjective evaluation and objective evaluation of the proposed fusion method with gradient pyramid (GP), contrast pyramid (CP), ratio pyramid (RP), PCA, DWT with DBSS, and SIDWT with Haar fusion methods. The comparative analysis of the fusion results has been performed with edge strength (), mutual information (MI), entropy (), standard deviation (SD), blind structural similarity index metric (BSSIM), spatial frequency (SF), and average gradient (AG) fusion metrics. Since the subjective and objective evaluations are separately not sufficient for analysis of fusion results, we have performed combined evaluation which proved the superiority of the proposed multiscale fusion approach over GP, CP, RP, PCA, DWT with DBSS, and SIDWT with Haar fusion methods.