Abstract
Medical image fusion plays an important role in diagnosis and treatment of diseases such as image-guided radiotherapy and surgery. Although numerous medical image fusion methods have been proposed, most of these approaches are sensitive to the noise and usually lead to fusion image distortion, and image information loss. Furthermore, they lack universality when dealing with different kinds of medical images. In this paper, we propose a new medical image fusion to overcome the aforementioned issues of the existing methods. It is achieved by combining with rolling guidance filter (RGF) and spiking cortical model (SCM). Firstly, saliency of medical images can be captured by RGF. Secondly, a self-adaptive threshold of SCM is gained by utilizing the mean and variance of the source images. Finally, fused image can be gotten by SCM motivated by RGF coefficients. Experimental results show that the proposed method is superior to other current popular ones in both subjectively visual performance and objective criteria.
1. Introduction
Multimodal medical image fusion is a hot research topic and drives a lot of attention for increasing demands for diagnosis and treatment of diseases. There are various modalities of medical images today such as computed tomography (CT), magnetic resonance angiography (MRA), magnetic resonance imaging (MRI), and functional MRI (fMRI) [1]. Different modality medical images can reflect different information of human organs such as CT can only provide dense structures like bones and implants with less distortion, while MR can provide normal and pathological soft tissues information. It is really helpful to doctor by combining complimentary features of different imaging modalities into one fused image. For example, MRI/CT imaging can be combined for diagnosis and treatment planning [2, 3].
This paper focuses on the pixel level medical image fusion technology. Up to now, a lot of medical image fusion algorithms have been proposed. Examples include principal component analysis fusion algorithm (PCA) [4], guided filtering fusion algorithm (GFF) [5], medical image fusion algorithm based on wavelet in [6], fusion algorithm based on Contourlet transform (CT) in [7], fusion algorithms based on nonsubsampled Contourlet transform (NSCT) in [8], fusion algorithm based on Ripplet in [9], and fusion algorithm based on Shearlet and PCNN in [10], and so on. Although these methods produce high-quality images, they also will lead to loss of information and pixel distortion due to nonlinear operations of fusion rules and blocky artifacts [11]. To address these problems, Wang et al. proposed a new medical fusion method based on SCM in [11], which can get much better fusion effects; but, in their method, the parameters of SCM are fixed to some constants which will obviously not be widely applicable to all kinds of medical image fusion. Although the gray values of images can be used as the input of SCM like Wang’s method, they are more sensitive to environment than the edge information [12].
In our paper, these disadvantages are overcome by using RGF and adaptive threshold in SCM. RGF is an edge aware filtering, and it can remove the small texture of images without blurring the image edge [13]. Therefore, in this paper, RGF is used to extract the saliency (edge information); and then, the coefficients of RGF are normalized and taken as the stimuli of the SCM. In order to be widely applied to all kinds of medical image fusion, adaptive threshold of SCM is proposed.
This paper is organized as follows. In Section 2, we give a brief review of RGF and SCM. In Section 3, we give the steps of the new image fusion algorithm. In Section 4, we demonstrate the experimental results of the proposed method and the comparisons with other typical fusion methods; and, in the last section, we explore some conclusions.
2. Rolling Guidance Filter and Spiking Cortical Model
2.1. Rolling Guidance Filter
Zhang et al. [13] proposed a new framework called RGF to filter images based on a rolling guidance with the complete control of detail smoothing under a scale measure. Compared to other edge preserving filters, RGF is implemented iteratively, which has a fast convergence property. It is simple and fast and also easy to understand. RGF can preserve large-scale structures automatically, where small structure removal and edge recovery are two main steps in RGF; see Figure 1 [13].

Firstly, Gaussian filter is used to remove the small structure. denotes the input image and denotes the output image. denotes the standard deviation of Gaussian filter. and are the indexes of pixel coordinates in the image. The filter is as follows:where is for normalization and denotes the set of pixels in the windows of Gaussian filter whose center is at .
Secondly, a joint bilateral filter is used to recover the edge iteratively. Initially, is set as the output of the Gaussian filtering. is the output of the th iteration of joint bilateral filtering with the input and . Considerwhere is for normalization. denotes the same input image in (2). controls the range weights.
Finally, two main steps in RGF can be combined into one by starting rolling guidance simply from a constant-value image. In (2), if we set all values in to a constant , that is, , , it updates to ; the new form is exactly the same as (2).
From Figure 1, we can see that the small structure in medical images is removed by RGF. RGF can remove small-scale structures while preserving other content and is parallel in terms of importance to previous edge-preserving filters. It enlists the power of distinguishing between structures in terms of scales without knowing the exact form (or model) of texture, details, or noise.
2.2. Spiking Cortical Model
The SCM [12] is derived from Eckhorn’s model and it conforms to the physiological characteristic of human visual neural system. In fact, Wang’s method [11] provides an effective means for fusion of the different kinds of medical images. In the spiking cortical model, each neuron consists of three parts: feeding and linking field, modulating product, and pulse generator; see Figure 2.

In the following expressions, the indexes and refer to the pixel location in the image, and refer to the locations of its neighboring pixels, and denotes the current iteration times. The receiving and linking field and modulating product are given bywhere is the internal activity and is the attenuation coefficient of . is the external stimulus. is the synaptic linking weight and is the previous output pulse.
The pulse generator determines the firing events in the model in (4). depends on the internal activity and threshold. Consider
The dynamic threshold of the neuron is defined aswhere denotes the attenuation coefficient and denotes the threshold magnitude coefficient. Normally, the size of internal activity matrix is the same as the external stimulus matrix, and is always initialized to zero matrices; and the image matrix can be input as external stimulus of SCM; that is, . However, the external stimulus of SCM in this paper is replaced by RGF coefficients of image.
In our paper, we find that the expectation and variance of the sources images can be used to calculate threshold which can reach better fusion results. The adaptive threshold is defined aswhere denotes sources images needed to fuse and denotes expectation function; denotes variance function; and the fired times can be computed as follows:where denotes the total number of the fired times of neurons after the current iteration.
3. Image Fusion Based on RGF and SCM
Without loss of generality, we suppose that and are two medical images with different sensor to fuse, and is the fused image.
Firstly, the RGF coefficients of and can be represented as follows. Note that all input images must be registered and also have the same size and identical resolution. Considerwhere denotes the RGF function.
Secondly, the normalized RGF coefficients are taken as the stimulus of the two SCMs to obtain where denotes the SCM with adaptive threshold by (3)–(7). and denote the total fired times motivated by RGF coefficients and , respectively.
Finally, the fused image can be refined as follows:
In conclusion, the framework of the proposed fusion algorithm is shown in Figure 3.

4. Experimental Results
4.1. The Comparison of Other Fusion Methods
In order to evaluate the performance of the proposed fusion method, we introduce some objective criteria such as mutual information (MI) [8], metric [14], metric [14], and metric [14]. MI measures the amount of information transferred to the fused image from the source images. utilizes Sobel edge detector to measure the amount of edge information which is transferred from the source images to the fused image. In general, the higher MI and values indicate the better fused result. is introduced to evaluate the information lost during the fusion process. The lost information is available in the source images but not in the fused image. represents fusion artifacts that were introduced into the fused image. It is clear that the smaller and the better the fused image. It is worth noting that the complimentary , , and indicate that the sum of all these should result in unity [14]. Furthermore, the fused algorithms are evaluated by using the Matlab codes on Intel Core2 2.6 GHz machines with a 4 GB RAM.
To evaluate the performance of the proposed fusion method, the experiments have been performed on four pairs of multimodal medical images as shown in Figure 4. These pairs of images are divided into the four groups. Group a contains Figures 4(a) and 4(e), and Figure 4(a) is a CT image of brain which can provide information of bones. Figure 4(e) is an MRI image of brain which can provide some information about soft tissue. Group b contains Figures 4(b) and 4(f), and Figure 4(b) is B ultrasound image of thyroid tumor, as the anatomical imaging, providing organ organization structure information. Figure 4(f) is SPECT image of thyroid tumors, as functional imaging, which can provide information of benignancy and malignanancy of thyroid tumor. Group c contains Figures 4(c) and 4(g), and they are CT image and T1-weighted MR-GAD image of several focal lesions involving basal ganglia. Group d contains Figures 4(d) and 4(h), and they are T1-MRI and T2-MRI that involve the lesion in the frontal lobe.

(a)

(b)

(c)

(d)

(e)

(f)

(g)

(h)
The following algorithms are used for comparison studies in the experiments: GFF based on guided filtering proposed in [5] (GFF), medical image fusion based on nonsubsampled direction complex wavelet transform proposed in [7] (NDCWT), NSCT-based multimodal medical image fusion using pulse-coupled neural network and modified spatial frequency proposed in [8] (NSCT-SF-PCNN), fusion algorithm based on Shearlet transform and PCNN proposed in [10] (ST-PCNN), and fusion algorithm based on SCM proposed in [11] (SCM). For fair comparison, we use the parameters that were reported by the authors to yield the best fusion results. In our method, and , and the iteration number of RGF is set to 4. The parameters of SCM are set as follows: , ; synaptic linking weight = [0.1091, 0.1409, 0.1091; 0.1409, 0, 0.1409; 0.1091, 0.1409, 0.1091]; iteration times and the constant threshold . To verify the effect of each part of our algorithm, fused method based on SCM with self-adaptive threshold (SCM-A) and fused method based on RGF and SCM with constant threshold (RGF-SCM-C) are also compared to our method (RGF-SCM).
The above methods are utilized to fuse four group images in Figure 4, respectively. Figures 5–8 show the fused images by eight fused methods. From the fusion results in Figures 5(a)–5(h), it can be clearly seen that the image fused by our method reaches a higher contrast among all the fused images. Comparing the fused images of each algorithm, we can see that the proposed fusion algorithm preserves the texture information of source images well at the upright of the fused image. At the same time, less useless image information such as block effect and artifacts are introduced in the fused images in present scheme.

(a)

(b)

(c)

(d)

(e)

(f)

(g)

(h)

(a)

(b)

(c)

(d)

(e)

(f)

(g)

(h)

(a)

(b)

(c)

(d)

(e)

(f)

(g)

(h)

(a)

(b)

(c)

(d)

(e)

(f)

(g)

(h)
The fusion results of the eight algorithms in Figures 6(a)–6(h) show that our method has the best visual effect in all the fused methods. Figures 6(a)–6(d) show that GFF, NDCWT, NSCT-SF-PCNN, and ST-PCNN cannot fuse this type of medical images well; and Figures 6(e)–6(h) show that the method based on SCM can achieve much better performances. Comparing the fused image of SCM, our method not only preserves the texture information of source images but also suppresses useless image information such as block effect and artifacts, which should be attributed to the adaptive threshold in SCM and the saliency of medical images which is captured by RGF.
From the objective criteria shown in Table 1, one can find that our algorithm has the best objective criteria. The highest MI and mean that most useful information and edge information are converted into the fused result by our algorithm. The least means that fewest information of source images is lost by our method. The least means that least fusion artifacts are introduced into the fused image by our method. Therefore, our method can be regarded as a kind of good medical image fusion algorithm.
Figures 7 and 8 show the fused images of group c and d by eight fused methods. The fusion results shown in Figures 7(a)–7(f) and 8(a)–8(f) indicate that our method both has a higher contrast in all the fused methods and preserves the texture information of source images, suppressing useless image information such as block effect and artifacts.
From the objective criteria shown in Table 2, we can find that our algorithm always has the best objective criteria. Therefore, our method can be regarded as a robust medical image fusion algorithm.
4.2. The Robust to Noise
In order to validate the robustness of the algorithm, Gaussian noise with different noise variance from 5 to 50 is added to group a. The peak signal to noise ratio (PSNR) [15] is used to evaluate the performance of different fused methods. As the perfect fused image does not exist, the average of PSNR between fused image and source images is computed as measurement. It is defined as follows:where denotes source images needed to fuse. denotes the fused image.
Figure 9(a) shows the fused image by SCM and Figure 9(b) shows the fused image by RGF-SCM. Obviously, our method has better visual performance than SCM.

(a)

(b)
Figure 10 shows PSNR of fused images by RGF-SCM and SCM. Obviously, the PSNR of fused image by RGF-SCM is higher than that by SCM when the source images have heavy noises. When the noise variance increases, the difference between PSNR of fused images by RGF-SCM and SCM increases too. It means that the performance of RGF-SCM becomes more efficient than SCM when the noise variance grows. Therefore, our method can be regarded as a robust medical image fusion algorithm.

5. Conclusions
A new fused method based on RGF and improved SCM is proposed to improve the medical fusion effect. The new fused method can enhance robustness to noise and extend SCM to fuse other kinds of medical images. Experimental results demonstrate that the proposed method is better than state-of-the-art medical image fusion methods in both visual appearance and objective criteria. In this paper, we just only cover the fusion of 2D images; however, 3D data sets are becoming increasingly important in medical procedure. It would be interesting to know whether and how an application to 3D data sets could be achieved. In the future research, we will extend current work to 3D data sets.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
Acknowledgments
The authors thank the associate editor and the anonymous reviewers for their careful work and valuable suggestions for an earlier version of this paper. They are also grateful to doctoral Zhu and Professor Wakin in Colorado School of Mines who kindly helped them to correct grammar and spelling mistakes in the paper. Moreover, their work is supported in part by Natural Science Foundation of China under Grant no. 61401308, Natural Science Foundation of Hebei Province under Grant no. 2013210094, Natural Science Foundation of Hebei University under Grant no. 2014-303, and Science and technology support project of Baoding City under Grant no. 15ZG036. Sources images can be downloaded from http://www.med.harvard.edu/aanlib/home.html and Hebei University.