Abstract
In imagebased medical decisionmaking, different modalities of medical images of a given organ of a patient are captured. Each of these images will represent a modality that will render the examined organ differently, leading to different observations of a given phenomenon (such as stroke). The accurate analysis of each of these modalities promotes the detection of more appropriate medical decisions. Multimodal medical imaging is a research field that consists in the development of robust algorithms that can enable the fusion of image information acquired by different sets of modalities. In this paper, a novel multimodal medical image fusion algorithm is proposed for a wide range of medical diagnostic problems. It is based on the application of a boundary measured pulsecoupled neural network fusion strategy and an energy attribute fusion strategy in a nonsubsampled shearlet transform domain. Our algorithm was validated in dataset with modalities of several diseases, namely glioma, Alzheimer’s, and metastatic bronchogenic carcinoma, which contain more than 100 image pairs. Qualitative and quantitative evaluation verifies that the proposed algorithm outperforms most of the current algorithms, providing important ideas for medical diagnosis.
1 Introduction
Multimodal medical imaging is a research field that has been getting increasing attention in the scientific community in the last few years, specially due to its significance in medical diagnosis, computer vision, and internet of things [3, 5, 15, 20, 28, 31, 32, 35]. Defined as the simultaneous production of signals belonging to different medical imaging techniques, one of the biggest challenges in this research field is how to combine (or fuse) in an effective and optimal way multimodal medical imaging sensors, such as positron emission tomography (PET), singlephoton emission computed tomography (SPECT), and magnetic resonance imaging (MRI). This image fusion process comprises many techniques and research areas, ranging from image processing techniques, computer vision to pattern recognition, with the goal of promoting more accurate medical diagnosis and more effective medical decisionmaking [8, 10, 18, 26, 45].
1.1 Current challenges in multimodal image fusion
Image fusion can usually be divided into three levels: pixellevel, featurelevel, and decision level [21, 31, 42,43,44, 47]. Since the aim is to fuse pixel information from source images, medical image fusion belongs to the pixellevel.
Multiscale transform (MST) method is one of the most famous categories [40]. Commonly, the MST fusion methods consist of three steps. First, the source images are transformed into MST domain. Then, the parameters in different scales merged in light of a specific fusion strategy. Finally, the fused image is reconstructed through the corresponding inverse transform. The MST methods mainly contain the Laplacian pyramid (LP) [6], the wavelet transform (WT) [27, 34], the nonsubsampled contourlet transform (NSCT) [49], and the nonsubsampled shearlet transform (NSST) [4, 23, 38]. However, if the MST method performs without other fusion measures, some unexpected block effect may appear [39].
To overcome this disadvantage, some fusion measures are applied in the MST method. For instance, spatial frequency (SF), local variance (LV), the energy of image gradient (EIG) and summodifiedLaplacian (SML) are commonly used as fusion measures [17, 41]. However, most of these measures are acquired in the spatial domain or loworder gradient domain, which means the fusion map may not be always precise. This imprecision may lead to blocking artefacts.
Except for traditional MST methods, the edgepreserving filtering (EPF)based MST decomposition method are also commonly used. In the EPFMST methods, Gaussian filtering and EPF are used to decompose the input image into two scalelayers and one base layer. Then, three layers are fused based on suitable fusion strategies. Finally, the fused image is reproduced by a reconstruction algorithm. The EPFMST methods contain bilateral filtering (BF)based [51], curvature filtering (CF)based [40], and cooccurrence filtering (CoF)based [37] methods.
1.2 A pulsecoupled neural network model for medical image fusion
To overcome this challenge, a method called pulsecoupled neural network (PCNN) has been proposed in the literature [46]. This method was initially proposed to emulate the underlying mechanisms of a cat’s visual cortex and became later an essential method in image processing [29]. Kong et al. presented an SF modulated PCNN fusion strategy in NSST domain with the solution of infrared and visible image fusion [19]. Inspired by this kind of fusion measure modulated by the PCNN model, one interesting research path would be a solution to a new measure to modulate PCNN in the medical image fusion field.
To further improve the fusion quality of medical images, we propose a medical image fusion method based on boundary measure modulated by a pulsecoupled neural network in the nonsubsampled shearlet domain. Firstly, the source images are transformed into the NSST domain with lowfrequency bands and highfrequency bands. Then, the lowfrequency bands are merged through an energy attributebased fusion strategy, and the highfrequency bands are merged through a boundary measure modulated PCNN strategy. Finally, the fused image is reconstructed by combining the inverse NSST. We evaluate the proposed algorithm by comparing its performance with several existing methods using both a quantitative and qualitative evaluation. Experimental results demonstrate that the proposed method performs better than most of the existing fusion methods.
1.3 Contribution
The main contributions of the proposed research article are the following:

1.
A medical image fusion framework based on boundary measured PCNN in NSST domain, which can complete the fusion task effectively;

2.
The application of a boundary measured PCNN model for highfrequency bands. In this method, the gradient information of the image can be easily extracted, and the size of the structure can be changed to adapt to the scale of structure;

3.
The application of an energy attributebased fusion strategy to lowfrequency bands.
Experiments conducted in this research paper suggest that the proposed boundary measured PCNNNSST achieves the best performance in most cases in qualitative and quantitative when compared to other stateoftheart image fusion techniques.
1.4 Organization
The rest of this paper is organized as follows. In Sect. 2, it is presented the most significant works in the image fusion domain. In Sect. 3, the proposed fusion method BMPCNNNSST is described. In Sect. 4, it is presented the set of experiments that were performed to evaluate the proposed algorithm. Finally, in Sect. 5, the main conclusions of this research work are presented.
2 Related work
In this section, we present an overview of the most significant image fusion algorithms in the literature, namely the nonsubsampled shearlet transform (Sect. 2.1), the multiscale morphological gradient (Sect. 2.2), and the pulsecoupled neural network (Sect. 2.3).
2.1 Nonsubsampled shearlet transform
The nonsampled shearlet transform is an image fusion method, originally proposed by Easley [13]. It consists in combining the nonsubsampled pyramid transform with different shearing filters, and it has the characteristics of multiscale and multidirectionality. The nonsubsampled pyramid transform makes it invariant, which is superior than the LP, and WT methods. Additionally, since the size of the shearing filter is smaller than the directional filter, NSST can represent smaller scales, which makes it better than NSCT.
Given the superiority of its underlying functions, NSST performs better than most commonly used MST. It is therefore widely used in the field of image denoising [36] and image fusion [22].
The NSST model can be described as follows. For the case, n = 2, the shearlet function is satisfied
where ψ ∊ L^{2}(R^{2}), both A and B are invertible matrices with size 2 × 2, and \( \left {\det B} \right = 1 \). For instance, A and B can be represented as
In this situation, the tiling of the frequency plane of NSST is shown in Fig. 1 It can be seen that (a) represents the decomposition, and (b) represents the size of the frequency support of the shearlet element \( \psi_{i,l,k} \).
For convenience, two related functions are used to represent the NSST and the inverse NSST
where \( {\text{nsst\_de}}\left( \cdot \right) \) represents the NSST decomposition function for the input image \( I_{\text{in}} \), and \( {\text{nsst\_re}}\left( \cdot \right) \) represents the NSST reconstruction steps for the reconstructed image \( I_{\text{re}} \). The parameters \( L_{n} \) and \( H_{n} \) represent lowfrequency subbands and highfrequency subbands, respectively.
2.2 Multiscale morphological gradient
Multiscale morphological gradient (MSMG) is an effective operator which extracts gradient information from an image in order to indicate the contrast intensity in the close neighborhood of a pixel in the image. For this reason, MSMG is a method that is highly efficient and used in edge detection and image segmentation. In image fusion, MSMG has been used as a type of focus measure in multifocus image fusion [50]. The specific details of MSMG are as follows.
A multiscale structuring element is defined as
where \( {\text{SE}}_{1} \) denotes a basic structure element, and t represents the number of scales.
The gradient feature \( G_{t} \) can be represented by the morphological gradient operators from the image f.
where \( \oplus \) and \( \odot \) denote the morphological dilation and erosion operators, respectively. \( \left( {x,y} \right) \) denotes the pixel coordinate.
From the multiscale structuring element and the gradient feature, then one can obtain the MSMG by computing the weighted sum of gradients over all scales.
where \( w_{t} \) represents the weight of gradient in tth scale, and it can be represented as
Figure 2 shows an example of MSMG. One can see that the boundary information of the images has been well extracted, which demonstrates the effectiveness of the boundary measure.
2.3 Pulsecoupled neural network
As the thirdgeneration artificial neural network, PCNN has achieved great success in the image fusion field. A PCNN model often contains three parts: the receptive field, the modulation field and the pulse generator. The expressions of a simplified dualchannel PCNN model can be defined as
As is shown in Fig. 3, \( S_{ij}^{1} \) and \( S_{ij}^{2} \) denote the pixel value of two input images at point \( \left( {i,j} \right) \) in the neural network; \( L_{ij} \) represents the linking parameter; \( \beta_{ij}^{1} \) and \( \beta_{ij}^{2} \) denote the linking strength; \( F_{ij}^{1} \) and \( F_{ij}^{2} \) represent the feedback of inputs. \( U_{ij} \) is the output of the dualchannel. \( \theta_{ij} \) is the threshold of step function, \( d_{e} \) is the declining extent of the threshold, \( V_{\theta } \) decides the threshold of the active neurons, and \( T_{ij} \) is the parameter to determine the number of iterations. \( Y_{ij} \left( k \right) \) is the kth output of PCNN.
3 A Bounded measured PCNN in NSST domain algorithm
In this section, we present the proposed algorithm for multimodal medical image fusion: a bounded measured PCNN approach in the NSST domain (BMPCNNNSST). The framework of the proposed algorithm is illustrated in Fig. 4. The fusion algorithm consists in four parts: the NSST decomposition, the lowfrequency fusion, the highfrequency fusion, and the NSST reconstruction.
The algorithm starts with a pseudocolor image source A which contains threebands (PET/SPECT image). The first step is to apply an intensityhuesaturation (IHS) transform in A, which will result in a pair containing the intensity image \( I_{A} \) and a source image B. After performing the fusion of this image pair, an inverse IHS transform is applied in order to obtain the final fused image.
3.1 NSST decomposition
An Nlevel NSST decomposition is performed on images \( I_{A} \) and B to acquire the decomposition bands \( L_{A} \), \( H_{A}^{l,k} \) and \( L_{B} \), \( H_{B}^{l,k} \) based on Eq (3), where L denotes lowfrequency subbands and \( H^{l,k} \) represents highfrequency subbands at level l with direction k.
3.2 Lowfrequency fusion
The lowfrequency subband contains most information of the source images (texture structure and background). In this paper, an energy attribute (EA) fusion strategy is presented in the lowfrequency fusion. This EA fusion strategy is divided into three steps:

1.
The intrinsic property values of the lowfrequency subband are computed as
$$ IP_{A} = \mu_{A} + Me_{A} $$(18)$$ IP_{B} = \mu_{B} + Me_{B} $$(19)where μ and Me represent the mean value and the median value of \( L_{A} \) and \( L_{B} \), respectively.

2.
The EA function \( E_{A} \) and \( E_{B} \) are calculated by
$$ E_{A} \left( {x,y} \right) = \exp \left( {\alpha \left {L_{A} \left( {x,y} \right)  IP_{A} } \right} \right) $$(20)$$ E_{B} \left( {x,y} \right) = \exp \left( {\alpha \left {L_{B} \left( {x,y} \right)  IP_{B} } \right} \right) $$(21)where \( \exp \left( {\alpha \left {L_{A} \left( {x,y  IP_{A} } \right)} \right} \right) \) represents the exponential operator, and α denotes the modulation parameter.

2.
The fused lowfrequency subband is obtained by a weighted mean
$$ L_{F} \left( {x,y} \right) = \frac{{E_{A} \left( {x,y} \right) \times L_{A} \left( {x,y} \right) + E_{B} \left( {x,y} \right) \times L_{B} \left( {x,y} \right)}}{{E_{A} \left( {x,y} \right) + E_{B} \left( {x,y} \right)}} $$(22)
3.3 Highfrequency fusion
While lowfrequency subband contains most information about the source images (such as background and texture), highfrequency subbands contain more information about details in images (for example, pixellevel information). Since in the PCNN model one pixel corresponds to one neuron, it is suitable to use PCNN in highfrequency subbands. In addition, modulating PCNN with MSMG can increase the spatial correlation in the image. Therefore, the MSMG operator can be used to adjust the linking strength between \( \beta_{ij}^{1} \) and \( \beta_{ij}^{2} \)
where \( M_{A} \) and \( M_{B} \) are computed by Eq. (7).
The highfrequency subbands are merged based on this MSMGPCNN model until all neurons are activated (equal to 1). The fused highfrequency subbands can be obtained by
where \( T_{xy,A} \) and \( T_{xy,B} \) can be computed using Eq (15).
3.4 NSST reconstruction
The fused image F is reconstructed by \( L_{F} \) and \( H_{F}^{l,k} \) through the inverse NSST according to Eq (4)
4 Experiments
To validate the proposed algorithm, a set of experiments was made using three datasets representing different diseases: (1) glioma, (2) mild Alzheimer’s, and (3) hypertensive encephalopathy. The proposed algorithm was compared with seven stateoftheart image fusion methods. Qualitative and quantitative analyses were made to assess its performance. The code of the paper is made available.^{Footnote 1}
4.1 Datasets
To verify the proposed algorithm, more than 100 pairs of multimodal medical images were used, including 30 image pairs of MRIPET and 13 image pairs of MRISPECT of glioma disease, 10 image pairs of MRIPET of mild Alzheimer’s disease, 11 image pairs of MRISPECT of Metastatic bronchogenic carcinoma, 10 image pairs of MRISPECT of hypertensive encephalopathy, 11 image pairs of MRISPECT of motor neuron disease, and 16 image pairs of MRISPECT of normal aging. All the image pairs can be downloaded from the Whole Brain Atlas dataset [1]. All the pairs have been perfectly registered, and the size of all images is 256 × 256.
4.2 Comparison methods
The proposed BMPCNNNSST algorithm is compared with seven stateoftheart fusion methods. There methods are the convolutional neural network (CNN) [ [24], [53] ], the convolutional sparsitybased morphological component analysis (CSMCA) [25], the information of interest in local Laplacian filtering domain (LLFIOI) [11], the neurofuzzy approach (NFA) [9], the parameteradaptive PCNN in NSST domain (NSSTPAPCNN) [48], the phase congruency and local Laplacian energy in NSCT domain (PCLLENSCT) [52], and the parallel saliency features (PSF) [12]. These methods are recently proposed fusion methods. The parameters that we used in our experiments are the same as in their papers.
4.3 Parameter settings
In the proposed BMPCNNNSST algorithm, the following parameters were used:

the NSST decomposition level N is set to 4;

the number of directions in each level is set to 16,16,8,8;

the modulation parameter is set to 4;

the scales number of MSMG operator t is set to 3.
4.4 Evaluation metrics
To analyze the performance of the proposed algorithm in a quantitative way, we evaluate the different fusion methods using five metrics: entropy (EN), standard deviation (SD), normalized mutual information (NMI) [14], Piella’s structure similarity (SS) [33], and visual information fidelity (VIF) [16]. In general, both SD and EN can measure the amount of information of the fused image. NMI evaluates the amount of information transferred from the source images to fused image. SS mainly evaluates the structure similarity between source images and fused image. VIF evaluates the visual information fidelity between the source images and fused image. More detailed information about these evaluation metrics can be found on the references related to each fusion method.
4.5 Experimental results
The results of medical image fusion cannot be completely dependent on visual effects evaluation. As long as the feature information is not lost, the medical diagnosis will not be misjudged because of this, and the visual effect will be acceptable. Therefore, in this paper, each disease demonstrates a set of experimental results, which is shown in Figs. 5, 6, 7, 8, 9, 10, and 11. Different methods have different visual effects, but the feature information does not seem to be lost. Therefore, objective evaluation indicators are needed for a further quantitative evaluation.
The mean value of each metrics of different fusion methods is listed in Tables 1, 2, 3, 4, 5, 6 and 7. Each column represents the same metrics for different methods. The highest value is shown in bold, while the second highest in italic. It can be seen that the proposed method performs the best in half of the cases. Even if it is not the highest value in one column, it is still the second highest value, except in the NMI of the normal aging case.
5 Conclusion
In this paper, a multimodal medical image fusion algorithm is proposed based on boundary measured PCNN and EA fusion strategies in NSST domain. The main advantage of the proposed algorithm is that the two fusion strategies are suitable for different scales. Decomposing images into different scales with NSST can give full play to the advantages of the two fusion strategies. Meanwhile, as an excellent decomposition method, NSST can well blend the differences of multimodal medical images. The performance of the proposed algorithm has been verified in public datasets, which represents it has reached stateoftheart level. One of the important outcomes of this paper is reported in “Appendix,” which showed the experimental performance of different values of α and t. One can see that when α = 4 and t = 3 the performance is the best in most cases. Since the deep learning technology has been widely used, in the future research, we will focus on the deep learning method in multimodal medical image fusion [2, 7, 30].
References
Whole brain atlas. http://www.med.harvard.edu/AANLIB/
Ahmed I, Din S, Jeon G, Piccialli F (2019) Exploring deep learning models for overhead view multiple object detection. IEEE Int Things J. https://doi.org/10.1109/JIOT.2019.2951365
Amato F, Moscato V, Picariello A, Piccialli F, Sperl G (2018) Centrality in heterogeneous social networks for lurkers detection: an approach based on hypergraphs. Concurr Comput Pract Exp 30(3):e4188
Asha C, Lal S, Gurupur VP, Saxena PP (2019) Multimodal medical image fusion with adaptive weighted combination of nsst bands using chaotic grey wolf optimization. IEEE Access 7:40782–40796
Bebortta S, Senapati D, Rajput NK, Singh AK, Rathi VK, Pandey HM, Jaiswal AK, Qian J, Tiwari P (2020) Evidence of powerlaw behavior in cognitive IoT applications. Neural Comput Appl pp 1–13
Burt P, Adelson E (1983) The Laplacian pyramid as a compact image code. IEEE Trans Commun 31(4):532–540
Casolla G, Cuomo S, Di Cola VS, Piccialli F (2020) Exploring unsupervised learning techniques for the internet of things. IEEE Trans Ind Inform 16(4):2621–2628. https://doi.org/10.1109/TII.2019.2941142
Chouhan V, Singh SK, Khamparia A, Gupta D, Moreira C, Damasevicius R, de Albuquerque VHC (2020) A novel transfer learning based approach for pneumonia detection in chest Xray images. Appl Sci 10(2):559
Das S, Kundu MK (2013) A neurofuzzy approach for medical image fusion. IEEE Trans Biomed Eng 60(12):3347–3353
Du J, Li W, Lu K, Xiao B (2016) An overview of multimodal medical image fusion. Neurocomputing 215:3–20
Du J, Li W, Xiao B (2017) Anatomicalfunctional image fusion by information of interest in local Laplacian filtering domain. IEEE Trans Image Process 26(12):5855–5866
Du J, Li W, Xiao B (2018) Fusion of anatomical and functional images using parallel saliency features. Inf Sci 430:567–576
Easley G, Labate D, Lim WQ (2008) Sparse directional image representations using the discrete shearlet transform. Appl Comput Harmon Anal 25(1):25–46
Estevez PA, Tesmer M, Perez CA, Zurada JM (2009) Normalized mutual information feature selection. IEEE Trans Neural Netw 20(2):189–201
Gochhayat SP, Kaliyar P, Conti M, Prasath V, Gupta D, Khanna A (2019) LISA: lightweight contextaware IoT service architecture. J Clean Prod 212:1345–1356
Han Y, Cai Y, Cao Y, Xu X (2013) A new image fusion performance metric based on visual information fidelity. Inf Fusion 14(2):127–135
Huang W, Jing Z (2007) Evaluation of focus measures in multifocus image fusion. Pattern Recognition Lett 28(4):493–500
Jaiswal AK, Tiwari P, Kumar S, Gupta D, Khanna A, Rodrigues JJ (2019) Identifying pneumonia in chest xrays: a deep learning approach. Measurement 145:511–518
Kong W, Zhang L, Lei Y (2014) Novel fusion method for visible light and infrared images based on NSSTSFPCNN. Infrared Phys Technol 65:103–112
Kumar S, Tiwari P, Zymbler M (2019) Internet of things is a revolutionary approach for future technology enhancement: a review. J Big Data 6(1):111
Li S, Kang X, Fang L, Hu J, Yin H (2017) Pixellevel image fusion: a survey of the state of the art. Inf Fusion 33:100–112
Liu S, Wang J, Lu Y, Li H, Zhao J, Zhu Z (2019) Multifocus image fusion based on adaptive dualchannel spiking cortical model in nonsubsampled shearlet domain. IEEE Access 7:56367–56388
Liu X, Mei W, Du H (2018) Multimodality medical image fusion based on image decomposition framework and nonsubsampled shearlet transform. Biomed Signal Process Control 40:343–350
Liu Y, Chen X, Cheng J, Peng H (2017) A medical image fusion method based on convolutional neural networks. In: 2017 20th international conference on information fusion (Fusion), pp 1–7. IEEE
Liu Y, Chen X, Ward RK, Wang ZJ (2019) Medical image fusion via convolutional sparsity based morphological component analysis. IEEE Signal Process Lett 26(3):485–489
Mallick PK, Ryu SH, Satapathy SK, Mishra S, Nguyen GN (2019) Brain MRI image classification for cancer detection using deep wavelet autoencoderbased deep neural network. IEEE Access 7:46278–46287
Nair RR, Singh T (2019) Multisensor medical image fusion using pyramidbased dwt: a multiresolution approach. IET Image Proc 13(9):1447–1459
Piccialli F, Bessis N, Jung JJ (2020) Data science challenges in industry 4.0. IEEE Trans Ind Inform
Piccialli F, Casolla G, Cuomo S, Giampaolo F, di Cola VS (2020) Decision making in iot environment through unsupervised learning. IEEE Intell Syst 35(1):27–35. https://doi.org/10.1109/MIS.2019.2944783
Piccialli F, Cuomo S, di Cola VS, Casolla G (2019) A machine learning approach for iot cultural data. J Ambient Intell Human Comput pp 1–12
Piccialli F, Cuomo S, Giampaolo F, Casolla G, di Cola VS (2020) Path prediction in iot systems through Markov chain algorithm. Fut Gen Comput Syst
Piccialli F, Yoshimura Y, Benedusi P, Ratti C, Cuomo S (2020) Lessons learned from longitudinal modeling of mobileequipped visitors in a complex museum. Neural Comput Appl 32:7785–7801. https://doi.org/10.1007/s00521019040998
Piella G, Heijmans H (2003) A new quality metric for image fusion. In: Proceedings 2003 international conference on image processing (Cat No 03CH37429), vol 3, pp III173. IEEE
Polinati S, Dhuli R (2020) Multimodal medical image fusion using empirical wavelet decomposition and local energy maxima. Optik 205:163947
Qian J, Tiwari P, Gochhayat SP, Pandey HM (2020) A noble double dictionary based ecg compression technique for ioth. IEEE Intern Things J
Rong S, Zhou H, Zhao D, Cheng K, Qian K, Qin H (2018) Infrared x pattern noise reduction method based on shearlet transform. Infrared Phys Technol 91:243–249
Tan W, Xiang P, Zhang J, Zhou H, Qin H (2020) Remote sensing image fusion via boundary measured dualchannel pcnn in multiscale morphological gradient domain. IEEE Access 8:42540–42549
Tan W, Zhang J, Xiang P, Zhou H, Thitøn W (2020) Infrared and visible image fusion via nsst and pcnn in multiscale morphological gradient domain. In: Optics, photonics and digital technologies for imaging applications VI, vol 11353, p 113531E. International society for optics and photonics
Tan W, Zhou H, Rong S, Qian K, Yu Y (2018) Fusion of multifocus images via a Gaussian curvature filter and synthetic focusing degree criterion. Appl Opt 57(35):10092–10101
Tan W, Zhou H, Song J, Li H, Yu Y, Du J (2019) Infrared and visible image perceptive fusion through multilevel Gaussian curvature filtering image decomposition. Appl Opt 58(12):3064–3073
Tan W, Zhou Hx, Yu Y, Du J, Qin H, Ma Z, Zheng R (2017) Multifocus image fusion using spatial frequency and discrete wavelet transform. In: AOPC 2017: Optical sensing and imaging technology and applications, vol 10462, p 104624 K. International society for optics and photonics
Tiwari P, Melucci M (2018) Towards a quantuminspired framework for binary classification. In: Proceedings of the 27th ACM international conference on information and knowledge management, pp 1815–1818
Tiwari P, Melucci M (2019) Binary classifier inspired by quantum theory. Proc AAAI Conf Artif Intell 33:10051–10052
Tiwari P, Melucci M (2019) Towards a quantuminspired binary classifier. IEEE Access 7:42354–42372
Tiwari P, Qian J, Li Q, Wang B, Gupta D, Khanna A, Rodrigues JJ, de Albuquerque VHC (2018) Detection of subtype blood cells using deep learning. Cogn Syst Res 52:1036–1044
Wang Z, Wang S, Guo L (2018) Novel multifocus image fusion based on PCNN and random walks. Neural Comput Appl 29(11):1101–1114
Yin H (2018) Tensor sparse representation for 3D medical image fusion using weighted average rule. IEEE Trans Biomed Eng 65(11):2622–2633
Yin M, Liu X, Liu Y, Chen X (2018) Medical image fusion with parameteradaptive pulse coupled neural network in nonsubsampled shearlet transform domain. IEEE Trans Instrum Meas 68(1):49–64
Zhang Q, Guo BL (2009) Multifocus image fusion using the nonsubsampled contourlet transform. Signal Process 89(7):1334–1346
Zhang Y, Bai X, Wang T (2017) Boundary finding based multifocus image fusion through multiscale morphological focusmeasure. Inf Fusion 35:81–101
Zhou Z, Wang B, Li S, Dong M (2016) Perceptual fusion of infrared and visible images through a hybrid multiscale decomposition with Gaussian and bilateral filters. Inf Fusion 30:15–26
Zhu Z, Zheng M, Qi G, Wang D, Xiang Y (2019) A phase congruency and local Laplacian energy based multimodality medical image fusion method in NSCT domain. IEEE Access 7:20811–20824
Pandey HM, Windridge D (2019) A comprehensive classification of deep learning libraries. In: Third international congress on information and communication technology. Springer, Singapore
Acknowledgements
The authors are grateful to the editors and the reviewers for their valuable comments and suggestions, the Whole Brain Atlas for providing the datasets, and Dr. Mengxue Zheng’s guidance on analyzing medical images. This study is supported by China Scholarship Council (CSC201906960047) and 111 Project (B17035).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Tan, W., Tiwari, P., Pandey, H.M. et al. Multimodal medical image fusion algorithm in the era of big data. Neural Comput & Applic (2020). https://doi.org/10.1007/s00521020051732
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s00521020051732
Keywords
 Multimodal medical imaging
 Medical image fusion
 Pulsecoupled neural network
 Nonsubsampled shearlet transform