paper - 1 - 副本
paper - 1 - 副本
Research Article
Multimodal Medical Image Fusion Based on Multiple Latent
Low-Rank Representation
Received 18 June 2021; Revised 31 August 2021; Accepted 2 September 2021; Published 29 September 2021
Copyright © 2021 Xi-Cheng Lou and Xin Feng. This is an open access article distributed under the Creative Commons
Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work
is properly cited.
A multimodal medical image fusion algorithm based on multiple latent low-rank representation is proposed to improve
imaging quality by solving fuzzy details and enhancing the display of lesions. Firstly, the proposed method decomposes the
source image repeatedly using latent low-rank representation to obtain several saliency parts and one low-rank part.
Secondly, the VGG-19 network identifies the low-rank part’s features and generates the weight maps. Then, the fused low-
rank part can be obtained by making the Hadamard product of the weight maps and the source images. Thirdly, the
fused saliency parts can be obtained by selecting the max value. Finally, the fused saliency parts and low-rank part are
superimposed to obtain the fused image. Experimental results show that the proposed method is superior to the traditional
multimodal medical image fusion algorithms in the subjective evaluation and objective indexes.
LatLRR
I LX XZ E
LatLRR
X LX XZ E
Superimpose
LatLRR
X1 LX1 X1Z E
rearranges these patches to sparse coefficient vectors, which based algorithm’s dictionary matrix cannot fully include
are the linear combination of vectors in the dictionary matrix. source image data, it fails to extract the source image’s detailed
Then, the fused image’s sparse coefficient vectors can be deter- texture information. Some scholars applied MST [24–26] or
mined through maximal l1 -norm, rearrange these vectors to filter [27–29] to decompose the source images. And SR can
patches of fused image, and put these patches return to the seat be used to fuse the low-frequency subbands. Theoretically,
can obtain the fused image. In terms of edge feature extraction, such methods can preserve the edge information of the image
SR has certain advantages over MST. Many improved versions better than using SR [30] to decompose the source images. Liu
of SR have appeared in recent years to increase computational et al. [31] proposed the low-rank representation (LRR), which
efficiency or improve fusion quality. Liu and Wang [21] pro- applies the source image as the dictionary matrix, and can
posed adaptive sparse representation (ASR), with seven sub- solve dictionary completeness. Liu and Yan [32] proposed
dictionaries trained in advance to match the patches of the latent low-rank representation (LatLRR), an improved ver-
images categorized by the gradient. Liu et al. [22] proposed sion of LRR, which can decompose the source image to
convolutional sparse representation (CSR), which does not saliency part, low-rank part, and noise part. Li et al. [33] pro-
use the sliding window to decompose the source image but posed MDLatLRR, which integrated SR and LatLRR by using a
applies a globe process. Liu et al. [23] proposed convolutional sliding window to sample the saliency parts of LatLRR and
sparsity-based morphological component analysis (CSMCA), processed the sparse coefficient vectors just like in SR-based
which simultaneously achieves multicomponent and global methods. LatLRR has an extraordinary capacity for extracting
SR by integrating CSR and the morphological component texture from the image, but the ability to extract high-
analysis into a unified framework. However, since the SR- frequency information is not as good as MST.
Computational and Mathematical Methods in Medicine 3
Conv3-64
Conv3-64
Pool1
Conv3-128
IL1 Conv3-128 W1
Pool2
Conv3-256
Conv3-256
Conv3-256
IL2 Pool3 W2
Pool4 C1 C2
Soft-max Conv3-512
Eq. (10)
FC1000 Conv3-512
Another type of image fusion method that is more widely CNN to generate weight maps and use the Laplacian pyramid
used is with the help of weighted maps [34, 35], and deep and Gaussian pyramid to decompose the source images and
learning-based methods are particularly suitable for generating weight maps, respectively, though the Hadamard product to
weighted maps due to their superior feature recognition capa- obtain fused layers, and the fused image can be reconstructed
bilities. The deep learning-based methods [36–42] have been through the Laplacian pyramid. Li et al. [39] applied the aver-
widely used in image fusion with the development of artificial age filter to decompose the source images, and the fused base
intelligence. These methods have a prominent ability to extract layer can be obtained by comparing the max absolute value of
feature information from the image. Therefore, it is wise to use images in four convolutional layers of the VGG-19 neural net-
deep learning-based methods to deal with high-frequency infor- work. The fused image can be obtained by superimposing the
mation after image decomposition or generate the weight map base layer and detail layer. Yin et al. [40], Tan et al. [41], and
as the basis of image region fusion. Wang et al. [36] applied a Panigrahy et al. [42] applied nonsubsampled shearlet trans-
convolutional neural network (CNN) to generate weight maps form (NSST) to decompose the source images and selected
and decompose the source images and weight maps by contrast the fused high-frequency subbands by more firing times in
pyramid and Laplacian pyramid, respectively, and make the the parameter-adaptive pulse coupled neural network
Hadamard product of each decomposition layer. Finally, the (PAPCNN), bounded measured pulse coupled neural network
fused image can be reconstructed through the contrast pyramid. (BMPCNN), and weighted parameter adaptive dual-channel
Xu et al. [37] applied LatLRR to decompose source images and pulse coupled neural network (WPADCPCNN), respectively.
processed the low-rank parts by CNN and pyramid-based As mentioned above, each method has its drawbacks and
methods, superimposed the fused low-rank part, and fused advantages. In this paper, the source images are repeatedly
saliency part to obtain the fused image. Liu et al. [38] applied decomposed through LatLRR to extract the saliency parts.
4 Computational and Mathematical Methods in Medicine
YUV space
CT
Proposed fusion
method
RGB space RGB space
Y Fused Y
The fused saliency parts can be obtained by selecting the max 2. Multiple Latent Low-Rank Representation
value. After superimposing these saliency parts, the structure
and edge information of the source images will be well pre- LatLRR is an improved version of LRR, whose principle is
served and enhance the lesion’s display. Then, the VGG-19 date X = fx1 , x2 ,⋯,xM g in space Rn can be represented by
network is used to extract features of the low-rank part, and a linear combination of vectors in an overcomplete dictio-
the weight maps can be generated to be the basis of the low- nary D ∈ Rn×m ðn < mÞ, as
rank part’s activity level. The weight maps and low-rank parts
then make the Hadamard product to obtain the fused low- X = DZ, ð1Þ
rank part. Finally, the fused image can be obtained by super-
imposing the fused saliency part and fused low-rank part. where Z = fz1 , z2 ,⋯,zM g is the coefficient matrix in space
The experimental results also show that the proposed method Rm ; it can be determined through
is significantly better than the comparison method regarding
image information retention. The main contributions of this min kZ k∗ s:t:X = DZ, ð2Þ
Z
paper are described as follows:
where k·k∗ denotes the nuclear norm. The idea of this algo-
(1) The proposed method applies the image detail rithm is similar to that of SR, in that it finds the coefficients
retention capability of LatLRR while fully extract- of an image under certain dictionary conditions. LRR using
ing the high-frequency information of an image date X itself as the dictionary, just like equation (3), that is
by iteratively decomposing the original image. It the reason LRR does not have the problems of dictionary
compensates for the deficiency of LatLRR and training or completeness.
enhances the display of the lesion by superimpos-
ing saliency parts min kZk∗ s:t:X = XZ: ð3Þ
Z
(2) The feature map of the low-rank part of the original
image is extracted using the VGG-19 network and The noise component E is added in equation (3); this is
then scaled up to match the size of the original because the original purpose of creating the low-rank algo-
image. The weight map generated in this way can rithm is to remove noise from the image. And Equation
well fit the low-rank part of the original image with (4) is the formula of LRR.
pixel information blockwise distributed
min kZk∗ + λkEk1,2 s:t:X = XZ + E, ð4Þ
The rest of this paper is organized as follows. Section 2 Z ,E
introduces the multiple LatLRR decomposition algorithm,
Section 3 introduces the fusion rules, Section 4 describes where λ > 0 is balance coefficient and kEk1,2 denotes the l1,2
the algorithmic structure of the proposed method, Section -norm of E.
5 provides a detailed discussion of the experimental results, However, there are two prerequisites for using X itself as a
and Section 6 concludes this paper. dictionary. One is that the data vector of X must be sufficiently
Computational and Mathematical Methods in Medicine 5
YUV
Color image? Yes space
No Y
I1 I2
U
V
Multiple LatLRR decomposition (N layers)
I1L I2L
I1S,N I2S,N
Max-rule VGG-19
Superpose
Superpose
IFS,N IL
IS
YUV convert
RGB
to RGB
IF Color fused
image
complete. Second, the noise of X must be controlled in a small equation (7) can be obtained.
range. In many practical conditions, such requirements are
challenging to achieve. For this reason, the method of adding min kZk∗ + kLk∗ + λkEk1 s:t:X = XZ + LX + E, ð7Þ
hidden items in the dictionary is proposed in [32] Z ,L ,E
Input:I1 and I2 .
Output: fused image I F
/∗ Part 1: N layer multiple LatLRR decomposition. ∗/
1 for each M = ½I1 , I2 do
2 for each i = ½1 : Ndo
3 Perform LatLRR decomposition on S to obtain fIS,i L S,i L
1 , I1 g and fI2 , I2 g.
4 end
5 end
/∗ Part 2: Fusion of saliency part. ∗/
6 for each i = ½1 : Ndo
7 perform max-rule on fIS,i S,i i
1 , I2 g to obtain IS as euqation (8).
8 end
N
9 Superimpose fIiS gi=1 to obtain the final fused saliency part IS as equation (9).
/∗ Part 3: Fusion of low-rank part.∗/
10 for each k = ½1, 2do
512
11 Input ILk to VGG-19 network to obtain fϕm k gm=1 extract from the 5th layer of the network;
m 512
12 Solve the l1 -norm of fϕk gm=1 to obtain activity level map Ck as equation (10);
13 Calculate the initial weight map W ̂ k by equation (11);
14 Enlarge W ̂ k to get the final weight map Wk as equation (12).
15 end
16 Calculate the final fused low-rank part IL as equation (13).
/∗Part 4: Reconstruction.∗/
17 Superimpose the fused saliency part IS and the fused low-rank part IL to obtain the fused image I F as equation (14).
may filter out some critical information. In this case, a reason- diagnostic information as much as possible. On the other
able approach is to superimpose the noise part and low-rank hand, Simonyan and Zisserman [44] first applied the VGG
part. Moreover, the saliency part LX contains most edge and network to extract features at different layers from images
structure information of the image so that the lesions may and obtain a splendid result. With the development of deep
mainly reflect in the saliency part. If the low-rank part of the learning, the operation efficiency and precision of the VGG
image is decomposed repeatedly, the saliency part will be fur- network have been significantly improved. As the number
ther extracted. As shown in Figure 2, a two-layer LatLRR of LatLRR decomposition layers increases, the source
decomposition structure, after completing the first layer of image’s low-rank part will contain less information. If the
LatLRR decomposition, the new object of LatLRR decomposi- recognition results of VGG-19 are extracted and processed,
tion could be obtained through X1 = XZ + E, which can be fur- the weight map with regional emphasis can be generated.
ther decomposed into saliency part LX1 , low-rank part X1 Z, Then, the fused low-rank part can be obtained by multiply-
and noise part E. The source images are denoted as I1 and I2. ing the weight map with the source image’s low-rank part.
If the number of LatLRR decomposition layers is N, there will Besides, because PET and SPECT images are in color, they
N N
be N saliency parts fIS1,i gi=1 or fIS2,i gi=1 and one low-rank part need to be converted into YUV color space before fusing
IL1 or IL2 for each source image. The display of the edge and them with the grayscale image as MRI.
structure information will be strengthened in the new image
to highlight the lesions by superimposing saliency parts. How- 3.1. Fusion of Saliency Parts. Each LatLRR decomposition of
ever, suppose the number N of LatLRR decomposition layers the two source images will produce a saliency part of each.
is blindly increased, which will reduce the efficiency of calcula- By adopting max-rule for all saliency parts of N layers of
tion. More importantly, the final fused image will display some LatLRR decomposition, N fused significant parts can be
artificial information unacceptable for medical images. In this obtained, as
paper, the optimal number of LatLRR decomposition layers will
be determined through the experiment in Section 5.1.
i N h N N i
IS i=1 ðx, yÞ = max IS1,i i=1 ðx, yÞ, IS2,i i=1 ðx, yÞ , ð8Þ
3. Fusion Regulation
The saliency parts of the image include most high-frequency N
information. For multimodal medical images, the critical where fIiS gi=1 ðx, yÞ denote the position ðx, yÞ of the ith
N N
diagnostic information reflected by a single image is not layer saliency part of fused image fIiS gi=1 , so as fIS1,i gi=1 ðx,
the same. Therefore, the max-rule is applied to fuse the N
saliency parts of the image can preserve a single image’s yÞ and fIS2,i gi=1 ðx, yÞ. The final fused saliency part can be
Computational and Mathematical Methods in Medicine 7
MR-PD CT MR-T1 CT
(a) (b) (c) (d)
MR-T2 SPECT-Tc
(i) (j)
Figure 6: Five sets of multimodal medical images for the experiment. (a) MR-PD. (b) CT. (c) MR-T1. (d) CT. (e) MR-T2. (f) SPECT-Tl. (g)
MR-T2. (h) PET-FDG. (i) MR-T2. (j) SPECT-Tc.
512
calculated by maps, where k ∈ f1, 2g. The l1 -norm of fϕmk gm=1 ðx, yÞ could
be the activity level measure of the low-rank part. So, the
N activity level map Ck can be calculated by
IS = 〠 IiS : ð9Þ
i=1 512
Ck ðx, yÞ = fϕm g
k m=1 ð x, y Þ : ð10Þ
1
3.2. Fusion of Low-Rank Parts. VGG-19 is a convolutional
neural network that is 19 layers deep, including 16 convolu- ̂ k can be obtained by
Then, the initial weight map W
tional layers and 3 fully connected layers. Its structure is
shown in Figure 3. Each convolutional layer is denoted as ̂ k ðx, yÞ = Ck ðx, yÞ
W : ð11Þ
conv‘size of the filter’-‘number of such filters,’ and max- C1 ðx, yÞ + C2 ðx, yÞ
pooling layers have 2 × 2 filter with the stride of 2. The infor-
mation of the low-rank part of the image after multiple
As feature maps are only ð0:5Þ5 , the size of the source
LatLRR decompositions is relatively fuzzy and presents a ̂ k , which is generated by
regional-like distribution. According to this feature, the fea- image, so the initial weight map W
5
ture maps extracted from the fifth convolution layer of feature maps, is ð0:5Þ the size of the source image too. For
VGG-19 can match the low-rank part of the image’s infor- matching the size of the source image, W ̂ k need the upsam-
mation distribution state after amplification. pling procedure as
512 m 512
For low-rank parts IL1 and IL2 , fϕm 1 gm=1 and fϕ2 gm=1
denote the feature maps extracted from the fifth convolu- ̂ k ðx + p, y + qÞ
Wk ðx, yÞ = W p, q ∈ f1, 2, ⋯ ,15g: ð12Þ
tional layer of VGG-19. As shown in Figure 3, the 5th con-
volutional layer is conv3-512, so there are 512 feature The fused low-rank part IL can be calculated by
maps of each low-rank part. Moreover, because of max-
pooling layers, these feature maps are only ð0:5Þ5 the size IL = W1 ∘ IL1 + W2 ∘ IL2 , ð13Þ
512
of the source image. According to [39], let fϕm k gm=1 ðx, yÞ
denote the ðx, yÞ position of the kth low-rank part’s feature where ∘ denotes the Hadamard product.
8 Computational and Mathematical Methods in Medicine
0.8000 0.7000
0.7000 0.6500
Value of QG
Value of QTE 0.6000 0.6000
0.5000 0.5500
0.4000 0.5000
0.3000 0.4500
1 2 3 4 1 2 3 4
(a) (b)
0.7350 0.6500
0.7100 0.6000
Value of QCB
Value of Qc
0.6850 0.5500
0.6600 0.5000
0.6350 0.4500
0.6100 0.4000
1 2 3 4 1 2 3 4
Figure 7: Parametric experimental results. (a) QTE of five sets of images. (b) QG of five sets of images. (c) QC of five sets of images. (d) QCB of
five sets of images.
Figure 8: The fusion result of each LatLRR decomposition level. (a) 1 layer. (b) 2 layers. (c) 3 layers. (d) 4 layers. (e) 1 layer. (f) 2 layers. (g) 3
layers. (h) 4 layers. (i) 1 layer. (j) 2 layers. (k) 3 layers. (l) 4 layers.
3.3. YUV Color Space. For color images such as SPECT and converted to YUV space and decomposed into one lumi-
PET, Yin et al. [40] proposed a YUV space to solve color and nance component, ‘Y’ and two chrominance components,
grayscale images’ fusion problems. The color image is first ‘U’ and ‘V.’ Then, the ‘Y’ component of the color image
Computational and Mathematical Methods in Medicine 9
Figure 9: Comparison of the fusion results in the first set images. (a) MR-PD. (b) CT. (c) CSR. (d) ASR. (e) CSMCA. (f) CNN. (g)
BMPCNN. (h) MDLatLRR. (i) VGG-19. (j) NSCT_SR. (k) LP_SR. (l) Proposed method.
Figure 10: Comparison of the fusion results in the first set images. (a) MR-T1. (b) CT. (c) CSR. (d) ASR. (e) CSMCA. (f) CNN. (g)
BMPCNN. (h) MDLatLRR. (i) VGG-19. (j) NSCT_SR. (k) LP_SR. (l) Proposed method.
5.1. Parametric Experiment. In order to determine the image. The larger the QC is, the better the structure of the
decomposition layers of LatLRR in this paper, five sets of source images is preserved. The calculation process of QCB
images in Figure 6 were fused by the proposed method. is complex and consists of five steps: contrast sensitivity fil-
The results are objectively evaluated by four fusion image tering, local contrast computation, contrast preservation cal-
evaluation indexes: fusion metric-based on Tsallis entropy culation, saliency map generation, and global quality map
(QTE ) [46], gradient-based fusion performance (QG ) [47], computation. QCB takes the mean value of the global quality
image structural similarity metric (QC ) [48], and human map. The larger the QCB value is, the richer the contrast
perception inspired fusion metric (QCB ) [49]. QTE is a diver- information of the fused image is.
gence measure of the degree of dependence between two dis- The test decomposition layers of LatLRR are set from 1
crete random variables, and it calculates information from to 4. Parameter experimental results of indexes of four sets
the source images is transferred to the fused image. There- of images are shown in Figures 7(a)–7(d). It can be observed
fore, the larger the QTE value, the better the fusion effect. that as the number of LatLRR decomposition layers
QG uses the Sobel edge operator to calculate the intensity increases, not all indexes show a uniform trend. The value
and direction information of the edges in the source image of QTE increased with the increase of decomposition layers,
and the fused image. The larger the QG value is, the richer while QCB are optimal in the case of one-level decomposi-
the edge information of the fused image is. QC is used to tion. As for QG and QC , the changing trend is related to
measure the preservation degree of structure of the fused the image set. It is reasonable because the more decomposi-
image, so it calculates how much of the salient information tion layers, the image background information contained in
in each source image has been transferred into the fused the low-rank part will be fuzzier and more contained in the
Computational and Mathematical Methods in Medicine 11
Figure 11: Comparison of the fusion results in the second set images. (a) MR-T2. (b) SPECT-T1. (c) CSR. (d) ASR. (e) CSMCA. (f) CNN.
(g) BMPCNN. (h) MDLatLRR. (i) VGG-19. (j) NSCT_SR. (k) LP_SR. (l) Proposed method.
saliency part. The max-rule selects the saliency part of the as possible, and artifacts should be strictly controlled. There-
fused image, so the background information of the source fore, the decomposition layer of LatLRR in this paper is set
images may also be strengthened in the fused image, as two.
enhancing the appearance of the lesion. If the greater the
amount of information, the larger the QTE value will be. 5.2. Contrast Experiment. Nine typical multimodal medical
However, strengthening image background information will image fusion methods are selected to compare with the pro-
weaken the boundary observation and reduce the image posed method: four sorts of SR-based methods, CSR [22],
contrast, undoubtedly leading to a lower QCB . Medical image ASR [21], CSMCA [23], and MDLatLRR [33]; three sorts
fusion aims to show the information of lesions in the fused of deep learning-based algorithm, CNN [36], VGG-19 [39],
image, so the QTE is more impotent than the other three and BMPCNN [41]; and two sorts of MST integrated SR
indexes. It can be seen from Figure 7(a) that QTE of two methods are nonsubsampled contourlet transform (NSCT)
decomposition layers is significantly improved than that of combined with SR (NSCT_SR) [24] and Laplacian pyramid
one decomposition layer. Still, the more decomposition (LP) combined with SR (LP_SR) [24]. The MST decomposi-
layers could not considerably improve the QTE . tion level is set to 4.
Besides, as Figure 8 shows, it can be seen that with the Figure 9 shows the fusion results of the first set images,
increase of the number of LatLRR decomposition layers, Figures 9(a) and 9(b) are the source images. It can be seen
the artifact around the object will be aggravated in several from Figure 9(l) of the proposed method, both the brain tis-
sets of images. In order to improve the image fusion quality, sue texture in the green box and the lesion edge in the red
the lesion in the fused image should be highlighted as much box are the clearest from other fusion methods. Besides,
12 Computational and Mathematical Methods in Medicine
Figure 12: Comparison of the fusion results in the third set images. (a) MR-T2. (b) PET-FDG. (c) CSR. (d) ASR. (e) CSMCA. (f) CNN. (g)
BMPCNN. (h) MDLatLRR. (i) VGG-19. (j) NSCT_SR. (k) LP_SR. (l) Proposed method.
the pixel consistency of bone in the proposed method is the serious color distortion, as can be seen in Figures 12(c),
best. Figure 10 shows the fusion results of the second set 12(d), and 12(j), the color rendering of the proposed method
images; Figures 10(a) and 10(b) are the source images. As is closest to the source image. Figure 13 shows the fusion
can be seen in Figure 10(l), the pixel consistency of the skel- results of the fourth set images, Figures 13(a) and 13(b) are
etal structure of the fused images is the best, and the widen- the source images. Compared with Figure 13(b), the graphic
ing of the brain tissue sulcus in the green box, as well as the structure of metabolic abnormalities (red box) in other
calcified lesions and edema in the red box, are also clearly fusion methods has been deformed to a certain extent. The
visible. Figure 11 shows the fusion results of the second set structure is kept intact in the proposed method, and the
images; Figures 11(a) and 11(b) are the source images. In chromatic aberration is most consistent with the source
Figure 11(l) of the proposed method, the boundary of the image.
metabolic abnormality in the red box is the clearest, and All the fusion images are evaluated by four indexes intro-
the chromatic aberration is most consistent with duced in Section 5.1. As shown in Tables 1–5, in all sets of
Figure 11(b). The texture information in the green box is images, the proposed method leads in QTE , especially in
also clearly visible. Figure 12 shows the fusion results of the first set of images, the lead is more than 30 percent. It
the third set images; Figures 12(a) and 12(b) are the source means that the proposed method is superior to other
images. In Figure 12(l) of the proposed method, the bound- methods in the information conversion of source images.
ary of the metabolic abnormality in the red box is the most Moreover, for the first set and the fourth set of images, the
distinct, and the texture in the green box also holds the best. proposed method also leads in the QC . And it indicates that
Besides, in this set of images, some fusion methods appear the proposed method can retain the structure of the source
Computational and Mathematical Methods in Medicine 13
Figure 13: Comparison of the fusion results in the fourth set images. (a) MR-T2. (b) SPECT-Tc. (c) CSR. (d) ASR. (e) CSMCA. (f) CNN. (g)
BMPCNN. (h) MDLatLRR. (i) VGG-19. (j) NSCT_SR. (k) LP_SR. (l) Proposed method.
Table 1: Objective evaluation of fusion methods in the first set of Table 2: Objective evaluation of fusion methods in the second set
images. of images.
Fusion method QTE QG QC QCB t/s Fusion method QTE QG QC QCB t/s
Proposed method 0.4359 0.5145 0.6436 0.5036 158.7860 Proposed method 0.4495 0.5647 0.6995 0.6061 50.6440
CSR 0.3096 0.6608 0.5621 0.7137 135.2650 CSR 0.3911 0.6363 0.7000 0.7454 33.1950
ASR 0.2976 0.6700 0.6121 0.6622 210.1220 ASR 0.3595 0.7122 0.7532 0.7328 64.9900
CSMCA 0.2945 0.5830 0.5587 0.6103 372.3040 CSMCA 0.3491 0.6315 0.7090 0.6974 78.8380
CNN 0.3247 0.6267 0.5560 0.5916 32.7860 CNN 0.4131 0.4757 0.6533 0.5848 13.8630
BMPCNN 0.3058 0.5554 0.5828 0.5943 67.0890 BMPCNN 0.3726 0.5354 0.6918 0.6210 15.7120
MDLatLRR 0.2882 0.6527 0.6077 0.6784 69.0270 MDLatLRR 0.3657 0.6516 0.7108 0.7357 21.1340
VGG-19 0.2835 0.5104 0.5569 0.6273 7.4480 VGG-19 0.4211 0.4030 0.6457 0.3627 5.1570
NSCT_SR 0.2953 0.6243 0.5895 0.6353 7.4890 NSCT_SR 0.3725 0.5514 0.6886 0.6389 2.2760
LP_SR 0.2948 0.6560 0.5734 0.6588 0.3380 LP_SR 0.3582 0.6025 0.6741 0.6521 0.2210
images better than other methods. On the other hand, the two indicators. As mentioned in Section 5.1, QG and QCB
proposed method has no advantage or even a considerable are mainly used to measure the degree to which the fused
gap over the best-performing method in terms of the other image retains the edge information of the source images.
14 Computational and Mathematical Methods in Medicine
Table 3: Objective evaluation of fusion methods in the third set of of CSR show undeniable advantages in the contrast experiment
images. of third set images. QTE measures the degree to which the fusion
image retains the information of the source images, and the dis-
Fusion method QTE QG QC QCB t/s tortion of image structure or color will lower this index. Image
Proposed method 0.7297 0.6602 0.6249 0.4645 233.0850 artifacts and structural distortions may mislead medical profes-
CSR 0.5025 0.8111 0.7920 0.7010 130.2410 sionals, so it is not advisable to blindly pursue high image struc-
ASR 0.4632 0.7901 0.7755 0.6611 188.6130 ture indexes. The proposed method has obvious advantages in
CSMCA 0.4576 0.7319 0.6790 0.5979 390.8500 the QTE index, and the degree of color distortion and artifact
CNN 0.7183 0.7822 0.7530 0.4966 38.6490
in the fused images is minimal. It is crucial for medical images.
Because artifacts in brain images can sometimes look very sim-
BMPCNN 0.3832 0.7900 0.8470 0.6505 75.5030
ilar to lesions, false lesions in images can directly affect the judg-
MDLatLRR 0.4303 0.7755 0.7298 0.6416 73.5810 ment of medical professionals. In addition, SPECT and PET
VGG-19 0.4828 0.7026 0.6891 0.6179 7.8900 show metabolic abnormalities through chromatic aberration.
NSCT_SR 0.4659 0.8054 0.8200 0.6757 9.2580 If significant color distortions appear in the image, it will cause
LP_SR 0.5044 0.7994 0.8120 0.5991 0.5030 deviation from the real. Based on the above analysis, the pro-
posed method is effective.
Program of Chongqing Education Commission of China [14] S. Kollem, K. R. Reddy, and D. S. Rao, “Improved partial differ-
(Grant Nos. KJQN201900821 and KJQN202000803), Inno- ential equation-based total variation approach to non-
vative Research Group of Universities in Chongqing (Grant subsampled contourlet transform for medical image denois-
No. CXQT21024), and Graduate Innovation Project of ing,” Multimedia Tools and Applications, vol. 80, no. 2,
Chongqing Technology and Business University (Grant pp. 2663–2689, 2021.
No. yjscxx2021-112-45). [15] J.-Y. Li and C.-Z. Zhang, “Blind watermarking scheme based
on Schur decomposition and non-subsampled contourlet
transform,” Multimedia Tools and Applications, vol. 79,
no. 39-40, pp. 30007–30021, 2020.
References
[16] T. Chu, Y. Tan, Q. Liu, and B. Bai, “Novel fusion method for
[1] D. L. Bailey, M. N. Maisey, D. W. Townsend, and P. E. Valk, SAR and optical images based on non-subsampled shearlet
Positron Emission Tomography, vol. 2, Springer, 2005. transform,” International Journal of Remote Sensing, vol. 41,
[2] Z. Wang, Z. Cui, and Y. Zhu, “Multi-modal medical image no. 12, pp. 4590–4604, 2020.
fusion by Laplacian pyramid and adaptive sparse representa- [17] H. Ullah, B. Ullah, L. Wu, F. Y. Abdalla, G. Ren, and
tion,” Computers in Biology and Medicine, vol. 123, article Y. Zhao, “Multi-modality medical images fusion based
103823, 2020. on local-features fuzzy sets and novel sum-modified-
[3] J. Du, W. Li, B. Xiao, and Q. Nawaz, “Union Laplacian pyra- Laplacian in non-subsampled shearlet transform domain,”
mid with multiple features for medical image fusion,” Neuro- Biomedical Signal Processing and Control, vol. 57, article
computing, vol. 194, pp. 326–339, 2016. 101724, 2020.
[4] F. Liu, L. Chen, L. Lu, A. Ahmad, G. Jeon, and X. Yang, “Med- [18] S. Goyal, V. Singh, A. Rani, and N. Yadav, “FPRSGF denoised
ical image fusion method by using Laplacian pyramid and con- non-subsampled shearlet transform-based image fusion using
volutional sparse representation,” Concurrency and sparse representation,” Signal, Image and Video Processing,
Computation: Practice and Experience, vol. 32, no. 17, article vol. 14, no. 4, pp. 719–726, 2020.
e5632, 2020. [19] B. Yang and S. Li, “Multifocus image fusion and restoration
[5] J. Bhardwaj, A. Nayak, and D. Gambhir, “Multimodal medical with sparse representation,” IEEE Transactions on Instrumen-
image fusion based on discrete wavelet transform and genetic tation and Measurement, vol. 59, no. 4, pp. 884–892, 2010.
algorithm,” in International Conference on Innovative Com- [20] B. A. Olshausen and D. J. Field, “Emergence of simple-cell
puting and Communications, pp. 1047–1057, Singapore, 2021. receptive field properties by learning a sparse code for natural
[6] L. Zhan, Y. Zhuang, and L. Huang, “Infrared and visible images,” Nature, vol. 381, no. 6583, pp. 607–609, 1996.
images fusion method based on discrete wavelet transform,” [21] Y. Liu and Z. Wang, “Simultaneous image fusion and denois-
Journal of Computers, vol. 28, pp. 57–71, 2017. ing with adaptive sparse representation,” IET Image Process-
[7] S. K. Panguluri and L. Mohan, “Discrete wavelet transform ing, vol. 9, no. 5, pp. 347–357, 2015.
based image fusion using unsharp masking,” Periodica Poly- [22] Y. Liu, X. Chen, R. K. Ward, and Z. Jane Wang, “Image fusion
technica Electrical Engineering and Computer Science, vol. 64, with convo-lutional sparse representation,” IEEE Signal Pro-
no. 2, pp. 211–220, 2020. cessing Letters, vol. 23, no. 12, pp. 1882–1886, 2016.
[8] A. Seal, D. Bhattacharjee, M. Nasipuri, D. Rodriguez-Esparra- [23] Y. Liu, X. Chen, R. K. Ward, and Z. J. Wang, “Medical image
gon, E. Menasalvas, and C. Gonzalo-Martin, “PET-CT image fusion via convolutional sparsity based morphological compo-
fusion using random forest and à-trous wavelet transform,” nent analysis,” IEEE Signal Processing Letters, vol. 26, no. 3,
International journal for numerical methods in biomedical pp. 485–489, 2019.
engineering, vol. 34, no. 3, article e2933, 2018. [24] Y. Liu, S. Liu, and Z. Wang, “A general framework for image
[9] K. Seethalakshmi and S. Valli, “A fuzzy approach to recognize fusion based on multi-scale transform and sparse representa-
face using contourlet transform,” International Journal of tion,” Information fusion, vol. 24, pp. 147–164, 2015.
Fuzzy Systems, vol. 21, no. 7, pp. 2204–2211, 2019. [25] L. Tan and X. Yu, “Medical image fusion based on fast finite
[10] M. Kumar, N. Ranjan, and B. Chourasia, “Hybrid methods of shearlet transform and sparse representation,” Computational
contourlet transform and particle swarm optimization for and Mathematical Methods in Medicine, vol. 2019, Article ID
multimodal medical image fusion,” in 2021 International Con- 3503267, 14 pages, 2019.
ference on Articial Intelligence and Smart Systems (ICAIS), [26] J. Xia, Y. Chen, A. Chen, and Y. Chen, “Medical image fusion
pp. 945–951, Coimbatore, India, 2021. based on sparse representation and PCNN in NSCT domain,”
[11] W. Wu, S. Guo, and Q. Cheng, “Fusing optical and synthetic Computational and Mathematical Methods in Medicine,
aperture radar images based on shearlet transform to improve vol. 2018, Article ID 2806047, 12 pages, 2018.
urban impervious surface extraction,” Journal of Applied [27] X. Li, F. Zhou, and H. Tan, “Joint image fusion and denoising
Remote Sensing, vol. 14, no. 2, article 024506, 2020. via three-layer decomposition and sparse representation,”
[12] A. Khare, M. Khare, and R. Srivastava, “Shearlet transform Knowledge-Based Systems, vol. 224, article 107087, 2021.
based technique for image fusion using median fusion rule,” [28] X. Li, F. Zhou, H. Tan, W. Zhang, and C. Zhao, “Multimodal
Multimedia Tools and Applications, vol. 80, no. 8, article medical image fusion based on joint bilateral filter and local
10184, pp. 11491–11522, 2021. gradient energy,” Information Sciences, vol. 569, pp. 302–325,
[13] M. Sayadi, H. Ghassemian, R. Naimi, and M. Imani, “A new 2021.
composite multimodality image fusion method based on [29] Z. Zhu, H. Yin, Y. Chai, Y. Li, and G. Qi, “A novel multi-
shearlet transform and retina inspired model,” in 2020 Inter- modality image fusion method based on image decomposition
national Conference on Machine Vision and Image Processing and sparse representation,” Information Sciences, vol. 432,
(MVIP), pp. 1–5, Qom, Iran, 2020. pp. 516–529, 2018.
16 Computational and Mathematical Methods in Medicine
[30] H. Li, Y. Wang, Z. Yang, R. Wang, X. Li, and D. Tao, “Discrim- [46] N. Cvejic, C. Canagarajah, and D. Bull, “Image fusion metric
inative dictionary learning-based multiple component decom- based on mutual information and Tsallis entropy,” Electronics
position for detail-preserving noisy image fusion,” IEEE Letters, vol. 42, no. 11, pp. 626-627, 2006.
Transactions on Instrumentation and Measurement, vol. 69, [47] C. Xydeas and V. Petrović, “Objective image fusion perfor-
no. 4, pp. 1082–1102, 2020. mance measure,” Electronics Letters, vol. 36, no. 4, pp. 308-
[31] G. Liu, Z. Lin, and Y. Yu, Robust Subspace Segmentation by 309, 2000.
Low-Rank Representation, vol. 1, Icml, 2010. [48] N. Cvejic, A. Loza, D. Bull, and N. Canagarajah, “A similarity
[32] G. Liu and S. Yan, “Latent low-rank representation for sub- metric for assessment of image fusion algorithms,” Interna-
space segmentation and feature extraction,” in 2011 Interna- tional journal of signal processing, vol. 2, pp. 178–182, 2005.
tional Conference on Computer Vision, pp. 1615–1622, [49] Y. Chen and R. S. Blum, “A new automated quality assessment
Barcelona, Spain, 2011. algorithm for image fusion,” Image and Vision Computing,
[33] H. Li, X.-J. Wu, and J. Kittler, “MDLatLRR: a novel decompo- vol. 27, no. 10, pp. 1421–1432, 2009.
sition method for infrared and visible image fusion,” IEEE
Transactions on Image Processing, vol. 29, pp. 4733–4746,
2020.
[34] A. Sengupta, A. Seal, C. Panigrahy, O. Krejcar, and A. Yazidi,
“Edge information based image fusion metrics using fractional
order differentiation and sigmoidal functions,” IEEE Access,
vol. 8, pp. 88385–88398, 2020.
[35] Z. Zhu, B. Zhu, H. H. T. Liu, and K. Qin, “A model-based
approach for measurement noise estimation and compensa-
tion in feedback control systems,” IEEE Transactions on
Instrumentation and Measurement, vol. 70, article 5001523,
pp. 1–23, 2020.
[36] K. Wang, M. Zheng, H. Wei, G. Qi, and Y. Li, “Multi-modality
medical image fusion using convolutional neural network and
contrast pyramid,” Sensors, vol. 20, no. 8, p. 2169, 2020.
[37] Z. Xu, W. Xiang, S. Zhu et al., “LatLRR-FCNs: latent low-rank
representation with fully convolutional networks for medical
image fusion,” Frontiers in Neuroscience, vol. 14, p. 1387, 2021.
[38] Y. Liu, X. Chen, J. Cheng, and H. Peng, “A medical image
fusion method based on convolutional neural networks,” in
2017 20th international conference on information fusion
(Fusion), pp. 1–7, Xi'an, China, 2017.
[39] H. Li, X.-J. Wu, and J. Kittler, “Infrared and visible image
fusion using a deep learning framework,” in 2018 24th interna-
tional conference on pattern recognition (ICPR), pp. 2705–
2710, Beijing, China, 2018.
[40] M. Yin, X. Liu, Y. Liu, and X. Chen, “Medical image fusion
with parameter adaptive pulse coupled neural network in non-
subsampled shearlet transform domain,” IEEE Transactions
on Instrumentation and Measurement, vol. 68, no. 1, pp. 49–
64, 2019.
[41] W. Tan, P. Tiwari, H. M. Pandey, C. Moreira, and A. K. Jais-
wal, “Multimodal medical image fusion algorithm in the era
of big data,” Neural Computing and Applications, pp. 1–21,
2020.
[42] C. Panigrahy, A. Seal, and N. K. Mahato, “MRI and SPECT
image fusion using a weighted parameter adaptive dual chan-
nel PCNN,” IEEE Signal Processing Letters, vol. 27, pp. 690–
694, 2020.
[43] Z. Lin, M. Chen, and Y. Ma, “The augmented lagrange multi-
plier method for exact recovery of corrupted low-rank matri-
ces,” 2010, https://arxiv.org/abs/1009.5055/.
[44] K. Simonyan and A. Zisserman, “Very deep convolutional net-
works for large-scale image recognition,” 2014, https://arxiv
.org/abs/1409.1556/.
[45] K. A. Johnson and J. A. Becker, “The whole brain atlas,” 2021,
http://www.med.harvard.edu/aanlib/.