
The angiography and color fundus images are of great assistance for the localization of central serous chorioretinopathy (CSCR) lesions. However, it brings much inconvenience to ophthalmologists because of these two modalities working independently in guiding laser surgery. Hence, a novel fundus image fusion method in non-subsampled contourlet transform (NSCT) domain, aiming to integrate the multi-modal CSCR information, is proposed. Specifically, the source images are initially decomposed into high-frequency and low-frequency components based on NSCT. Then, an improved deep learning-based method is employed for the fusion of low-frequency components, which helps to alleviate the tedious process of manually designing fusion rules and enhance the smoothness of the fused images. The fusion of high-frequency components based on pulse-coupled neural network (PCNN) is closely followed to facilitate the integration of detailed information. Finally, the fused images can be obtained by applying an inverse transform on the above fusion components. Qualitative and quantitative experiments demonstrate the proposed scheme is superior to the baseline methods of multi-scale transform (MST) in most cases, which not only implies its potential in multi-modal fundus image fusion, but also expands the research direction of MST-based fusion methods.
Citation: Jianguo Xu, Cheng Wan, Weihua Yang, Bo Zheng, Zhipeng Yan, Jianxin Shen. A novel multi-modal fundus image fusion method for guiding the laser surgery of central serous chorioretinopathy[J]. Mathematical Biosciences and Engineering, 2021, 18(4): 4797-4816. doi: 10.3934/mbe.2021244
[1] | Fang Zhu, Wei Liu . A novel medical image fusion method based on multi-scale shearing rolling weighted guided image filter. Mathematical Biosciences and Engineering, 2023, 20(8): 15374-15406. doi: 10.3934/mbe.2023687 |
[2] | Hao Wang, Guangmin Sun, Kun Zheng, Hui Li, Jie Liu, Yu Bai . Privacy protection generalization with adversarial fusion. Mathematical Biosciences and Engineering, 2022, 19(7): 7314-7336. doi: 10.3934/mbe.2022345 |
[3] | Keying Du, Liuyang Fang, Jie Chen, Dongdong Chen, Hua Lai . CTFusion: CNN-transformer-based self-supervised learning for infrared and visible image fusion. Mathematical Biosciences and Engineering, 2024, 21(7): 6710-6730. doi: 10.3934/mbe.2024294 |
[4] | Hao Chen, Shengjie Li, Xi Lu, Qiong Zhang, Jixining Zhu, Jiaxin Lu . Research on bearing fault diagnosis based on a multimodal method. Mathematical Biosciences and Engineering, 2024, 21(12): 7688-7706. doi: 10.3934/mbe.2024338 |
[5] | Shuaiyu Bu, Yuanyuan Li, Wenting Ren, Guoqiang Liu . ARU-DGAN: A dual generative adversarial network based on attention residual U-Net for magneto-acousto-electrical image denoising. Mathematical Biosciences and Engineering, 2023, 20(11): 19661-19685. doi: 10.3934/mbe.2023871 |
[6] | Xiao Zou, Jintao Zhai, Shengyou Qian, Ang Li, Feng Tian, Xiaofei Cao, Runmin Wang . Improved breast ultrasound tumor classification using dual-input CNN with GAP-guided attention loss. Mathematical Biosciences and Engineering, 2023, 20(8): 15244-15264. doi: 10.3934/mbe.2023682 |
[7] | Yang Pan, Jinhua Yang, Lei Zhu, Lina Yao, Bo Zhang . Aerial images object detection method based on cross-scale multi-feature fusion. Mathematical Biosciences and Engineering, 2023, 20(9): 16148-16168. doi: 10.3934/mbe.2023721 |
[8] | Tian Ma, Boyang Meng, Jiayi Yang, Nana Gou, Weilu Shi . A half jaw panoramic stitching method of intraoral endoscopy images based on dental arch arrangement. Mathematical Biosciences and Engineering, 2024, 21(1): 494-522. doi: 10.3934/mbe.2024022 |
[9] | Zhijing Xu, Jingjing Su, Kan Huang . A-RetinaNet: A novel RetinaNet with an asymmetric attention fusion mechanism for dim and small drone detection in infrared images. Mathematical Biosciences and Engineering, 2023, 20(4): 6630-6651. doi: 10.3934/mbe.2023285 |
[10] | Yong Tian, Tian Zhang, Qingchao Zhang, Yong Li, Zhaodong Wang . Feature fusion–based preprocessing for steel plate surface defect recognition. Mathematical Biosciences and Engineering, 2020, 17(5): 5672-5685. doi: 10.3934/mbe.2020305 |
The angiography and color fundus images are of great assistance for the localization of central serous chorioretinopathy (CSCR) lesions. However, it brings much inconvenience to ophthalmologists because of these two modalities working independently in guiding laser surgery. Hence, a novel fundus image fusion method in non-subsampled contourlet transform (NSCT) domain, aiming to integrate the multi-modal CSCR information, is proposed. Specifically, the source images are initially decomposed into high-frequency and low-frequency components based on NSCT. Then, an improved deep learning-based method is employed for the fusion of low-frequency components, which helps to alleviate the tedious process of manually designing fusion rules and enhance the smoothness of the fused images. The fusion of high-frequency components based on pulse-coupled neural network (PCNN) is closely followed to facilitate the integration of detailed information. Finally, the fused images can be obtained by applying an inverse transform on the above fusion components. Qualitative and quantitative experiments demonstrate the proposed scheme is superior to the baseline methods of multi-scale transform (MST) in most cases, which not only implies its potential in multi-modal fundus image fusion, but also expands the research direction of MST-based fusion methods.
CSCR is a common macular disease characterized by neurosensory retinal detachment (NRD) with or without pigment epithelium detachment (PED) [1,2], which often leads to blurred vision, central scotoma and deformed vision, and even results in permanent visual loss in serious cases. At present, laser surgery intervention is an effective measure for the treatment of CSCR, among which the angiography and color fundus images are the key auxiliary tools. As shown in Figure 1, the angiography image can clearly display the location and shape of the leakage areas (hereinafter referred to as lesions), but it has obvious color difference with the actual fundus field of view. On the contrary, the color fundus image cannot provide the location and shape information of the lesions, but it has the advantage of reflecting the actual color of fundus field of vision. In view of the characteristics of multi-modal fundus images, during the laser surgery procedure, ophthalmologists have to read the angiography image carefully to memorize the location and shape of the lesions, and then the corresponding lesion information will be marked on the color fundus image, and the marked positions of which are consistent with the angiography image. With the help of the marked color fundus image, the location and range of lesions in the real field of view are obtained in time by ophthalmologists through memory, which can provide guidance information for the location and area of laser surgery. However, it can be found that the multi-modal fundus images work independently when guiding the laser surgery of CSCR, that is, the ophthalmologists need to process the two kinds of images separately in order to get the final lesion information of real field of view, which not only increases the burden of ophthalmologists, but also causes inaccurate lesion information due to the cross-modal hand-marked deviation. Therefore, by utilizing multi-modal fundus information fusion strategy, we attempt to provide more accurate lesion information for the laser surgery of CSCR and simultaneously reduce the pressure of ophthalmologists. To the best of our knowledge, the multi-modal fundus image fusion for CSCR has not been done yet. Fortunately, the research of natural image fusion and brain image fusion has been widely carried out, which can provide some inspiration for the work of this paper.
Generally, image fusion can be divided into three categories, namely, pixel level, feature level, and decision level, of which the pixel-level image fusion performs directly on the source images and can retain more information. Next, the related research of this fusion scheme is described briefly. The pixel-level image fusion is mainly composed of spatial domain and transform domain [3]. In terms of the former, it aims to synthesize an image using techniques such as averaging the source images [4], the principal component analysis [5], the intensity-hue-saturation transform [6] and so on. Although spatial domain methods achieve better results, spectral degradation and contrast reduction [7] may be followed. Then the latter, namely transform domain, has emerged as an excellent fusion way to alleviate the above issues. The representation learning-based and multi-scale transform (MST) based methods are typical representatives of this fusion scheme. Yang et al. [8] applied the sparse representation (SR) theory to image fusion field. Specifically, the sparse coefficients of source images were firstly gained using an overcomplete dictionary. A choose-max fusion rule was then employed to combine the coefficients. With the help of global dictionary, the final fused image can be reconstructed accordingly. Li et al. [9] presented a low-rank representation-based fusion method. The low-rank representation was used instead of sparse representation to extract features from source images, and after that the L1-norm and the same fusion strategy as [8] were jointly adopted to reconstruct the fused image. In addition, the MST-based methods have marked predominance in multi-modal image fusion, where the techniques such as discrete wavelet transform (DWT) [10], Shearlet [11], Contourlet transform (CT) [12], NSCT [13] and non-subsampled shearlet transform (NSST) [14] are frequently applied to decompose the source images into high-frequency and low-frequency components. In particular, the NSCT has achieved great success in multi-modal image fusion by combining with various methods. Yong et al. [15] utilized a type-2 fuzzy logic-based rule to fuse the high-frequency components and a local energy algorithm to combine the low-frequency components, and the inverse NSCT was then followed to construct the final brain image. Similarly, a multi-modal image fusion method [16] combined with SR and PCNN was raised based on the NSCT decomposition. Especially, they employed the number of ignition times of PCNN to adjust the fusion parameters for the high-frequency components and acquired more detailed information. By applying the maximum entropy of square of the coefficients and the maximum weighted sum-modified laplacian in NSCT domain, respectively, Ganasala et al. [17] achieved the fusion of different components successfully. A fuzzy-adaptive reduced PCNN-based fusion way in NSCT domain was presented by Das et al. [18], the difference of which was that the fusion of high-frequency components and low-frequency components adopted the same method, namely using PCNN-based strategy with fuzzy-adaptive linking strengths.
NSCT is a shift-invariant, multiscale, and multi-direction expansion with fast implementability [19]. However, it cannot be better competent for specific image fusion tasks independently and may cause the degradation of image quality. Therefore, the combination with other fusion techniques like [15,16,17,18] is a common solution for NSCT and have achieved promising performance. Nevertheless, in order to make the fusion result possess both the content and details of the source images and shield the noise to the greatest extent, manually designing the fusion rules of different components plays an important role in MST-based methods, which makes these methods complicated and tediously. Fortunately, deep learning provides an effective solution for this problem and some research has preliminarily proved its potential. Hou et al. [20] put forward an unsupervised image fusion framework for adaptively extracting, fusing and reconstructing the deep features from the source images, which shows the applicability in infrared and visible image fusion task. Li et al. [21] also explored the deep learning-based method for image fusion, but only for extracting the multi-layer features of source images. After that the max selection strategy was still carried out on the candidate features generated by the L1-norm and weighted-average rule. Moreover, a fusion scheme [22] based on the generative adversarial network (GAN), was intelligently designed. In their method, the idea of generative adversarial training was cleverly integrated into the image fusion process where the fusion parameters of the source images were obtained in an end-to-end way, and achieved gratifying performance.
The aforementioned deep learning methods are beneficial to alleviate the multi-modal image fusion issue in NSCT domain, among which the GAN-based method has given us the most inspiration. In view of the dilemma of information acquisition of CSCR lesions in the laser surgery and the solid research foundation of image fusion mentioned above, this paper attempts to seek an adaptive parameter fusion method without too much manual intervention for multi-modal CSCR fundus image fusion and aims to obtain an assistant tool with the characteristics of both the angiography and color fundus images. The significance of this work is not only to relieve the pressure of ophthalmologists to capture the lesion information of CSCR in traditional laser surgery, but also to further facilitate the safety and efficiency of automatic laser surgery by providing clearer image data. In general, the main contributions of our research are as follows: 1) To the best of our knowledge, the image fusion method is firstly employed in the multi-modal CSCR fundus image scene for promoting to obtain more accurate information of the CSCR lesions, which provides a feasible scheme for better guiding the laser surgery of CSCR. 2) The data set of low-frequency and high-frequency components for the angiography and color fundus images is established though NSCT, which lays the foundation for better boosting the fusion of image content (structure) and detail information in NSCT domain. 3) An improved deep learning-based fusion method is employed to obtain the adaptive fusion parameters of low-frequency components in an end-to-end learning way, which enhances the smoothness of the fused images. Additionally, the PCNN-based fusion scheme is adopted to combine the high-frequency components, which facilitates the integration of detailed information. 4) Qualitative and quantitative experiments are performed on the multi-modal CSCR fundus images to verify the effectiveness of the proposed image fusion method in NSCT domain.
The rest of the paper is organized as follows. Section 2 presents the related work of NSCT, GAN and PCNN. Section 3 describes the details of the proposed image fusion method. Section 4 shows the experiments and discussions. Section 5 concludes this paper.
NSCT is a multi-scale transform technique developed on the basis of CT, which can effectively suppress pseudo-Gibbs phenomena [23]. Driven by non-subsampled pyramid filter bank (NSPFB) and non-subsampled directional filter bank (NSDFB), NSCT achieves multi-scale and multi-directional successfully [13]. Specifically, on the one hand, NSPFB can complete the iterative decomposition task on the lower-frequency component obtained by the previous decomposition, which ensures the multi-scale characteristics of NSCT. And on the other hand, NSDFB performs multi-directional decomposition with l directional levels on the high-frequency components, which can make NSCT obtain 2l sub-images with more precise directional details on the current scale. Besides, due to the removal of down-sampling and up-sampling operations in the process of multi-scale transform and multi-directional decomposition processes, NSCT has the advantage of shift-invariance property. It is also the above superiorities that make NSCT shine in a variety of image fusion scenes.
GAN [24] enhances its ability to capture data distribution by skillfully establishing an adversarial game between the generator G and the discriminator D. Since its birth, it has been widely used in various visual tasks [25,26]. Yu et al. [25] effectively enhanced the capability of cervical cell detection model by introducing images generated from GAN (i.e., data augmentation) when training the convolutional neural network. In the aspect of image fusion, with the help of adversarial training, the integration of infrared and visible images was achieved successfully [26]. Whether it is the process of data augmentation or image fusion, the application of GAN are essentially a minimax two-player game between G and D from a mathematical point of view, which can be described as the following formula:
minGmaxDVGAN(G,D)=Ex∼Pdata(x)[logD(x)]+Ez∼Pz(z)[log(1−D(G(z)))] | (2.1) |
where x is a sample from a real distribution Pdata(x) and z denotes a sample from a concrete prior distribution Pz(z). Because it is difficult to express Pdata(x) explicitly, so GAN tries its best to fit the distribution of Pdata(x) by fooling D using the generated samples based on z. The whole process can be further described separately as follows:
D∗=maxDVGAN(G,D)=Ex∼Pdata(x)[logD(x)]+Ez∼Pz(z)[log(1−D(G(z)))] | (2.2) |
G∗=minGVGAN(G,D)=Ez∼Pz(z)[log(1−D(G(z)))] | (2.3) |
where D∗ and G∗ represent the optimal G and D, respectively. Alternating optimization drives G to generate more realistic data distribution close to Pdata(x), and the capability of D to distinguish true from false is gradually enhanced. Once the adversarial training reaches equilibrium, the final generative model can be applied in the corresponding scene.
PCNN is an artificial neural network that simulates the information processing mechanism of mammalian visual cortex, which processes the input signal autonomously through the cooperation of three sub-modules, namely, the receiving field, modulation field and pulse generator module, and has been well implemented in the field of image fusion [27,28]. To perform object detection, the PCNN-based image fusion network [27] was firstly chosen as an architecture for the fusion of the original and filtered images, and then the target can be highlighted consequently and thus be easily detected. In order to enhance the ability to capture edge information, Tan et al. [28] properly introduced multi-scale morphological gradient into the PCNN-based fusion method to improve its performance in remote sensing image fusion task. The same technique is also used in multi-modal brain image fusion task for integrating the information from the fine-structure and coarse-structure layers [29]. It can be found that PCNN has performed well in the above fields, especially in the fusion task, which also encourages us to yearn for it in our multi-modal CSCR fundus image fusion scene.
The CSCR image fusion scheme proposed in this paper is an innovative application of GAN-based and PCNN-based image fusion methods in NSCT domain. Firstly, NSCT decomposes the CSCR images into high-frequency and low-frequency components, which promotes the smooth fusion of information in the angiography and color fundus images. Secondly, the GAN-based technique is employed for the fusion of low-frequency components, which can determine the adaptive fusion parameters in a learning way and greatly avoid the manual intervention. Simultaneously, PCNN is adopted to fuse the high-frequency components, which can effectively integrate the edge and contour information from the two modal images. Finally, the required fused CSCR image can be obtained by applying a color conversion operation on the fused image YAB achieved based on the inverse transform on the above fusion components. The flow chart of the proposed method is shown in Figure 2.
It should be noted that the preprocessing operation needs to be done before the fusion task is formally executed, the parts of which mainly involve the registration of multi-modal fundus images, the clipping of macular region, and the establishment of data set. Generally, registration is used for the alignment of sequence images or multi-modal images to ensure the consistency of spatial information of image content, which is very important for the analysis of medical images [30]. Considering the accuracy and efficiency of registration, the scheme in [31] is adopted in our CSCR fundus image registration task. After that, in order to speed up the fusion of multi-modal information, the macular region is roughly localized by clipping, which can limit the image fusion range to the occurrence area of CSCR. Following that is the color space conversion and NSCT operation, among which the YIQ color conversion is applied to separate the brightness and color channels and followed by the NSCT operation on the brightness channel to establish the data set based on the high-frequency and low-frequency components. Furthermore, the low-frequency components will be further divided into patches with size of 24 in this paper, which can not only expand the training samples, but also reduce the training cost.
In recent years, deep learning techniques have been involved in various image fusion scenes and achieved good performance [20,21,22,26]. Among them, the GAN-based fusion schemes [22,26] have shown great advantages, which provides a new solution for the determination of adaptive fusion parameters. This section is the work extension of GAN-based image fusion scheme (as shown in Figure 3), which is embodied in three aspects, namely, a new attempt in application domain, a new incorporation in the MST domain and a new adjustment in technique domain. Firstly, we apply the GAN-based fusion scheme to the multi-modal CSCR fundus image information integration task, which broadens its application field and scope. Secondly, the scheme is intriguingly integrated into the MST domain, aiming at capturing the data distribution of the low-frequency components in an end-to-end learning way and driving the generation of adaptive fusion parameters to enhance the smoothness of the fused images. Thirdly, due to the computational cost of NSCT, not all details are filtered into the high-frequency components [13], which means that the structure and detailed information need to be processed synchronously in the fusion of low-frequency components. To this issue, the cycle-consistency loss term, inspired by [32], is integrated into the deep learning-based fusion scheme to further enhance the learning ability of GAN model for low-frequency component data distribution.
Considering the loss of information, the up-sampling and down-sampling operations in [22] are removed in our work and other network structures remain unchanged. Finally, the objective optimization framework corresponding to Eq (2.1) is as follow:
minGABmaxDA,DBVGAN(G,DA,DB)=E[logDA(LA)]+E[log(1−DA(GAB(LA,LB)))]+E[logDB(LB)]+E[log(1−DB(GAB(LA,LB)))] | (3.1) |
where GAB s the generator for obtaining the adaptive fusion parameters and generating fusion results, and DA and DB denote the discriminators that drive GAB to capture the data distributions of the low- frequency components LA and low-frequency component LB synchronously. Similarly, corresponding to Eqs (2.2) and (2.3), we define the loss functions for DA, DB and GAB as follows:
Loss_DA=E[logDA(LA)]+E[log(1−DA(GAB(LA,LB)))] | (3.2) |
Loss_DB=E[logDB(LB)]+E[log(1−DB(GAB(LA,LB)))] | (3.3) |
Loss_GAB=E[log(1−DA(GAB(LA,LB)))]+E[log(1−DB(GAB(LA,LB)))] | (3.4) |
As mentioned above, we adopt the cycle-consistency term to improve the performance of GAN model, that is to say, to enhance the learning ability of the generator GAB for low-frequency component data distribution. Specifically, through two generators with the same encoding and decoding structures as GAB, the fused image is translated back to the corresponding modals respectively to force GAB to grasp the data distribution of the source images in detail. The cycle-consistency loss can be written as follows:
LossAB→A=E[‖GA(GAB(LA,LB))−LA‖1] | (3.5) |
LossAB→B=E[‖GB(GAB(LA,LB))−LB‖1] | (3.6) |
Loss_cycle=ηLossAB→A+βLossAB→B | (3.7) |
where GA and GB represent the generators of corresponding modals respectively, and ‖⋅‖1 is the L1-norm term that reduces blurs in the fused images. Then, the total loss function can be defined as follows:
Loss_T=Loss_GAB+Loss_DA+Loss_DB+λLoss_cycle | (3.8) |
where the η, β and λ denote the trade-off parameters adjusting the proportion of the corresponding term, which are set to 0.05, 1.5 and 0.8, respectively. Then, the final optimized framework Eq (3.1) can be rewritten as:
G∗AB,D∗A,D∗B=minGABmaxDA,DBLoss_T | (3.9) |
where G∗AB, D∗A and D∗B represent the optimal GAB, DA and DB respectively. In practice, the image fusion method based on GAN is not stable in the training process. In order to reduce the model oscillation, this paper follows the training strategy in [22], that is, the training quota is allocated according to the performance of the generators and discriminators instead of training in turns.
The existing PCNN model is developed on the basis of biological neuron model [33]. Due to simulating the information processing mechanism of mammalian visual cortex, the signal processed by the model is closer to the characteristics of human vision. However, setting a large number of parameters is a defect of the traditional PCNN model. Therefore, a simplified PCNN [34] model is introduced into our multi-modal CSCR fundus image fusion research, which is specifically used for the information integration of the high-frequency components in this section. The expressions of the simplified PCNN model can be defined as follows:
FAij=SAij,FBij=SBij | (3.10) |
LAij(n)=e−aALLAij(n−1)+VAL∑WAijklYAkl(n−1),LBij(n)=e−aBLLBij(n−1)+VBL∑WBijklYBkl(n−1) | (3.11) |
where SAij and SBij denote the external stimuli determined by the high-frequency components in this paper, and FAij and FBij are the feedbacks input of the above inputs respectively. LAij and LBij represent the linking parameters of neurons, and the corresponding amplification and attenuation coefficients are aAL,VAL,aBL,VBL, among which aAL equals aBL and VAL equals VBL in this paper. YAkl and YBkl are the corresponding outputs of neuron (k, l). WAijkl and WBijkl are the connection matrixes between neuron (i, j) and neuron (k, l) in the respective low-frequency components, and they can be express as follow:
WAijkl=1√(iA−kA)2+(jA−lA)2,WBijkl=1√(iB−kB)2+(jB−lB)2 | (3.12) |
where (iA,jA) and (iB,jB) denote the current neurons in the high-frequency component A and the high-frequency component B respectively, and (kA,lA) and (kA,lA) are the corresponding neighbor neurons.
UAij(n)=FAij(1+βALAij(n−1)),UBij(n)=FBij(1+βBLBij(n−1)) | (3.13) |
EAij(n)=e−aAEEAij(n−1)+VAEYAij(n−1),EBij(n)=e−aBEEBij(n−1)+VBEYBij(n−1) | (3.14) |
YAij(n)={1,ifEAij(n)≤UAij(n)0,else,YBij(n)={1,ifEBij(n)≤UBij(n)0,else | (3.15) |
where UAij and UBij denote the modulation output results, and βA and βB are the link strength factors that take the same value.EAij and EBij are dynamic thresholds used to adjust the firing frequency of neurons, and the corresponding amplification and attenuation coefficients are aAE,VAE,aBE,VBE, among which aAE equals aBE and VAE equals VBE in this paper. Then, based on the step function, the pulse value YAij and YBij of PCNN model can be obtained. Finally, we acquire the fusion map of the high-frequency components by the following formula:
FA=N∑n=1YAij(n),FB=N∑n=1YBij(n),Fmap={1,ifFA≥FB0, else | (3.16) |
where FA and FB denote the sum of the pulses with N iterations. Based on Eq (3.16), the final fusion result of the high-frequency components can be obtained, that is, HAB shown in Figure 4.
Through Sections 3.2 and 3.3, the low-frequency components and high-frequency components of multi-modal CSCR fundus images can be integrated by deep learning-based fusion method and PCNN-based fusion method respectively, which greatly reduces the manual intervention of fusion parameters and strengthens the integration of detail information. On this basis, the preliminary fusion results can be obtained by performing the NSCT inverse transform on the low-frequency fusion and high- frequency fusion results, followed by the YIQ-based color space conversion on the above results to obtain the final fusion image based on the RGB color space.
In the task of multi-modal CSCR fundus image fusion in this paper, a total of 396 angiography and color fundus image pairs are used to verify the performance of the proposed image fusion method based on NSCT. By relying on the NVIDIA-2080ti GPU with the Tensorflow framework, 356 image pairs are utilized to train the deep learning-based fusion method and the remaining image pairs are used for testing. Moreover, in order to expand the training samples and reduce the training cost, the low-frequency components are further divided into 297,457 patches with the size of 24. Meanwhile, in the training process, the learning rate, batch size and learning decay rate are set to 0.0002, 64 and 0.9 respectively, and the size of all test images is adjusted to 256 to ensure the fusion effect of the whole CSCR area in the test process.
Quantitative analysis of image fusion results is one of the important ways to evaluate the fusion methods effectively. In this paper, six commonly used evaluation indexes are applied to analyze the fusion effect of multi-modal CSCR fundus image, namely, average gradient (AG), entropy (EN) [35], mutual information (MI) [36], Qabf [37], Nabf [38], structural similarity (SSIM) [39]. The details are as follows:
● AG: AG is the average value of the image gradient, which represents the clarity of the fused image and can be expressed as follow:
AG=1(M−1)×(N−1)M−1∑i=1N−1∑j=1√((F(i+1,j)−F(i,j))2+(F(i,j+1)−F(i,j))2)/2 | (4.1) |
where F denotes the final fused image, and M and N represent the size of F.
● EN: EN is an objective evaluation to measure the amount of information contained in the fused image. The larger the value is, the more information the fused image carries. The formula is as follow:
EN=−W∑l=1p(l)log2p(l) | (4.2) |
where W denotes all gray levels in the final fused image and p(l) represents the proportion of gray value l to all gray levels.
● MI: MI can indicate how much information the fused image carries about the source images, and the greater the MI value is, the better the fusion effect is. The formula of MI can be expressed as follow:
MI=MIAF+MIBF | (4.3) |
MIAF=W∑lW∑ppAF(l,p)log2(pAF(l,p)pA(l)pF(p)),MIBF=W∑lW∑ppBF(l,p)log2(pBF(l,p)pB(l)pF(p)) | (4.4) |
where MIAF and MIBF denote the mutual information between the source image A and the fused image F, the source image B and the fused image F, respectively. pAF(l,p) and pBF(l,p) are the corresponding joint distribution of gray levels.
● Qabf and Nabf: Qabf reflects the quality of the visual information obtained from the input image, a higher value of which means that the quality of the fused image is better. Details of the formula can be found in [37]. Nabf represents the noise or artifacts added in fused image due to fusion process, a smaller value of the index indicates a higher image quality. Refer to reference [38] for details.
● SSIM: SSIM is a measure to characterize the structural similarity between the fused image and the source images. A larger value means that the fused image is more consistent with the source images in terms of structure. The formula of SSIM is shown as follow:
SSIM=(SSIMAF+SSIMBF)/2 | (4.5) |
SSIMAF=(2μAμF+C1)(2σAF+C2)(μ2A+μ2F+C1)(σ2A+σ2F+C2),SSIMBF=(2μBμF+C1)(2σBF+C2)(μ2B+μ2F+C1)(σ2B+σ2F+C2) | (4.6) |
where SSIMAF and SSIMBF denote the structural similarity between the source image A and the fused image F, the source image B and the fused image F, respectively. μA, μB and μF are the corresponding mean intensity respectively. σA and σB are the standard deviations, and σAF and σBF denote the covariances between the source images and the fused image F respectively. C1=(K1∗255)2 and C2=(K2∗255)2 are two constants to prevent an unstable results, and K1 and K2 are set to 0.01 and 0.03 respectively.
In this section, we conduct the ablation experiment to verify the effectiveness of the cycle-consistency loss module in the GAN-based fusion method for the low-frequency components. The results of quantitative comparison are shown in Figure 5.
Significantly, in the five evaluation indexes, namely, EN, Qabf, MI, Nabf and SSIM, the deep learning-based fusion scheme equipped with the cycle-consistency loss module has achieved good results in most cases. Especially the MI index, in the absence of the cycle-consistency loss module, the final fusion images obtain less information from the source images, which may lead to the lack of key lesion information and is very unfavourable for the multi-modal CSCR fundus image fusion task. In addition, in terms of the AG index, it can be found that the low-frequency fusion scheme integrated with the cycle-consistency loss module will be slightly inferior to that without the same module, which means a slight difference in image clarity. However, on the basis of other indexes, it can be observed that this is at the cost of image information reduction and noise increase, which is obviously not desirable.
Besides, we show the effect of the cycle-consistency loss module in the fusion scheme based on deep learning through the specific low-frequency fusion experiments. As shown in Figure 6, the first line and the second line are the low-frequency components of the angiography and color fundus images, represented by LA and LB respectively. The third line and the fourth line denote the GAN-based fusion method with and without the cycle-consistency loss module respectively, represented by Scheme 1 and Scheme 2. Under the same experimental setup, Scheme 1 can better learn the data distribution of LA and LB, and the final low-frequency fusion results not only contain normal structural information, but also better retain the lesion content. But the performance of Scheme 2 is on the contrary. Specifically, the low-frequency fusion images under this scheme are generally not smooth enough, and the fusion quality of the lesion information is poor, which seriously affects the ophthalmologists' judgment of the shape and location of the lesions, and thus cannot guide the CSCR laser surgery well.
As shown in Figure 7, it further conveys the difference of the final fusion results obtained by the proposed fusion method using the above two low-frequency fusion schemes. It can be clearly observed that part of the lesion information is lost in Scheme 2 and the vessels and macular area are not smooth enough, which are consistent with the performance in Figure 6 and confirm the necessity of introducing the cycle-consistency loss module. On the contrary, Scheme 1 not only captures the distribution of lesions in the source images, but also better integrates the normal structural information. More importantly, the transition of each region of the fused image is natural and smooth, which can provide better reference for ophthalmologists to identify the lesion information.
After the previous section, the importance of the cycle-consistency loss module in the multi-modal CSCR fundus image fusion task is reflected. On this basis, this section further discusses the fusion strategy of high-frequency components, specifically involving the PCNN-based fusion strategy (i.e., strategy 1) and the commonly used choose-max fusion strategy (i.e., strategy 2). On the premise that the GAN-based fusion scheme is the low-frequency component fusion strategy, the results shown in Figure 8 can be obtained by using the above two high-frequency component fusion strategies. Apparently, except for the Nabf index, the PCNN-based fusion strategy can make the proposed fusion method more attractive in the task of this paper. In particular, the three indexes, namely AG, Qabf and MI, show that the strategy can promote the whole fusion method to obtain high-quality and high-definition fusion results which inherit more information from the source images. Admittedly, the choose-max fusion strategy can indeed make the fusion results avoid too much noise, but it cannot make up for the losses of other aspects of the fusion results, such as the image clarity, image quality and the source image information carried by the fused images.
In addition, corresponding to the above analysis, it can be clearly observed from Figure 9 that the fused images obtained by strategy 1 are significantly better than those obtained by strategy 2 in terms of clarity. Moreover, the latter results in a weak contrast of image fusion results, which makes it difficult to distinguish the fundus background area, normal structure and lesion information to a certain extent. On the contrary, the performance of strategy 1 in this aspect is commendable. The above discussions prove the applicability and feasibility of the PCNN-based fusion strategy in the multi-modal CSCR fundus image fusion task in this paper, which also lay the foundation for the combination of the PCNN-based and deep learning-based fusion strategies.
In this part, the proposed multi-modal CSCR fundus image fusion method in NSCT domain is compared with the MST baseline methods, and the fusion techniques of low-frequency components and high-frequency components in the baseline methods adopt the common fusion strategies [8,40]. Additionally, both the proposed and baseline methods perform image fusion tasks in the Y channel of YIQ color space and acquire the final fusion images after the same color space conversion operation. The results of quantitative comparison are shown in Figure 10. It can be found that the AG, EN and Qabf values of the proposed method are better than the baseline methods in most cases. Among them, the AG and Qabf indexes have significant advantages, which imply that the fused images are clearer and of good quality. Compared with the Shearlet and NSST fusion methods, it should be pointed out that the proposed fusion scheme is competitive in terms of the MI index. Furthermore, although there are still some gaps in the Nabf and SSIM indexes between our proposed fusion method and some baseline schemes, this slight difference of which also confirms the potential of the proposed method. At the same time, in order to display the fusion performance of multi-modal CSCR fundus images with different fusion methods more conveniently, the average fusion indexes of 40 pairs of CSCR images are specially counted and summarized in Table 1. The quantitative results in the table show that the proposed method is superior to other baseline fusion methods in the average values of AG, EN and Qabf and projects its competitiveness in most cases in other indexes, which is consistent with the above analysis.
DWT | NSCT | NSST | Shearlet | Contourlet | Proposed | |
AG | 5.9231 | 3.9179 | 5.0483 | 5.4747 | 5.8684 | 6.5489 |
EN | 6.9128 | 6.8538 | 6.8928 | 6.8963 | 6.9133 | 7.1129 |
MI | 3.0255 | 2.9946 | 3.4880 | 3.6318 | 2.9755 | 3.3504 |
Qabf | 0.4039 | 0.3823 | 0.5174 | 0.5583 | 0.3843 | 0.6699 |
Nabf | 0.1730 | 0.0233 | 0.0786 | 0.0883 | 0.1946 | 0.1083 |
SSIM | 0.5776 | 0.6403 | 0.6273 | 0.6260 | 0.5705 | 0.6067 |
After the same color space conversion, the MST baseline methods and the proposed method can obtain the final CSCR fusion images. The specific fusion results are shown in Figure 11. From the perspective of image clarity, the results gained by the proposed fusion method based on NSCT are relatively clear, and the contours of vessels and lesions are easy to identify. From the perspective of contrast, the proposed method can make the final fusion images have better contrast, and make the vessels, macular area and background easy to distinguish. From the perspective of preserving image information, all the methods can achieve the fusion task of the angiography and color fundus images, but on the whole our scheme can better deal with the fusion of structure and detail information of the source images.
Generally speaking, compared with the MST baseline fusion methods, the proposed fusion scheme based on NSCT can be able to better realize the information integration of multi-modal CSCR fundus images, which shows its potential value. The fusion scheme not only innovatively realizes the combination of the deep learning-based and PCNN-based fusion techniques, but also provides a new idea for the research of fusion rules based on MST fusion methods. Moreover, through the information integration of multi-modal fundus images, such an image fusion scheme provides a more convenient auxiliary tool characterized by similar fundus color and clear lesion and structure information for guiding the operation of traditional CSCR laser surgery.
In this paper, a multi-modal CSCR fundus image fusion method in NSCT domain is presented to assist ophthalmologists to identify the CSCR lesion information and normal fundus structure more convenient so as to better guide the laser surgery. This fusion idea is an intriguing combination of the GAN-based and PCNN-based fusion schemes, which is not only the application extension of deep learning-based fusion technique in the low-frequency component fusion task, but also the reappearance of PCNN fusion rule in the information integration field of high-frequency components. On the one hand, in order to alleviate the tedious process of manually designing fusion rules, we introduce the GAN-based method into the fusion scene of low-frequency components and improve its learning ability of the data distribution by integrating the cycle-consistency loss module. On the other hand, the PCNN-based method is employed to promote to fuse the high-frequency components, aiming to enhance the integration of detail information of multi-modal CSCR fundus images. Quantitative and qualitative experiments demonstrate the effectiveness and competitiveness of the proposed method in the fusion scene of CSCR fundus images. Although the proposed fusion method achieves considerable performance and provides a new idea for the research of MST-based fusion methods, we will continue to make efforts from the following two aspects that include: 1) further optimizing the training strategy of the GAN fusion method to improve the training efficiency and 2) further improving the fusion rules of the low-frequency components and high-frequency components to enhance the quality of fused images.
This work was financially supported by Fundamental Research Funds for the Central Universities (No.NP2020420) and China Postdoctoral Science Foundation funded project (No.2019M661832). We also appreciate it very much for editors and reviewers with your attention to our paper.
The authors declare that we have no conflict of interests regarding this paper.
[1] |
A. Daruich, A. Matet, A. Dirani, E. Bousquet, M. Zhao, N. Farman, et al., Central serous chorioretinopathy: Recent findings and new physiopathology hypothesis, Prog. Retinal Eye Res., 48 (2015), 82-118. doi: 10.1016/j.preteyeres.2015.05.003
![]() |
[2] |
J. Yu, C. Jiang, G. Xu, Study of subretinal exudation and consequent changes in acute central serous chorioretinopathy by optical coherence tomography, Am. J. Ophthalmol., 158 (2014), 752-756. doi: 10.1016/j.ajo.2014.06.015
![]() |
[3] |
P. Balasubramaniam, V. P. Ananthi, Image fusion using intuitionistic fuzzy sets, Inf. Fusion, 20 (2014), 21-30. doi: 10.1016/j.inffus.2013.10.011
![]() |
[4] |
H. Yin, Tensor sparse representation for 3-D medical image fusion using weighted average rule, IEEE Trans. Biomed. Eng., 65 (2018), 2622-2633. doi: 10.1109/TBME.2018.2811243
![]() |
[5] |
J. Li, M. Song, Y. Peng, Infrared and visible image fusion based on robust principal component analysis and compressed sensing, Infrared Phys. Technol., 89 (2018), 129-139. doi: 10.1016/j.infrared.2018.01.003
![]() |
[6] |
Y. Leung, J. Liu, J. Zhang, An improved adaptive intensity-hue-saturation method for the fusion of remote sensing images, IEEE Geosci. Remote Sens. Lett., 11 (2014), 985-989. doi: 10.1109/LGRS.2013.2284282
![]() |
[7] | Y. Yang, S. Tong, S. Huang, P. Lin, Multi-focus image fusion based on NSCT and focused area detection, IEEE Sens. J., 15 (2015), 2824-2838. |
[8] |
B. Yang, S. Li, Multi-focus image fusion and restoration with sparse representation, IEEE Trans. Instrum. Meas., 59 (2010), 884-892. doi: 10.1109/TIM.2009.2026612
![]() |
[9] | H. Li, X. Wu, Multi-focus image fusion using dictionary learning and low-rank representation, in International Conference on Image and Graphics, Springer, 10666 (2017), 675-686. |
[10] |
A. B. Hamza, Y. He, H. Krim, A. Willsky, A multiscale approach to pixel-level image fusion, Integr. Comput.-Aided Eng., 12 (2005), 135-146. doi: 10.3233/ICA-2005-12201
![]() |
[11] |
L. Wang, B. Li, L. Tian, Multi-modal medical image fusion using the inter-scale and intra-scale dependencies between image shift-invariant shearlet coefficients, Inf. Fusion, 19 (2014), 20-28. doi: 10.1016/j.inffus.2012.03.002
![]() |
[12] | M. N. Do, M. Vetterli, Contourlets: a directional multiresolution image representation, in Proceedings. International Conference on Image Processing, Rochester, NY, USA, 2002. |
[13] |
Z. Zhu, M. Zheng, G. Qi, D. Wang, Y. Xiang, A phase congruency and local Laplacian energy based multi-modality medical image fusion method in NSCT domain, IEEE Access, 7 (2019), 20811-20824. doi: 10.1109/ACCESS.2019.2898111
![]() |
[14] |
G. Easley, D. Labate, W.Q. Lim, Sparse directional image representations using the discrete shearlet transform, Appl. Comput. Harmonic Anal., 25 (2008), 25-46. doi: 10.1016/j.acha.2007.09.003
![]() |
[15] |
Y. Yang, Y. Que, S. Huang, P. Lin, Multimodal sensor medical image fusion based on type-2 fuzzy logic in NSCT domain, IEEE Sens. J., 16 (2016), 3735-3745. doi: 10.1109/JSEN.2016.2533864
![]() |
[16] | J. Xia, Y. Chen, A. Chen, Y. Chen, Medical image fusion based on sparse representation and PCNN in NSCT domain, Comput. Math. Methods Med., 2018 (2018), 1-12. |
[17] |
P. Ganasala, V. Kumar, CT and MR image fusion scheme in non-subsampled contourlet transform domain, J. Digit. Imag., 27 (2014), 407-418. doi: 10.1007/s10278-013-9664-x
![]() |
[18] |
S. Das, M. K. Kundu, A neuro-fuzzy approach for medical image fusion, IEEE Trans. Biomed. Eng., 60 (2013), 3347-3353. doi: 10.1109/TBME.2013.2282461
![]() |
[19] |
A. L. D. Cunha, J. Zhou, M. N. Do, The non-subsampled contourlet transform: Theory, design, and applications, IEEE Trans. Image Process., 15 (2006), 3089-3101. doi: 10.1109/TIP.2006.877507
![]() |
[20] |
R. Hou, D. Zhou, R. Nie, D. Liu, L. Xiong, Y. Guo, et al., VIF-Net: an unsupervised framework for infrared and visible image fusion, IEEE Trans. Comput. Imaging, 6 (2020), 640-651. doi: 10.1109/TCI.2020.2965304
![]() |
[21] | H. Li, X. Wu, J. Kittler, Infrared and visible image fusion using a deep learning framework, in 2018 24th international conference on pattern recognition (ICPR), Beijing, China, (2018), 2705-2710. |
[22] |
J. Ma, H. Xu, J. Jiang, X. Mei, X. Zhang, DDcGAN: A dual-discriminator conditional generative adversarial network for multi-resolution image fusion, IEEE Trans. Image Process., 29 (2020), 4980-4995. doi: 10.1109/TIP.2020.2977573
![]() |
[23] |
W. Kong, Y. Lei, X. Ni, Fusion technique for grey-scale visible light and infrared images based on non-subsampled contourlet transform and intensity-hue-saturation transform, IET Signal Process., 5 (2011), 75-80. doi: 10.1049/iet-spr.2009.0263
![]() |
[24] | I. J. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, et al., Generative adversarial networks, Adv. Neural Inf. Process. Syst., 3 (2014), 2672-2680. |
[25] |
S. Yu, S. Zhang, B. Wang, H. Dun, L. Xu, X. Huang, et al., Generative adversarial network based data augmentation to improve cervical cell classification model, Math. Biosci. Eng., 18 (2021), 1740-1752. doi: 10.3934/mbe.2021090
![]() |
[26] |
J. Ma, W. Yu, P. Liang, C. Li, J. Jiang, FusionGAN: A generative adversarial network for infrared and visible image fusion, Inf. Fusion, 48 (2019), 11-26. doi: 10.1016/j.inffus.2018.09.004
![]() |
[27] |
R. P. Broussard, S. K. Rogers, M. E. Oxley, G. L. Tarr, Physiologically motivated image fusion for object detection using a pulse coupled neural network, IEEE Trans. Neural Networks, 10 (1999), 554-563. doi: 10.1109/72.761712
![]() |
[28] |
W. Tan, P. Xiang, J. Zhang, H. Zhou, H. Qin, Remote sensing image fusion via boundary measured dual-channel pcnn in multi-scale morphological gradient domain, IEEE Access, 8 (2020), 42540-42549. doi: 10.1109/ACCESS.2020.2977299
![]() |
[29] |
W. Tan, W. Thitøn, P. Xiang, H. Zhou, Multi-modal brain image fusion based on multi-level edge-preserving filtering, Biomed. Signal Process. Control, 64 (2021), 102280. doi: 10.1016/j.bspc.2020.102280
![]() |
[30] |
D. Paquin, D. Levy, E. Schreibmann, L. Xing, Multiscale image registration, Math. Biosci. Eng., 3 (2006), 389-418. doi: 10.3934/mbe.2006.3.389
![]() |
[31] |
J. Wang, J. Chen, H. Xu, S. Zhang, X. Mei, J. Huang, et al., Gaussian field estimator with manifold regularization for retinal image registration, Signal Process., 157 (2019), 225-235. doi: 10.1016/j.sigpro.2018.12.004
![]() |
[32] | K. Li, L. Yu, S. Wang, P. A. Heng, Unsupervised retina image synthesis via disentangled representation learning, in International Workshop on Simulation and Synthesis in Medical Imaging, Springer, Cham, 11827 (2019), 32-41. |
[33] |
E. M. Izhikevich, Class 1 neural excitability, conventional synapses, weakly connected networks, and mathematical foundations of pulse-coupled models, IEEE Trans. Neural Networks, 10 (1999), 499-507. doi: 10.1109/72.761707
![]() |
[34] |
S. Yang, M. Wang, Y. Lu, W. Qi, L. Jiao, Fusion of multi-parametric SAR images based on SW-non-subsampled contourlet and PCNN, Signal Process., 89 (2009), 2596-2608. doi: 10.1016/j.sigpro.2009.04.027
![]() |
[35] |
J. V. Aardt, Assessment of image fusion procedures using entropy, image quality, and multispectral classification, J. Appl. Remote Sens., 2 (2008), 023522. doi: 10.1117/1.2945910
![]() |
[36] |
G. Qu, D. Zhang, P. Yan, Information measure for performance of image fusion, Electron. Lett., 38 (2002), 313-315. doi: 10.1049/el:20020212
![]() |
[37] |
C. S. Xydeas, V. S. Petrovic, Objective image fusion performance measure, Electron. Lett., 36 (2000), 308-309. doi: 10.1049/el:20000267
![]() |
[38] |
B. K. S. Kumar, Multifocus and multispectral image fusion based on pixel significance using discrete cosine harmonic wavelet transform, Signal, Image Video Process., 7 (2013), 1125-1143. doi: 10.1007/s11760-012-0361-x
![]() |
[39] |
Z. Wang, A. C. Bovik, H. R. Sheikh, E. P. Simoncelli, Image quality assessment: From error visibility to structural similarity, IEEE Trans. Image Process., 13 (2004), 600-612. doi: 10.1109/TIP.2003.819861
![]() |
[40] | H. Li, X. Wu, T. Durrani, Multi-focus noisy image fusion using low-rank representation, preprint, arXiv: 1804.09325. |
1. | Junqiang Zhao, Yi Lu, Shaojun Zhu, Keran Li, Qin Jiang, Weihua Yang, Systematic Bibliometric and Visualized Analysis of Research Hotspots and Trends on the Application of Artificial Intelligence in Ophthalmic Disease Diagnosis, 2022, 13, 1663-9812, 10.3389/fphar.2022.930520 | |
2. | Bo Zheng, Yunfang Liu, Kai He, Maonian Wu, Ling Jin, Qin Jiang, Shaojun Zhu, Xiulan Hao, Chenghu Wang, Weihua Yang, Yi Shao, Research on an Intelligent Lightweight-Assisted Pterygium Diagnosis Model Based on Anterior Segment Images, 2021, 2021, 1875-8630, 1, 10.1155/2021/7651462 | |
3. | Shaojun Zhu, Bing Lu, Chenghu Wang, Maonian Wu, Bo Zheng, Qin Jiang, Ruili Wei, Qixin Cao, Weihua Yang, Screening of Common Retinal Diseases Using Six-Category Models Based on EfficientNet, 2022, 9, 2296-858X, 10.3389/fmed.2022.808402 | |
4. | Sukun Tian, Renkai Huang, Zhenyang Li, Luca Fiorenza, Ning Dai, Yuchun Sun, Haifeng Ma, Cosimo Ieracitano, A Dual Discriminator Adversarial Learning Approach for Dental Occlusal Surface Reconstruction, 2022, 2022, 2040-2309, 1, 10.1155/2022/1933617 | |
5. | Shahid Karim, Geng Tong, Jinyang Li, Akeel Qadir, Umar Farooq, Yiting Yu, Current advances and future perspectives of image fusion: A comprehensive review, 2023, 90, 15662535, 185, 10.1016/j.inffus.2022.09.019 | |
6. | 青山 陈, The Comparison Study of Tablet Aescin Therapy versus High-Density Subthreshold Micropulse Treatment Laser for Central Serous Chorioretinopathy—A Prospective Cohort Study, 2023, 12, 2167-6542, 45, 10.12677/HJO.2023.121007 | |
7. | Chen Qin, Shimin Wen, Fan Fei, Yuemei Han, Haiting Wang, Hao Chen, Quankui Lin, NIR-triggered thermosensitive polymer brush coating modified intraocular lens for smart prevention of posterior capsular opacification, 2023, 21, 1477-3155, 10.1186/s12951-023-02055-2 | |
8. | Xiaofen Wang, Automatic Alignment of Ring-Grid Laser Images based on an Improved Classical Iterative Algorithm, 2024, 32, 0218-4885, 433, 10.1142/S0218488524400026 | |
9. | Sukun Tian, Haifeng Ma, Pan Huang, Xiang Wang, Tianxiang Li, Renkai Huang, RADDA-Net: Residual attention-based dual discriminator adversarial network for surface defect detection, 2024, 136, 09521976, 108887, 10.1016/j.engappai.2024.108887 |
DWT | NSCT | NSST | Shearlet | Contourlet | Proposed | |
AG | 5.9231 | 3.9179 | 5.0483 | 5.4747 | 5.8684 | 6.5489 |
EN | 6.9128 | 6.8538 | 6.8928 | 6.8963 | 6.9133 | 7.1129 |
MI | 3.0255 | 2.9946 | 3.4880 | 3.6318 | 2.9755 | 3.3504 |
Qabf | 0.4039 | 0.3823 | 0.5174 | 0.5583 | 0.3843 | 0.6699 |
Nabf | 0.1730 | 0.0233 | 0.0786 | 0.0883 | 0.1946 | 0.1083 |
SSIM | 0.5776 | 0.6403 | 0.6273 | 0.6260 | 0.5705 | 0.6067 |
DWT | NSCT | NSST | Shearlet | Contourlet | Proposed | |
AG | 5.9231 | 3.9179 | 5.0483 | 5.4747 | 5.8684 | 6.5489 |
EN | 6.9128 | 6.8538 | 6.8928 | 6.8963 | 6.9133 | 7.1129 |
MI | 3.0255 | 2.9946 | 3.4880 | 3.6318 | 2.9755 | 3.3504 |
Qabf | 0.4039 | 0.3823 | 0.5174 | 0.5583 | 0.3843 | 0.6699 |
Nabf | 0.1730 | 0.0233 | 0.0786 | 0.0883 | 0.1946 | 0.1083 |
SSIM | 0.5776 | 0.6403 | 0.6273 | 0.6260 | 0.5705 | 0.6067 |