Abstract
Most multi-focus image fusion evaluation methods are based on focus detection and measuring the similarity between the fused image and the whole source images including defocused regions, which is liable to result in the difference between the evaluation result and the real image fusion quality. To overcome the problem above, we proposed a novel objective measure for multi-focus image fusion assessment in feature level. Firstly, the corners in source images and the fused image are separately detected based on Smallest Univalue Segment Assimilating Nucleus (SUSAN) algorithm. Then, a corner similarity measure based on overlapping rate is proposed to measure the fusion quality. The proposed method avoids focus detection in the assessment procedure, which make evaluation results more reliable. Experimental results demonstrate that the proposed measure is more consistent with subjective evaluation. Comparing with other objective metrics, two meta-measures including correct ranking (CR) and subjective relevance (R) give our proposed measure the highest scores, 0.8377 and 0.7384, respectively. The area under the ROC curve (AUC) gives our metric the second-best scores of 0.8428.
Similar content being viewed by others
References
Bouzos O, Andreadis I, Mitianoudis N (2019) Conditional random field model for robust multi-focus image fusion. IEEE Trans Image Process 28(11):5636–5648
Chen YB, Guan JW, Cham WK (2018) Robust multi-focus image fusion using edge model and multi-matting. IEEE Trans Image Process 27(3):1526–1541
Fang Y, Zhu H, Ma K, Wang Z, Li S (2019) Perceptual evaluation for multi-exposure image fusion of dynamic scenes. IEEE Trans Image Process 29:1127–1138
Han YY (2015) Multimodal gray image fusion metric based on complex wavelet structural similarity. Optik 126(24):5842–5844
Han Y, Cai YZ, Cao Y et al (2013) A new image fusion performance metric based on visual information fidelity. Information fusion 14(2):127–135
Hassen R, Wang Z, Salama MMA (2015) Objective quality assessment for multiexposure multifocus image fusion. IEEE Trans Image Process 24(9):2712–2724
Hossny M, Nahavandi S, Creighton D (2008) Comments on ‘Information measure for performance of image fusion’. Electron Lett 44(18):1066–1067
Li ST, Kang XD, Fang LY et al (2017) Pixel-level image fusion: a survey of the state of the art. Information fusion 33:100–112
Li H, Zhang L, Jiang M, Li Y (2021) Multi-focus image fusion algorithm based on supervised learning for fully convolutional neural network. Pattern Recogn Lett 141:45–53
Liu Y, Wang L, Cheng J, Li C, Chen X (2020) Multi-focus image fusion: a survey of the state of the art. Information Fusion 64:71–91
Ma K, Zeng K, Wang Z (2015) Perceptual quality assessment for multi-exposure image fusion. IEEE Trans Image Process 24(11):3345–3356
Martinez J, Pistonesi S, Maciel MC et al (2019) Multi-scale fidelity measure for image fusion quality assessment. Information Fusion 50:197–211
Nejati M, Samavi S, Shirani S (2015) Multi-focus image fusion using dictionary-based sparse representation. Information Fusion 25:72–84. https://doi.org/10.1016/j.inffus.2014.10.004
Nian Z, Jung C (2019) Cnn-based multi-focus image fusion with light field data. 2019 IEEE International Conf Image Process (ICIP), 1044–1048
Petrovic V (2007) Subjective tests for image fusion evaluation and objective metric validation. Information fusion 8(2):208–216
Petrovic V, Dimitrijevic V (2015) Focused pooling for image fusion evaluation. Information fusion 22:119–126
Piella G, Heijmans H (2003) A new quality metric for image fusion. Proceedings of the 10th International Conference on Image Processing, 3: 173–176
Possa PR, Mahmoudi SA, Harb N, Valderrama C, Manneback P (2014) A multi-resolution FPGA-based architecture for real-time edge and corner detection. IEEE Trans Comput 63(10):2376–2388
Qiu X, Li M, Zhang L, Yuan X (2019) Guided filter-based multi-focus image fusion through focus region detection. Signal Process Image Commun 72:35–46
Qu G, Zhang D, Yan P (2002) Information measure for performance of image fusion. Electron Lett 38(7):313–315
Smith SM, Brady JM (1997) SUSAN-a new approach to low level image processing. Int J Comput Vis 23(1):45–78
Tan J, Zhang T, Zhao L, Luo X, Tang YY (2021) Multi-focus image fusion with geometrical sparse representation. Signal Process Image Commun 92:116130
Tang H, Xiao B, Li WS et al (2018) Pixel convolutional neural network for multi-focus image fusion. Inf Sci 433:125–141
Tsai CC Standard images for multifocus image fusion (https://www.mathworks.com/matlabcentral/fileexchange/45992-standard-images-for-multifocus-image-fusion). MATLAB Central File Exchange, Retrieved March 10, 2020.
Wang Z, Bovik AC (2002) A universal image quality index. Signal Processing Letters 9(3):81–84
Xing L, Cai L, Zeng HQ et al (2018) A multi-scale contrast-based image quality assessment model for multi-exposure image fusion. Signal Process 145:233–240
Xu K, Qin Z, Wang G et al (2018) Multi-focus image fusion using fully convolutional two-stream network for visual sensors. KSII Transactions on Internet and Information Systems 12(5):2253–2272
Xydeas S, Petrovic V (2000) Objective image fusion performance measure. Electron Lett 36(4):308–309
Yu S, Li X, Ma M, Zhang X, Chen S (2021) Multi-focus image fusion based on L1 image transform. Multimed Tools Appl 80(4):5673–5700
Zhang XL, Li XF, Feng YC et al (2015) The use of ROC and AUC in the validation of objective image fusion evaluation metrics. Signal Process 115:38–48
Zhang Y, Bai X, Wang T (2017) Boundary finding based multi-focus image fusion through multi-scale morphological focus-measure. Information fusion 35:81–101
Zhao W, Wang D, Lu H (2019) Multi-focus image fusion with a natural enhancement via a joint multi-level deeply supervised convolutional neural network. IEEE Transactions on Circuits and Systems for Video Technology 29(4):1102–1115
Zhu H (2020) Image quality assessment model based on multi-feature fusion of energy internet of things. Futur Gener Comput Syst 112:501–506
Xiao Zuzhang. multi-focus-image-fusion-dataset (https://www.mathworks.com/matlabcentral/fileexchange/70109-multi-focus-image-fusion-dataset), MATLAB Central File Exchange Retrieved July 8, 2021.
Acknowledgements
The work was supported by Youth Growth Science and Technology Plan Project of Jilin Provincial Department of Science and Technology (NO.20210508039RQ), “Thirteenth Five-Year Plan” Scientific Research Planning Project of Education Department of Jilin Province (NO.JJKH20200678KJ, NO.JJKH20210752KJ, NO.JJKH20200677KJ), Fundamental Research Funds for the Central Universities, JLU(NO.93K172020K05), and National Natural Science Foundation of China (NO.61806024, NO.61876070, NO.61801190).
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Feng, Y., Guo, R., Shen, X. et al. A measure for the evaluation of multi-focus image fusion at feature level. Multimed Tools Appl 81, 18053–18071 (2022). https://doi.org/10.1007/s11042-022-11976-3
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-022-11976-3