Abstract
Background: Leaning redundant and complementary relationships is a critical step in the human visual system. Inspired by the infrared cognition ability of crotalinae animals, we design a joint convolution auto-encoder (JCAE) network for infrared and visible image fusion. Methods: Our key insight is to feed infrared and visible pair images into the network simultaneously and separate an encoder stream into two private branches and one common branch, the private branch works for complementary features learning and the common branch does for redundant features learning. We also build two fusion rules to integrate redundant and complementary features into their fused feature which are then fed into the decoder layer to produce the final fused image. We detail the structure, fusion rule and explain its multi-task loss function. Results: Our JCAE network achieves good results in terms of both visual quality and objective evaluation metrics.










Similar content being viewed by others
References
Bertinetto L, Valmadre J, Henriques JF, Vedaldi A, Torr PHS (2016) Fully-convolutional siamese networks for object tracking. arXiv:1606.09549
Chen M (2016) Image fusion of visual and infrared image based on nsct and compressed sensing. J Image Graph 21(1):0039–0044
Deshmukh M, Bhosale U (2010) Image fusion and image quality assessment of fused images. Int J Image Process 4(5):484–508
Dong C, Chen C, Tang X (2016) Accelerating the super-resolution convolutional neural network. In: European conference on computer vision, pp 391–407
Epstein B, Meir R, Michaeli T (2017) Joint auto-encoders: a flexible multi-task learning framework. In: ArXiv e-prints
Feng Y, Chen F, Ji Y-M, Wu F, Sun J (2021) Efficient cross-modality graph reasoning for rgb-infrared person re-identification. IEEE Signal Process Lett 28:1425–1429
Han Y, Cai Y, Cao Y, Xu X (2013) A new image fusion performance metric based on visual information fidelity. Inf Fusion 14(2):127–135
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778
Li Q, Li J, Li Z (2009) Image fusion based on nonsubsampled contourlet transform and evaluation. Appl Res Comput 26(3):1138–1134
Li S, Kang X, Hu J (2013) Image fusion with guided filtering. IEEE Trans Image Process 22(7):2864–2875
Li H, Wu X (2018) Densefuse: a fusion approach to infrared and visible images. IEEE transactions on image processing: a publication of the IEEE Signal Processing Society
Liu Z, Blasch E, Xue Z, Zhao J, Laganiere R, Wu W (2011) Objective assessment of multiresolution image fusion algorithms for context enhancement in night vision: a comparative study. IEEE Trans Pattern Anal Mach Intell 34 (1):94–109
Liu Y, Liu S, Wang Z (2015) A general framework for image fusion based on multi-scale transform and sparse representation. Inf Fusion 24:147–164
Liu Y, Chen X, Ward RK, Wang ZJ (2016) Image fusion with convolutional sparse representation. IEEE Signal Process Lett 23(12):1882–1886
Liu Y, Chen X, Peng H, Wang Z (2017) Multi-focus image fusion with a deep convolutional neural network. Inf Fusion 36:191–207
Luo X, Zhang Z, Wu X (2016) A novel algorithm of remote sensing image fusion based on shift-invariant shearlet transform and regional selection. AEU-Int J Electron Commun 70(2):186–197
Luo X, Zhang Z, Zhang B, Wu X-J (2017) Image fusion with contextual statistical similarity and nonsubsampled shearlet transform. IEEE Sensors J 17(6):1760–1771
Luo X, Gao Y, Wang A, Zhang Z, Wu X-J (2021) Ifsepr: a general framework for image fusion based on separate representation learning. IEEE Trans Multimed:1–1
Luo X, Wang A, Zhang Z, Xiang X, Wu X-J (2021) Latraivf: an infrared and visible image fusion method based on latent regression and adversarial training. IEEE Trans Instrum Meas 70:1–16
Ma J, Chen C, Li C, Huang J (2016) Infrared and visible image fusion via gradient transfer and total variation minimization. Inf Fusion 31(C):100–109
Ma J, Ma Y, Li C (2018) Infrared and visible image fusion methods and applications: a survey. Inf Fusion 45:153–178
Ma J, Yu W, Liang P, Li C, Jiang J (2019) Fusiongan: a generative adversarial network for infrared and visible image fusion. Inf Fusion 48:11–26
Pang H, Zhu M, Guo L (2012) Multifocus color image fusion using quaternion wavelet transform. In: International congress on image and signal processing, pp 543–546
Prabhakar KR, Srikar VS, Babu RV (2017) Deepfuse: a deep unsupervised approach for exposure fusion with extreme exposure image pairs. In: IEEE international conference on computer vision, pp 4724–4732
Qi C, Ye W, Miao B, Qi G (2008) Fusion algorithm of infrared and visible light images based on nsct transform. Syst Eng Electron 30(4):593–596
Shelhamer E, Long J, Darrell T (2014) Fully convolutional networks for semantic segmentation. IEEE Trans Pattern Anal Mach Intell 39(4):1–1
Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. In: Computer science
Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition. In: Bengio Y, LeCun Y (eds) 3rd international conference on learning representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings
Toet A, Hogervorst MA (2012) Progress in color night vision. Opt Eng 51(1):1–20
Wang Z, Bovik AC (2002) A universal image quality index. IEEE Signal Process Lett 9(3):81–84
Wang W, Chang F (2011) A multi-focus image fusion method based on laplacian pyramid. J Comput 6(12):2559–2566
Wu F, Jing X-Y, Dong X, Hu R, Yue D, Wang L, Ji Y-M, Wang R, Chen G (2020) Intraspectrum discrimination and interspectrum correlation analysis deep network for multispectral face recognition. IEEE Trans Cybern 50 (3):1009–1022
Wu F, Jing X-Y, Feng Y, Ji Y-M, Wang R (2021) Spectrum-aware discriminative deep feature learning for multi-spectral face recognition. Pattern Recognit 111(2):107632
Xu H, Ma J, Jiang J, Guo X, Ling H (2020) U2fusion: a unified unsupervised image fusion network. IEEE Trans Pattern Anal Mach Intell
Yu Z, Yu L, Sun P, Yan H, Zhao X, Zhang L (2020) Ifcnn: a general image fusion framework based on convolutional neural network. Inf Fusion 54:99–118
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
This study was funded by the National Natural Science Foundation of P. R. China (grant number 61772237) and the Six Talent Peaks Project in Jiangsu Province (grant numbe XYDXX-030).
Ethics approval and consent to participate
This article does not contain any studies with human participants or animals performed by any of the authors.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Zhang, Z., Gao, Y., Xiong, M. et al. A joint convolution auto-encoder network for infrared and visible image fusion. Multimed Tools Appl 82, 29017–29035 (2023). https://doi.org/10.1007/s11042-023-14758-7
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-023-14758-7