Abstract
Dueto the limitations of digital image capturing equipment, it is usually difficult for the photographer to obtain a complete and clear image of a certain scene in the case of dual targets and multiple targets. This is because most digital imaging systems have a limited depth of field control range, so they can only focus on one or a few objects in the far or near distance, resulting in clear and blurred areas with clear boundaries, that is multi-focus image. This kind of image limits further image processing, such as target recognition, image segmentation, target tracking and so on. Often, two multi-focus images can basically integrate all scene information completely, and multiple multi-focus images can also be fused by cascading all images. Inspired by this, we propose a new image fusion method based on binocular depth estimation and binocular image difference, called depth-differential mapping fusion network (DDFN). In detail, DDFN is based on the idea of residual U-Net and the network structure. It takes two multi-focus images as input, extracts rich hierarchical features through the convolutional pooling pyramid, and learns the residuals between them and the corresponding groundtruth. In this process, DDFN will use their differential information to encode, merge the depth information, and finally perform the decoding process, so the features in the multi-focus image pair will be fully extracted. Finally a clear image without defocusing blur area is formed. We have conducted multiple ablation experiments and comparative experiments, furthermore, a large number of results fully demonstrate the effectiveness of our network structure.
Similar content being viewed by others
References
Alantas V, Kurban R (2010) Fusion of multi-focus images using differential evolution algorithm. Expert Syst Appl 37(12):8861–8870
Chen C, Mu S, Xiao W, Ye Z, Wu L, Ju Q (2019) Improving image captioning with conditional generative adversarial nets. In: Proceedings of the AAAI conference on artificial intelligence, vol 33, pp 8142–8150
Eigen D, Puhrsch C, Fergus R (2014) Depth map prediction from a single image using a multi-scale deep network. In: Advances in neural information processing systems, pp 2366–2374
Eigen D, Puhrsch C, Fergus R (2014) Depth map prediction from a single image using a multi-scale deep network. In: Advances in neural information processing systems, pp 2366–2374
Everingham M, Winn J (2011) The pascal visual object classes challenge 2012 (voc2012) development kit. Pattern Analysis, Statistical Modelling and Computational Learning, Tech. Rep, 8
Feng S, Zhao H, Shi F, Cheng X, Chen X (2020) CPFNet: context pyramid fusion network for medical image segmentation. IEEE Trans Med Imaging 1–1:99
Guo C, Li C, Guo J, Cong R, Fu H, Han P (2018) Hierarchical features driven residual learning for depth map super-resolution. IEEE Trans Image Process, 1–1
Han J, Pauwels EJ, De Zeeuw P (2013) Fast saliency-aware multi-modality image fusion. Neurocomputing 111:70–80
Hirschmuller H (2007) Stereo processing by semiglobal matching and mutual information. IEEE Trans Pattern Anal Mach Intell 30(2):328–341
Jaritz M, Vu TH, Charette RD, Wirbel E, Pérez P (2020) xmuda: cross-modal unsupervised domain adaptation for 3d semantic segmentation. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 12605–12614
Joshi K, Kirola M, Chaudhary S, Diwakar M, Joshi NK (2019) Multi-focus image fusion using discrete wavelet transform method. In: International conference on advances in engineering science management & technology (ICAESMT)-2019, Uttaranchal University, Dehradun, India
Jung H, Kim Y, Jang H, Ha N, Sohn K (2020) Unsupervised deep image fusion with structure tensor representations. IEEE Trans Image Process 29:3845–3858
Lai R, Li Y, Guan J, Xiong A (2019) Multi-scale visual attention deep convolutional neural network for multi-focus image fusion. IEEE Access 7:114385–114399
Li M, Cai W, Tan Z (2006) A region-based multi-sensor image fusion scheme using pulse-coupled neural network. Pattern Recogn Lett 27(16):1948–1956
Li J, Guo X, Lu G, Zhang B, Xu Y, Wu F, Zhang D (2020) DRPL: deep regression pair learning for multi-focus image fusion. IEEE Trans Image Process 29:4816–4831
Li S, Kang X, Fang L, Hu J, Yin H (2017) Pixel-level image fusion: a survey of the state of the art. Inform Fus 33:100–112
Li S, Kwok JT, Wang Y (2001) Combination of images with diverse focuses using the spatial frequency. Inform Fus 2(3):169–176
Li H, Li L, Zhang J (2015) Multi-focus image fusion based on sparse feature matrix decomposition and morphological filtering. Opt Commun 342:1–11
Liu Z, Blasch E, Xue Z, Zhao J, Laganiere R, Wu W (2011) Objective assessment of multiresolution image fusion algorithms for context enhancement in night vision: a comparative study. IEEE Trans Pattern Anal Mach Intell 34 (1):94–109
Liu Y, Chen X, Peng H (2017) Multi-focus image fusion with a deep convolutional neural network. Inform Fus 36:191–207
Liu Y, Jiang J, Sun J, Bai L, Wang Q (2020) A survey of depth estimation based on computer vision. In: 2020 IEEE Fifth international conference on data science in cyberspace (DSC), pp 135–141
Liu Y, Liu S, Wang Z (2015) Multi-focus image fusion with dense sift. Inform Fus 23:139–155
Ma B, Ban X, Huang H, Zhu Y (2019) SESF-fuse: an unsupervised deep model for multi-focus image fusion. arXiv:1908.01703
Ma T, Kuang P, Tian W (2020) An improved recurrent neural networks for 3D object reconstruction. Appl Intell 50:905–923
Ma H, Liao Q, Zhang J, Liu S, Xue JH (2020) An α-matte boundary defocus model-based cascaded network for multi-focus image fusion. IEEE Trans Image Process 29:8668–8679
Ma J, Ma Y, Li C (2019) Infrared and visible image fusion methods and applications: a survey. Information Fusion, 153–178
Ma J, Yu W, Chen C, Liang P, Jiang J (2020) Pan-GAN: an unsupervised pan-sharpening method for remote sensing image fusion. Inform Fus 62:110–120
Ma J, Zhou Z, Wang B, Dong M (2017) Multi-focus image fusion based on multi-scale focus measures and generalized random walk. In: 2017 36th Chinese control conference (CCC), pp 5464–5468
Ma J, Zhou Z, Wang B, Miao L, Zong H (2019) Multi-focus image fusion using boosted random walks-based algorithm with two-scale focus maps. Neurocomputing 335:9–20
Mao X, Li Q, Xie H, Lau RY, Wang Z, Paul Smolley S (2017) Least squares generative adversarial networks. In: Proceedings of the IEEE international conference on computer vision, pp 2794– 2802
Mingrui C, Junyi Y, Guanghui C (2015) Multi-focus image fusion algorithm using LP transformation and PCNN. In: 2015 6th IEEE International conference on software engineering and service science (ICSESS), pp 237–241
Naji MA, Aghagolzadeh A (2015) A new multi-focus image fusion technique based on variance in DCT domain. In: 2015 2nd International conference on knowledge-based engineering and innovation (KBEI), pp 478–484
Nejati M, Samavi S, Shirani S (2015) Multi-focus image fusion using dictionary-based sparse representation. Inform Fus 25:72–84
Nencini F, Garzelli A, Baronti S, Alparone L (2007) Remote sensing image fusion using the curvelet transform. Inform Fus 8(2):143–156
Qiu X, Li M, Zhang L, Yuan X (2019) Guided filter-based multi-focus image fusion through focus region detection. Signal Process Image Commun 72:35–46
Ronneberger O, Fischer P, Brox T (2015) U-net: convolutional networks for biomedical image segmentation. In: International conference on medical image computing and computer-assisted intervention, pp 234–241
Sahu A, Bhateja V, Krishn A (2014) Medical image fusion with Laplacian pyramids. In: 2014 International conference on medical imaging, m-health and emerging communication systems (MedCom), pp 448–453
Savić S, Babić Z (2012) Multifocus image fusion based on empirical mode decomposition. In: 19th IEEE International conference on systems, signals and image processing (IWSSIP)
Saxena A, Sun M, Ng AY (2008) Make3d: learning 3d scene structure from a single still image. IEEE Trans Pattern Anal Mach Intell 31(5):824–840
Shin K (2020) Binocular depth estimation, https://githubcom/kkenshin1/Binocular-Depth-Estimation
Shita D, Chanda B (2013) Multi-focus image fusion using a morphology-based focus measure in a quad-tree structure. Inform Fus 14(2):136–146
Tang H, Xiao B, Li W, Wang G (2018) Pixel convolutional neural network for multi-focus image fusion. Inform Sci 433:125–141
Xu H, Fan F, Zhang H, Le Z, Huang J (2020) A deep model for multi-focus image fusion based on gradients and connected regions. IEEE Access 8:26316–26327
Yan X, Gilani SZ, Qin H, Mian A (2018) Unsupervised deep multi-focus image fusion. ArXiv:1806.07272
Yang B, Li S (2010) Multifocus image fusion and restoration with sparse representation. IEEE Trans Instrum Measur 59(4):884–892
Zafar R, Farid MS, Khan MH (2020) Multi-focus image fusion: algorithms, evaluation, and a library. J Imag 6(7):60
Zhang X (2020) Multi-focus image fusion: a benchmark. arXiv:2005.01116
Zhang Q, Li G, Cao Y, Han J (2020) Multi-focus image fusion based on non-negative sparse representation and patch-level consistency rectification. Pattern Recogn 104:107325
Zhang Q, Liu Y, Blum RS, Han J, Tao D (2018) Sparse representation based multi-sensor image fusion for multi-focus and multi-modality images: a review. Inform Fus 40:57–75
Zhang Y, Liu Y, Sun P, Yan H, Zhao X, Zhang L (2020) IFCNN: a general image fusion framework based on convolutional neural network. Inform Fus 54:99–118
Zhang Q, Shi T, Wang F, Blum RS, Han J (2018) Robust sparse representation based multi-focus image fusion with dictionary construction and local spatial consistency. Pattern Recogn 83:299–313
Zhao W, Wang D, Lu H (2018) Multi-focus image fusion with a natural enhancement via a joint multi-level deeply supervised convolutional neural network. IEEE Trans Circ Syst Video Technol 29(4):1102–1115
Acknowledgments
This research was supported by the National Natural Science Foundation of China (61772319, 61773244, 61976125, 61976124), Shandong Natural Science Foundation of China (ZR2017MF049) and Yantai Key Research and Development Plan (2019XDHZ081).
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Jiang, L., Fan, H. & Li, J. DDFN: a depth-differential fusion network for multi-focus image. Multimed Tools Appl 81, 43013–43036 (2022). https://doi.org/10.1007/s11042-022-12075-z
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-022-12075-z