SIViP DOI 10.1007/s11760-017-1155-y ORIGINAL PAPER Block-based image fusion using multi-scale analysis to enhance depth of field and dynamic range Vishal Chaudhary 1 · Vinay Kumar 1 Received: 12 February 2017 / Revised: 21 June 2017 / Accepted: 26 July 2017 © Springer-Verlag London Ltd. 2017 Abstract A novel technique for integrating information by exploring multi-scale positions with block-based fusion and to address blocking effects is discussed in the present manuscript. The source images are split into local and global layers using neighbor distance filter by extracting informa- tion at multi-scale positions. Recombined local and global layers are constructed using block-based and weighted aver- age methods, respectively. The spatial frequency as well as exposedness factor is used to find the texture information and exposure level for respective blocks. Resulting local and global layers are then fused to generate final fused image. The method is applicable to any number of source images. Extensive experimental results are provided to show the effectiveness of proposed technique. Keywords Depth of field (DoF) · Image fusion · Multi-focus image · Multi-exposure image · Shutter speed 1 Introduction Image fusion [1] is a process of collecting complemen- tary information from number of captured images (source images) of same scene and combine them into a single image, refer Fig. 1. There are wide range of applications of image fusion right from defense to medical systems [2 6]. Image fusion can be categorized in number of ways Electronic supplementary material The online version of this article (doi:10.1007/s11760-017-1155-y) contains supplementary material, which is available to authorized users. B Vishal Chaudhary vishalch.nsit@gmail.com 1 ECED, Thapar University, Patiala, India based on image capturing conditions: multi-focus, multi- exposure, multi-view and multi-modal image fusion. Present manuscript utilizes multi-focus and multi-exposure methods to combine entire information into single image. Varying shutter speed captures diverse details of scene as source images [7]. To get an image with optimum detail, com- plementary information is combined from captured images. The process is known as multi-exposure image fusion. Fig- ure 1 shows the images with different exposure times and corresponding fused image. Lens is another device feature which directs the infor- mation from scene to sensor. It maintain sharpness in some sections while blurs remaining, the process is defined as depth of field (DoF) [8, 9]. Figure 2 shows images with limited DoF and fused image. To improve the DoF, source images with diverse focused regions are combined and the process is known as multi-focus image fusion. The techniques available in literature for multi-focus image fusion are broadly classified as spatial and transform domain techniques. In spatial domain, operations are directly performed on pixels. The simplest way to perform this is by averaging intensities of source images on pixel to pixel basis. De et al. [8] proposed a method using mathematical morphology on individual pixels. Bai et al. [10] used morphology based top-hat transform to perform fusion on pixel to pixel basis. These methods are prone to contrast reduction and sensitive to misregistration. To overcome these problems, researchers have adopted block-based fusion, where blocks with highest information measure among source images are combined to construct fused image. Li et al. [11] proposed block-based fusion based on spatial frequency components. Li et al. [12] trained neural networks to select the best block using spa- tial frequency, visibility and edge. Miao and Wang [13] and Huang and Jing [14] used pulse coupled neural network, 123