Please cite this article in press as: A.S. Malik, et al., A Fuzzy-Neural approach for estimation of depth map using focus, Appl. Soft Comput.
J. (2010), doi:10.1016/j.asoc.2010.05.030
ARTICLE IN PRESS
G Model
ASOC-901; No. of Pages 14
Applied Soft Computing xxx (2010) xxx–xxx
Contents lists available at ScienceDirect
Applied Soft Computing
journal homepage: www.elsevier.com/locate/asoc
A Fuzzy-Neural approach for estimation of depth map using focus
Aamir Saeed Malik
a,∗
, Humaira Nisar
b
, Tae-Sun Choi
b
a
Department of Electrical & Electronic Engineering, Universiti Teknologi Petronas, Bandar Sei Iskandar, 31750 Tronoh, Perak, Malaysia
b
Department of Mechatronics, Gwangju Institute of Science and Technology, Republic of Korea
article info
Article history:
Received 1 October 2009
Received in revised form 5 May 2010
Accepted 30 May 2010
Available online xxx
Keywords:
Depth map estimation
3D shape recovery
Fuzzy logic
Neural Network
Back propagation
abstract
Depth map is used for recovery of three-dimensional structure of the object which is required in many
high level vision applications. In this paper, we present a new algorithm for the estimation of depth map
for three-dimensional shape recovery. This algorithm is based on Fuzzy-Neural approach using shape
from focus (SFF). A Fuzzy Inference System (FIS) is designed for the calculation of the depth map and an
initial set of membership functions and fuzzy rules are proposed. Then Neural Network is used to train the
FIS. The training is done using back propagation algorithm in combination with the least squares method.
Hence, a new set of input membership functions are generated while discarding the initial ones. Lastly,
the trained FIS is used to obtain final depth map. The results are compared with five other methods
including the traditional SFF method and the Focused Image Surface SFF method (FISM). Six different
types of objects are used for testing the proposed algorithm.
© 2010 Elsevier B.V. All rights reserved.
1. Introduction
Depth estimation is an important parameter for robot vision,
3D scene modeling, motion, etc. using single or multiple cameras.
Various passive techniques have been proposed in the literature
to estimate depth including depth from focus, defocus, stereo, etc.
[1,2]. We study depth/shape from focus. Under suitable conditions,
different three-dimensional (3D) scenes generate different accom-
modation cues, and therefore it is possible to distinguish them
from their images. Focus is an accommodation cue that can be
measured from blurring in the image, which increases with the dis-
tance of imaging system from the plane of focus. Techniques that
retrieve spatial information, by looking at multiple images of the
same scene, taken with different geometries or position of imaging
devices, are classified as shape from focus (SFF) or depth from focus
[3,4]. The objective of shape from focus (SFF) is to find out the depth
of every point of the object from the camera lens.
The shape from focus method is based on obtaining a series
of images for an object by a single camera at different distances
between lens and object. Then by using a quality measure of focus
for each point, the image frame for which that point is well-focused
is calculated. The basic image formation geometry is shown in Fig. 1.
In the figure, the parameters related to the camera are already
known. We need to calculate ‘u’, i.e., depth of object from the lens.
We make a depth map by calculating ‘u’ for every pixel. We can use
∗
Corresponding author.
E-mail address: aamir saeed@petronas.com.my (A.S. Malik).
the lens formula to calculate ‘u’. If the image detector (ID) is placed
exactly at a distance v, sharp image P
′
of the point P is formed at v
(see Fig. 1) while P
′′
is the blurred circle (defocused points). Then
the relationship between the object distance u, focal distance of the
lens f, and the image distance v is given by the Gaussian lens law:
1
f
=
1
u
+
1
v
Given the position of the focused image for each point (image
frame for which a point is well-focused), its position in the scene is
uniquely determined by using the lens formula.
To measure the true focus point requires large number of images
with incremental distance moved towards focus plane. To detect
the true focus point from finite number of images, various focus
measures have been proposed by researchers. A focus measure is
a quantity which measures the degree of blurring of an image; its
value is a maximum when the image is best focused and decreases
as blurring increases. Fig. 2 shows a focus measure curve for a
point in the image calculated using Sum of Modified Laplacian
(SML) (camera parameters: focal length = 35 mm, aperture diam-
eter = 35/4 mm, F-number = 4). In Fig. 2, the peak area shows the
pixel being focused in those frames while being defocused in other
frames. However, as we move away (on both sides of the peak), the
value decreases indicating increase in blur.
Depth estimation, like all aspects of computer vision, has some
inherent error. Some sources of this error are noise in the image,
imprecision in computations, ambiguity in interpreting depth
information, camera lens aberrations, etc. These errors combine to
create an uncertainty and error in the result. Therefore, various
1568-4946/$ – see front matter © 2010 Elsevier B.V. All rights reserved.
doi:10.1016/j.asoc.2010.05.030