Omnidirectional Depth Computation from a Single Image * Radu Orghidan and El Mustapha Mouaddib Centre de Robotique Electrotechnique et d’Automatique Universit´ e de Picardie Jules Verne, Amiens, France. mouaddib@u-picardie.fr Joaquim Salvi Computer Vision and Robotics Group Institute of Informatics and Applications University of Girona, Girona, Catalonia, Spain. {radu,qsalvi}@eia.udg.es Abstract— Omnidirectional cameras offer a much wider field of view than the perspective ones and alleviate the problems due to occlusions. However, both types of cameras suffer from the lack of depth perception. A practical method for obtaining depth in computer vision is to project a known structured light pattern on the scene avoiding the problems and costs involved by stereo vision. This paper is focused on the idea of combining omnidirectional vision and structured light with the aim to provide 3D information about the scene. The resulting sensor is formed by a single catadioptric camera and an omnidirectional light projector. It is also discussed how this sensor can be used in robot navigation applications. Index Terms— catadioptrics, omnidirectional vision, cali- bration, structured light, 3D reconstruction I. I NTRODUCTION The omnidirectional vision sensors enhance the field of view of traditional cameras by means of special optics, structures of still or gyratory cameras or combinations of lenses and mirrors. Yagi [21] surveyed the existing techniques for building cameras with a wide field of view and Svoboda [19] proposed several classifications of the existing omnidirectional cameras according to their most important features. The catadioptric sensors use at least one mirror coupled to a conventional camera. The catadioptric cameras can be classified depending on the way they gather the light rays. When all the observed light rays cross into a point, called focus, the sensors are known as Single View Point (SVP). The class of catadioptric sensors that have a SVP was derived by Baker and Nayar [1]. The SVP is a desir- able property that enables distortion-free reconstruction of panoramic images in a familiar form for the human users. The catadioptric sensors that do not possess a single focal point (non-SVP) are less used but proved to be helpful for applications with specific requirements such as prescribed distortions [10] or with linear projection constraints [4]. Still, neither the standard cameras nor the catadioptric ones can provide depth information of the scene when used independently. Stereoscopic vision combines separate images taken from distinct points of view and permits to visually per- * This work is partially supported by the Spanish project CICYT TIC 2003-08106-C02-02 and by the AIRE mobility grant provided by the Generalitat of Catalunya that allowed a four month stay in the CREA lab from Amiens, France ceive depth. Stereo catadioptric sensors are special struc- tures of mirrors and lenses designed for obtaining depth from images with a wide field of view. In order to obtain distinct points of view of the scene the camera is pointed towards a structure of convex [13] [9] [2] or planar [8] mirrors. The 3D information is obtained by triangulation. This method leans on the assumption that the correspon- dences of the points between the observed images can be accurately found. However, correspondence matching is deteriorated in the case of catadioptric sensors because the resolution of the omnidirectional images is lower than the resolution of the conventional ones since the number of the scene points is significantly different while both images are represented using the same number of pixels. A solution to this problem is the use of a structured light pattern projected onto the scene [18] [17]. Using this technique is similar to placing visible landmarks in the scene so that image points can be identified and matched faster. It is noticeable that the use of 360 degrees images and of scene-depth information is ideal for robot navigation tasks. An efficient approach to the navigation problem is to choose a reasonable tradeoff between localization accuracy and travelled distance. Similarly, in real life a high precision of movements is required in small spaces such as narrow halls or offices while a low accuracy is needed in wide areas where the possible obstacles are more spaced. Consequently, robot navigation can be divided in two branches as shown by Gaspar [20]: topological navigation and visual path following. Topological navigation gives a qualitative characterization of the robot’s global position using omnidirectional images. This technique is suitable for navigation tasks that implies travelling large distances with low precision of localization. Visual path following is mainly used for short-distance segments that require high accuracy navigation. The goal of this paper is to present an omnidirectional sensor that provides 3D information using a single camera. From this point of view, a robot navigation application is also discussed. The sensor is formed by a single- camera catadioptric configuration with an embedded om- nidirectional structured light projector. By mounting the omnidirectional sensor on a mobile robot applications such as 3D map building, robot navigation and localization, active surveillance with real-time object detection or 3D Proceedings of the 2005 IEEE International Conference on Robotics and Automation Barcelona, Spain, April 2005 0-7803-8914-X/05/$20.00 ©2005 IEEE. 1222 Authorized licensed use limited to: UNIVERSITAT DE GIRONA. Downloaded on April 26,2010 at 10:44:08 UTC from IEEE Xplore. Restrictions apply.