GENERATING VIRTUAL MICROPHONE SIGNALS USING GEOMETRICAL INFORMATION GATHERED BY DISTRIBUTED ARRAYS Giovanni Del Galdo 1 , Oliver Thiergart 2 , Tobias Weller 1 , and Emanu¨ el A.P. Habets 2 1 Fraunhofer Institute for Integrated Circuits IIS, Erlangen, Germany 2 International Audio Laboratories Erlangen, Germany Email: ame-info@iis.fraunhofer.de ABSTRACT Conventional recording techniques for spatial audio are limited to the fact that the spatial image obtained is always relative to the po- sition in which the microphones have been physically placed. In many applications, however, it is desired to place the microphones outside the sound scene and yet be able to capture the sound from an arbitrary perspective. This contribution proposes a method to place a virtual microphone at an arbitrary point in space, by computing a signal perceptually similar to the one which would have been picked up if the microphone had been physically placed in the sound scene. The method relies on a parametric model of the sound field based on point-like isotropic sound sources. The required geometrical in- formation is gathered by two or more distributed microphone arrays. Measurement results demonstrate the applicability of the proposed method and reveal its limitations. Index TermsSpatial sound, Sound localization, Audio recording, Parameter estimation 1. INTRODUCTION Spatial sound acquisition aims at capturing either an entire sound scene or just certain desired components, depending on the applica- tion at hand. Several recording techniques providing different advan- tages and drawbacks are available for these purposes. For instance, close talking microphones are often used for recording individual sound sources with high SNR and low reverberation, while more distant configurations such as XY stereophony represent a way for capturing the spatial image of an entire sound scene. More flexibility in terms of directivity can be achieved with beamforming, where a microphone array can be used to realize steerable pick-up patterns. Even more flexibility is provided by parametric methods, such as directional audio coding (DirAC) [1], in which it is possible to real- ize spatial filters with arbitrary pick-up patterns [2] as well as other signal processing manipulations of the sound scene [3, 4]. All these methods have in common that they are limited to a rep- resentation of the sound field with respect to only one point, namely the measurement location. Thus, the required microphones must be placed at very specific, carefully selected positions, e. g., close to the sources or such that the spatial image can be captured optimally. In many applications however, this is not feasible and therefore it would be beneficial to place several microphones further away from the sound sources and still be able to capture the sound as desired. There exist several field reconstruction methods for estimating the sound field in a point in space other than where it was measured. One method is acoustic holography [5], which allows to compute the sound field at any point within an arbitrary volume given that the sound pressure and particle velocity is known on its entire surface. Therefore, when the volume is large, an unpractically large num- ber of sensors is required. Moreover, the method assumes that no sound sources are present inside the volume, making the algorithm unfeasible for our needs. The related wave field extrapolation [5] aims at extrapolating the known sound field on the surface of a vol- ume to outer regions. The extrapolation accuracy however degrades rapidly for larger extrapolation distances as well as for extrapola- tions towards directions orthogonal to the direction of propagation of the sound [6]. In [7] a plane wave model is assumed, such that the field extrapolation is possible only in points far from the actual sound sources, i. e., close to the measurement point. To overcome the drawbacks of these field reconstructing meth- ods, this contribution proposes a parametric method capable of es- timating the sound signal of a virtual microphone placed at an arbi- trary location. In contrast to the methods previously described, the proposed method does not aim directly at reconstructing the sound field, but rather at providing sound that is perceptually similar to the one which would be picked up by a microphone physically placed at this location. This is possible thanks to a parametric model of the sound field based on isotropic point-like sound sources (IPLS). The required geometrical information, namely the instantaneous po- sition of all IPLS, is gathered via triangulation of the directions of arrival (DOA) estimated with two or more distributed microphone ar- rays. Therefore, knowledge on the relative position and orientation of the arrays is required. Notwithstanding, no a priori knowledge on the number and position of the actual sound sources is neces- sary. Given the parametric nature of the method, the virtual micro- phone can possess an arbitrary directivity pattern as well as physical or non-physical behaviors, e. g., with respect to the pressure decay with distance. The presented approach is verified by studying the pa- rameter estimation accuracy based on measurements in a reverberant environment. The paper is structured as follows: In Section 2 the sound field model is introduced and the geometric parameter estimation algo- rithm is derived. In Section 3 the virtual microphone approach is presented and discussed in detail. The algorithm is verified with measurement results in Section 4. Section 5 concludes the paper. 2. GEOMETRIC PARAMETER ESTIMATION 2.1. Sound Field Model The sound field is analyzed in the time-frequency domain, for in- stance obtained via a short-time Fourier transform (STFT), in which k and n denote the frequency and time indices, respectively. The complex pressure P v (k, n) at an arbitrary position pv for a certain 2011 Joint Workshop on Hands-free Speech Communication and Microphone Arrays May 30 - June 1, 2011 978-1-4577-0999-9/11/$26.00 ©2011 IEEE 185