Three-dimensional Gaussian Mixture Texture Model Michal Haindl and Vojtˇ ech Havl´ ıˇ cek Institute of Information Theory and Automation of the CAS Prague, Czech Republic 182 08 Email: {haindl,havlicek}@utia.cas.cz Abstract—Visual texture modeling based on multidimensional mathematical models is the prerequisite for both robust material recognition as well as for image restoration, compression or numerous physically correct virtual reality applications. A novel multispectral visual texture modeling method based on a de- scriptive, unusually complex, three-dimensional, spatial Gaussian mixture model is presented. Texture synthesis benefits from easy computation of arbitrary conditional distributions from the model. The model is inherently multispectral thus it does not suffer with the spectral quality compromises of the spectrally factorized alternative approaches. The model is especially well suited for multispectral textile textures and it can also describe the most advanced textural representation in the form of a bidirectional texture function (BTF). I. I NTRODUCTION Human observer’s visual scene recognition is based on shapes and materials. Unfortunately, the surface material ap- pearance vastly changes under variable observation conditions which significantly complicates and negatively affects its mathematical synthesis as well as machine analysis. Reliable computer-based interpretation of visual information which would approach human cognitive capabilities is very challeng- ing and impossible without significant improvement of the cor- responding sophisticated visual information models capable to handle huge variations of possible observation conditions. The appropriate paradigm for such a surface material reflectance function models is a multidimensional visual texture. Texture synthesis approaches may be divided primarily into sampling and model-based methods. Sampling methods [1], [2], [3], [4], [5], [6] rely on sophisticated sampling from real texture measurements while the model-based techniques [7], [8], [9], [10], [11], [12], [13] describe texture data using multidimensional mathematical models and their synthesis is based on the estimated model parameters only. Generative visual texture models are useful not only for modelling physically correct virtual objects material surfaces in virtual or augmented reality environments, image restoration or compression but also for contextual recognition applications such as segmentation, classification or image retrieval. Physically correct surface material reflectance model (RM) is sixteen-dimensional function [14] RM (λ i ,x i ,y i ,z i ,t i i i v ,x v ,y v ,z v ,t v v v i,T v,T ). RM describes incident light with spectral value λ i illuminating surface location x i ,y i ,z i in time t i under spherical reflectance angle θ i i and observed at time t v from surface location x v ,y v ,z v under spherical reflectance angle θ v v and spec- trum λ v . θ i,T v,T are the corresponding transmittance angles. The model height parameters z i ,z v indicate that even radiance along light rays is not constant but depends on the height. Such a RM model is too complex and there neither exist any measurement of such data nor any mathematical representation allowing its synthesis. One of the early compromised attempts to capture real material appearance was done by Nicodemus et al. [15] and later elaborated by Dana et al. [16] in the form of Bidirectional Texture Function (BTF). Even if a BTF model assumes several strong simplifying assumptions [17], [18], [14] its measurement, compression and synthesis is on the leading edge of current mathematical modelling and techno- logical capabilities. BTF is a seven-dimensional function [14] which considers not only measurement dependency on planar material position and spectral channel but also its dependence on illumination and viewing angles: BTF θiivv r) (1) where θ,φ are elevation and azimuthal angles of illumination and view direction vector, the multiindex ˜ r = [r 1 ,r 2 ,r 3 ] specifies planar horizontal and vertical position in a material sample image and r 3 is the spectral index. Reliable parameters estimation of such a seven-dimensional stochastic model is very difficult not only because it requires very demanding numerical optimization but because the learning textural data are always too limited to obtain robust and reliable estimates. The solution is to factorize the original seven-dimensional measurement space into a set of less dimensional textural fac- tors. The realistic modeling strives not necessarily to recover the exact pixel-wise correspondence with some original target texture but rather a texture which is visually indiscernible from the original one. In our previous paper [19] we have introduced three two- dimensional probabilistic mixture models, where a measured 3D multi-spectral texture had to be spectrally factorized and the corresponding multivariate mixture models were further learned from single orthogonal mono-spectral components and used to synthesise and enlarge these mono-spectral factor components. The presented model (BTF-3DGMM), on the contrary, is fully multispectral and thus it does not need to compromise spectral modeling quality in multicoloured textures. We applied this model for simpler task of high quality texture restoration [20] where the model can exploit information from corrupted textural data. The presented ap- plication to BTF (color / multispectral) texture synthesis is 2016 23rd International Conference on Pattern Recognition (ICPR) Cancún Center, Cancún, México, December 4-8, 2016 978-1-5090-4846-5/16/$31.00 ©2016 IEEE 2026