Computational Statistics and Data Analysis 56 (2012) 2334–2346 Contents lists available at SciVerse ScienceDirect Computational Statistics and Data Analysis journal homepage: www.elsevier.com/locate/csda Supervised classification for functional data: A weighted distance approach Andrés M. Alonso , David Casado, Juan Romo Departamento de Estadística, Universidad Carlos III de Madrid, 28903, Spain article info Article history: Received 9 March 2010 Received in revised form 10 January 2012 Accepted 10 January 2012 Available online 24 January 2012 Keywords: Supervised classification Discriminant analysis Functional data Weighted distances abstract A natural methodology for discriminating functional data is based on the distances from the observation or its derivatives to group representative functions (usually the mean) or their derivatives. It is proposed to use a combination of these distances for supervised classification. Simulation studies show that this procedure performs very well, resulting in smaller testing classification errors. Applications to real data show that this technique behaves as well as – and in some cases better than – existing supervised classification methods for functions. © 2012 Elsevier B.V. All rights reserved. 1. Introduction Functional data have great – and growing – importance in Statistics. Most of the classical techniques for the finite- and high-dimensional frameworks have been adapted to cope with the infinite dimensions, but due to the curse of dimensionality, new and specific treatments are still required. As with other types of data, statisticians must put forward different steps – registration, missing data, representation, transformation, typicality – and tackle different tasks—modelization, discrimination or clustering, amongst others. In practice, curves can neither be registered continuously nor at infinitely many points. Thus, techniques dealing with high-dimensional data can sometimes be applied: Hastie et al. (1995), for example, adapt the discriminant analysis to cope with many highly correlated predictors, ‘‘such as those obtained by discretizing a function’’. Among the approaches designed for supervised classification or discrimination of functional data, some project the data onto a finite-dimensional space of functions and work with the coefficients. James and Hastie (2001) model the coefficients with ‘‘Gaussian distribution with common covariance matrix for all classes, by analogy with LDA [linear discriminant analysis]’’; their classification minimizes the distance to the group mean. The discriminant method of Hall et al. (2001) maximizes the likelihood, and although they propose a fully nonparametric density estimation, in practice multivariate Gaussian densities are considered, leading to quadratic discriminant analysis. Biau et al. (2003) apply k-nearest neighbors to the coefficients, while Rossi and Villa (2006) apply support vector machines. Berlinet et al. (2008) consider wavelet bases. The following proposals are designed to make direct use of the continuity of the functional data. Ferraty and Vieu (2003) classify new curves in the group with the highest posterior probability of membership kernel estimate. On the other hand, López-Pintado and Romo (2006) also take into account the continuous feature of the data and propose two classification methods based on the notion of depth for curves; in their first proposal new curves are assigned to the group with the closest trimmed mean, while the second method minimizes a weighted average distance to each element in the group. Nerini and Ghattas (2007) classify density functions with functional regression trees. Baíllo and Cuevas (2008) provide Corresponding author. E-mail addresses: andres.alonso@uc3m.es (A.M. Alonso), david.casado@uc3m.es (D. Casado), juan.romo@uc3m.es (J. Romo). 0167-9473/$ – see front matter © 2012 Elsevier B.V. All rights reserved. doi:10.1016/j.csda.2012.01.013