AbstractIn this paper, we study the application of Extreme Learning Machine (ELM) algorithm for single layered feedforward neural networks to non-linear chaotic time series problems. In this algorithm the input weights and the hidden layer bias are randomly chosen. The ELM formulation leads to solving a system of linear equations in terms of the unknown weights connecting the hidden layer to the output layer. The solution of this general system of linear equations will be obtained using Moore-Penrose generalized pseudo inverse. For the study of the application of the method we consider the time series generated by the Mackey Glass delay differential equation with different time delays, Santa Fe A and UCR heart beat rate ECG time series. For the choice of sigmoid, sin and hardlim activation functions the optimal values for the memory order and the number of hidden neurons which give the best prediction performance in terms of root mean square error are determined. It is observed that the results obtained are in close agreement with the exact solution of the problems considered which clearly shows that ELM is a very promising alternative method for time series prediction. KeywordsChaotic time series, Extreme learning machine, Generalization performance. I. INTRODUCTION RTIFICIAL Neural Networks (ANNs) have been extensively applied for pattern classification and regression problems. The major reason for the success of ANNs is their ability in obtaining a non-linear approximation model function describing the association between the dependent and independent variables using the given input samples. Since ANNs adaptively select the model from the features presented in the input data, they are applied to a large number of classes of problems of importance like optical character recognition [7], face detection [11], gene prediction [14], credit scoring [6] and time series forecasting [12], [17].Though ANNs have many advantages such as better approximation capabilities and simple network structures, however, it suffers from several problems such as presence of local minima's, imprecise learning rate, selection of the number of hidden neurons and over fitting. Moreover, the gradient descent based learning algorithms such as Back Propagation (BP) will generally lead to slow convergence during the training of the networks. Rampal Singh is with Department of Computer Science, Deen Dayal Upadhyaya College, University of Delhi, New Delhi-110015, India (phone:+91-11-27570620, 09350647546, e-mail: rpsrana@ddu.du.ac.in). S. Balasundaram is with School of Computer & Systems Sciences, Jawaharlal Nehru University, New Delhi-110067, India (phone: +91-11- 26704724, e-mail: balajnu@hotmail.com). Time series forecasting is an important and challenging problem of regression. In the regression problem by analyzing the given input samples the best fit functional model describing the relationship between the dependent and independent variables is obtained. There are many prediction models exist in the literature [4], [9], [12] for time series. The important and widely used among them are Auto Regressive Integrated Moving Average (ARIMA) [1], ANNs [12], [17] and Support Vector Regression (SVR) [8], [9], [13], [15] methods. Among the above methods, ARIMA assumes the existence of a linear relationship in the time series values, i.e. its prediction value will be a linear function of the past observations and therefore it is not always suitable for complex real world problems [18]. Also it is proposed to combine several methods in order to obtain improved forecasting accuracy. For the study of a hybrid approach of combining ARIMA and ANN for time series forecasting we refer the reader to [18]. For time series involving seasonality, combining Seasonal time series ARIMA (SARIMA) and ANN is discussed in [16] and for the study of a combined SARIMA and SVR approach see [3]. Huang et al [5] have proposed a new learning algorithm for Single hidden Layer Feedforward Neural Network (SLFN) architecture called Extreme Learning Machine (ELM) which overcomes the problems caused by gradient descent based algorithms such as BP applied in ANNs. In this algorithm the input weights and the hidden layer bias are randomly chosen. The ELM formulation leads to solving a system of linear equations in terms of the unknown weights connecting the hidden layer to the output layer. The solution of this general system of linear equations is obtained using Moore-Penrose generalized pseudo inverse [10]. In this work we discuss briefly the ELM algorithm and study its feasibility of application for chaotic time series prediction problems. Throughout this paper, we assume all vectors to be column vectors. For any two vectors x, y in the m- dimensional real space m , we denote the inner product of the two vectors by y x . where x is the transpose of the vector x and the norm of a vector by . || || The paper is organized as follows. In Section 2, we define Moore- Penrose generalized inverse, the minimum norm least squares solution of a general linear system of equations and state the relation between them. In Section 3, we revive the ELM algorithm for SLFN. For the application of this algorithm we considered Mackey Glass delay differential equation with different time delays, Santa Fe-A and UCR heart beat rate (ECG) time series in Section 4 and the results Application of Extreme Learning Machine Method for Time Series Analysis Rampal Singh, and S. Balasundaram A International Journal of Intelligent Systems and Technologies 2;4 © www.waset.org Fall 2007 256