Adaptive Activity Recognition with Dynamic Heterogeneous Sensor Fusion Ming Zeng, Xiao Wang, Le T. Nguyen, Pang Wu, Ole J. Mengshoel, Joy Zhang Department of Electrical and Computer Engineering Carnegie Mellon University Moffett Field, CA, USA Email: {ming.zeng, sean.wang, le.nguyen, pang.wu, ole.mengshoel, joy.zhang}@sv.cmu.edu Abstract—In spite of extensive research in the last decade, activity recognition still faces many challenges for real-world applications. On one hand, when attempting to recognize various activities, different sensors play different on different activity classes. This heterogeneity raises the necessity of learning the optimal combination of sensor modalities for each activity. On the other hand, users may consistently or occasionally annotate activities. To boost recognition accuracy, we need to incorporate the user input and incrementally adjust the model. To tackle these challenges, we propose an adaptive activity recognition with dynamic heterogeneous sensor fusion framework. We dynamically fuse various modalities to characterize different activities. The model is consistently updated upon arrival of newly labeled data. To evaluate the effectiveness of the proposed framework, we incorporate it into popular feature transformation algorithms, e.g., Linear Discriminant Analysis, Marginal Fisher’s Analysis, and Maximum Mutual Information in the proposed framework. Finally, we carry out experiments on a real-world dataset col- lected over two weeks. The result demonstrates the practical implication of our framework and its advantage over existing approaches. KeywordsActivity Recognition, Deep Learning, Convolutional Neural Network I. I NTRODUCTION The rapid spread of wearable devices with sensing capa- bilities offers the opportunity for human activity recognition. Knowing a user’s activity over a period of time enables applications such as continuous monitoring of user behavior, physical activity monitoring [18], abnormal activity detec- tion [3], elderly care [23] and physical activity recognition [2]. The activity recognition is usually formulated as a classi- fication problem [15]. Many classification methods have been leveraged in previous studies. The decision table, decision tree and naive Bayes classifier are experimented to recognize twenty predefined daily activities [2]. The support vector machine (SVM) and k-nearest neighbor (kNN) algorithm are used to perform fall detection [26]. The linear discriminant analysis and hidden Markov models are introduced to recog- nize predefined workshop activities [13]. However, most of the aforementioned activity recognition approaches frame activity recognition as a “static” machine learning problem, which assumes the types of activities to be recognized are predefined. This assumption does not hold for many real-life applications such as Lifelogger [6], social activity pattern detection, etc. In these systems, the number of activities is not constant. Moreover, different users have their own definition of a “meaningful activity”. It is infeasible to foresee activities that users may be interested in. So in the training phase, the systems are required to learn the most useful sensor modality combination according to different kinds of activity classes. We call these systems Adaptive Activity Recognition Systems In order to recognize personal, unseen activities, some incremental methods [22], [1] are proposed. However, their results are similar to those of non-personalized models [12], indicating that the feature selection is crucial for activity recog- nition [12]. The semantic attribute sequence based models are also used for recognizing unseen new activities [5], [4], but still fail to consider the influence of different features. We have developed a dynamic heterogeneous sensor fusion framework for adaptive activity recognition. The key idea is to find the most discrimnative combination of sensor modalities (motion, sound, location, time of the day, WiFi environment, etc.) for each activity. For example, if all sensor modalities are leveraged, the system will not be able to recognize that the user is walking unless he walks with the same motion, at the same location and at the same time as the training walking examples. On the other hand, when the user annotates new types of activities, the system needs to adjust the model to use additional sensor modalities in order to discriminate a new activity from existing activities. Specifically, when the user labels an activity as walking, the system learns that motion feature is sufficient to recognize this activity. Several days later if the user labels a new type of activity: grocery shopping, which has very similar motion as walking, the system will need to incorporate location information to distinguish these two types of activities. Then the “motion” and “location” sensors play important roles in this case. The sensor weight is a value representing the importance of a sensor. To examine the effectiveness of the proposed framework, we integrate several feature transformation meth- ods including Linear Discriminant analysis (LDA), Marginal Fisher’s Analysis (MFA) and Maximum Mutual Information (MMI) algorithm. To summarize, we develop a practical dynamic heteroge- neous sensor fusion framework, which addresses the challenge of dynamic sensor fusion in adaptive activity recognition. The key contributions of the paper are highlighted as follows: We propose a sensor fusion framework to learn sensor weights for each activity class so that activities are eas- ier to be discriminated in the new distance space. We implement several feature transformation algorithms MobiCASE 2014, November 06-07, Austin, United States Copyright © 2014 ICST DOI 10.4108/icst.mobicase.2014.257787