i*Chameleon: A Unified Web Service Framework for Integrating Multimodal Interaction Devices Kenneth W.K. Lo, Wai Wa Tang, Hong Va Leong Department of Computing The Hong Kong Polytechnic University Hong Kong e-mail: {cskenneth,cswwtang, cshleong}@comp.polyu.edu.hk Alvin Chan , Stephen Chan, Grace Ngai Department of Computing The Hong Kong Polytechnic University Hong Kong e-mail: {csschan,cstschan,csgngai}@comp.polyu.edu.hk Abstract— Multimodality inputs are becoming increasingly popular in supporting pervasive applications, due to the demand for highly responsive and intuitive human control interfaces beyond the traditional keyboard and mouse. However, the heterogeneous nature of novel multimodal input devices and the tight coupling between input devices and applications complicate their deployment, rendering their dynamic integration to the intended applications rather difficult. i*Chameleon exploits device abstraction in a web services-based framework to alleviate these problems. Developers can dynamically register new devices with the i*Chameleon framework. They can also map specific device inputs to keyboard and mouse events efficiently. A number of input modalities such as tangible devices, speech, and finger gestures have been implemented to validate the feasibility of the i*Chameleon framework in supporting multimodal input for pervasive applications. Keywords - component; multimodal; human computer interaction; middleware; web services; I. INTRODUCTION Mark Weiser coined the term pervasive computing, the third wave in computing describing the creation of environments with communication capability, yet gracefully integrated with human users [1]. Riding on the success of mature distributed systems and well-developed hardware technologies, nowadays users consider usability and functionality to be equally important. The proliferation of smart and human-natural devices, such as motion capture cameras and speech recognition engines, renders the integration of different modalities into one single application a concrete reality. Users prefer a combination of different modalities rather than just a single one when interacting with computer systems. With multimodality, reliability of human computer interaction was found to improve [2]. The term multimodal” has been used in many contexts across several research areas [3, 4]. The purpose of multimodal human computer interaction is to study how computer technology can be made more user-friendly, through the interplay among three aspects: user, system and interaction [5]. Exploiting a diversity of networking technologies and multimodal interface techniques, cooperative interaction on pervasive computing system has become a new trend for human computer interaction. Many well-known multimodal systems like ICARE [6] and OpenInterface [7] were developed. OpenInterface is a component-based tool for rapidly deploying multimodal input for different kinds of applications. The framework includes a kernel, a builder and a design tool. After coding their own multimodal application and device drivers, developers will port and deploy them on the platform. However, it does not support plug-and-play and cannot readily support collaborative interaction over the Internet. Promising and intuitive interaction methods with heterogeneous input devices and interaction over the web browser have been developed by W3C. However, there remains much room for the development of multimodal interaction for pervasive systems. A unified multimodal framework is largely needed. This paper presents the i*Chameleon framework, which allows users to customize their interaction, capable of adapting to changes in the environment as well as user requirements according to the features of applications or the needs of different users, like a chameleon. Since this type of dynamic adaptation tremendously increases the complexity in the interaction design, a middleware which can mitigate the tight coupling of applications and specific input devices offers a practical solution. It enables the abstraction of hardware devices and communication technologies, and provides a well-defined interface to interact with other applications or widgets [8]. i*Chameleon focuses on the design considerations of multimodality for pervasive computing. It aims to develop a web-service framework with a separate analytic co-processor for collaborative multimodal interaction, providing a standard and semantic interface that facilitates the integration of new widgets with a wide variety of computer applications. This paper is organized as follows: Section II highlights the desirable system features and requirements. Section III presents the framework workflow and architecture, followed by system implementation and evaluation in Section IV. Brief concluding remarks are given in Section V. II. SYSTEM REQUIREMENTS The system design for supporting non-traditional interfaces such as a multimodal interaction framework is much more complex than those for traditional interfaces. Design efforts tend to be focused on implementing particular interface devices to fulfill specific requirements. Hence, the product interface is usually not flexible enough to be reused on new applications and new widgets or to 9th IEEE International Workshop on Managing Ubiquitous Communications and Services 2012, Lugano (19 March 2012) 978-1-4673-0907-3/12/$31.00 ©2012 IEEE 106