CiiT International Journal of Digital Image Processing, Vol 7, No 08, August 2015 234 0974-9691/CIITIJ-6055/08/$20/$100 © 2015 CiiT Published by the Coimbatore Institute of Information Technology Abstract---Lip password is composed of a password embedded with motions of lip and point out the characteristic of lip motion. To provides security of a speaker verification system by using private password and behavioral biometrics of a lip motion simultaneously. The target speaker saying wrong password then rejected and the target speaker saying correct password then detected. Here a Hidden Markov Model (HMM) learning approach based on multi boosted scheme is presented for a security speaker system. This method first extract the visual features and to characterize each frame. The lip password segmentation algorithm is used for the segmentation of lip sequences. Hidden Markov Models with boosting learning framework contains random subspace method and data sharing scheme. Finally, the lip-password is verified based on verification results provided by all the subunit learned from HMM based multi-boosted scheme and it will check whether the password is spoken by the speaker with the already-recorded password or not. Keywords---Lip Motion, HMM, GMM, RSM, DSS I. INTRODUCTION OWDAYS in the community considerable attention must needed for speaker verification because of its several applications such as secure access control, financial transaction authentication, human-computer interfaces, security protection [1], [3]. This verification process is done by following. Using previously stored information first it verifies a claimed speaker under a certain criterion; thereby the speaker would be either accepted or rejected based on verification [1]. Traditionally, not only the linguistic information is conveyed by speech also it characterizes the speaker identity that can be utilized for speaker verification. To achieve speaker verification the most natural modality may probably be acoustic speech signals. Speaker verification system based on acoustic model is effective in domain of application, but the performance of the purely acoustic SPV would be degraded dramatically because of the environment corrupted by the multiple talkers or background noise. For instance, Kenny et al. derives the spatial and temporal models Manuscript received on July 28, 2015, review completed on July 28, 2015 and revised on February 08, 2015. A.Jebaselvi is with the Department of Information Technolgy in Manonmaniam Sundaranar University, Tirunelveli, TamilNadu, India. E-Mail: jebaselvia21@gmail.com. Dr. Kumar Parasuraman is with the Centre for Information Technology and Engineering, Manonmaniam Sundaranar University, Tirunelveli, Tamilnadu, India. E-Mail: kumarcite@msuniv.ac.in / kumarcite@gmail.com Dr. T. Arumuga Maria Devi is with the Centre for Information Technology and Engineering, Manonmaniam Sundaranar University, Tirunelveli, Tamilnadu, India. E-Mail: deviececit@gmail.com. Digital Object Identifier: DIP082015003. for the identification of speaker by extracting the group of visual lip region features and then utilizes the Hidden Markov Model with Gaussians Mixtures [10],[15],[16], while Sao et al. Gaussian Mixture Model are utilized to build the statistical speaker models. Later, Shafait et al. utilized the GMM classifier and extracted a suitable visual features group from the sequential mouth regions. Faraj et al. Utilizes GMM and lip motion group features are extracted with person verification. The lip motion alone concentrated on this paper, although the other modalities can also be fused with this. The password is attached with the movement of lip and the basic characteristic of motion of lip sequences is composed to derive a concept of lip motion password. This approach provides a double security lip-password protected speaker verification system [9]. Simultaneously, the password information and behavioral biometrics of lip motions are verified by the speaker. Whether the wrong password said by the target speaker or the accurate password known by wrong person will be discovered and refused accordingly. The speaker verification system is efficient means it has some advantages and they are as follows. Motions of lip sequences are not sensitive to the environmental noise; the lip motion acquisition is not affected by distance; this type of speaker verification system can be carried out in a hidden way silently; simply applicable to a weakened speech person. In literature, almost all related speaker verification systems analyze the whole utterance as the basic unit of processing for the lip motions. The lip password protected system is designed based on the capability that it simultaneously detects both of the following two cases. That is the speaker saying incorrect password and wrong person saying the accurate correct password. Such tasks are incompetent in these traditional methods. Generally, the lip-password protection system has always comprises of many sub block units. A short period of lip motions are indicated by these subunits and to describe the underlying lip-password information these subunits always have diverse moving between various elements that are considered individually, but not as a whole. Here, we present a multi-boosted Hidden Markov Model learning approach to such a speaker verification system based on lip-password. A group of representative visual features are extracted first to characterize each lip frame and then an effective algorithm is used to segment the lip-password sequence. Then these sequences of lip-password are splitted into a small set of sub block units. For subunit verification we integrate Hidden Markov Models with multi boosted learning Security Based Speaker Verification for Lip-Password using Learning Multi-Boosted HMMS A. Jebaselvi, Kumar Parasuraman and T. Arumuga Maria Devi N