This work is licensed under a Creative Commons Attribution 4.0 International License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. ech T Press Science Computers, Materials & Continua DOI:10.32604/cmc.2022.022857 Article A Real-Time Oral Cavity Gesture Based Words Synthesizer Using Sensors Palli Padmini 1 , C. Paramasivam 1 , G. Jyothish Lal 2 , Sadeen Alharbi 3, * and Kaustav Bhowmick 4 1 Department of Electronics & Communication Engineering, Amrita School of Engineering, Bengaluru, Amrita Vishwa Vidyapeetham, India 2 Center for Computational Engineering and Networking (CEN), Amrita School of Engineering, Coimbatore, Amrita Vishwa Vidyapeetham, India 3 Department of Software Engineering, College of Computer and Information Sciences, King Saud University, Riyadh, Saudi Arabia 4 Department of Electronics and Communication Engineering, PES University, Bengaluru, India *Corresponding Author: Sadeen Alharbi. Email: sadalharbi@ksu.edu.sa Received: 20 August 2021; Accepted: 21 October 2021 Abstract: The present system experimentally demonstrates a synthesis of syllables and words from tongue manoeuvers in multiple languages, captured by four oral sensors only. For an experimental demonstration of the system used in the oral cavity, a prototype tooth model was used. Based on the principle developed in a previous publication by the author(s), the proposed system has been implemented using the oral cavity (tongue, teeth, and lips) features alone, without the glottis and the larynx. The positions of the sensors in the proposed system were optimized based on articulatory (oral cavity) gestures estimated by simulating the mechanism of human speech. The system has been tested for all English alphabets and several words with sensor-based input along with an experimental demonstration of the developed algorithm, with limit switches, potentiometer, and flex sensors emulating the tongue in an artificial oral cavity. The system produces the sounds of vowels, consonants, and words in English, along with the pronunciation of meanings of their translations in four major Indian languages, all from oral cavity mapping. The experimental setup also caters to gender mapping of voice. The sound produced from the hardware has been validated by a perceptual test to verify the gender and word of the speech sample by listeners, with 98% and 95% accuracy, respectively. Such a model may be useful to interpret speech for those who are speech-disabled because of accidents, neuron disorder, spinal cord injury, or larynx disorder. Keywords: English vowels and consonants; oral cavity; proposed system; sensors; speech-disabled; speech production; vocal tract model 1 Introduction Communication is essential in modern society, in every environment or social aspect of people’s life, private or public. Statistics show that there are 400 million disabled people in the developing