International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 09 Issue: 01 | Jan 2022 www.irjet.net p-ISSN: 2395-0072
© 2022, IRJET | Impact Factor value: 7.529 | ISO 9001:2008 Certified Journal | Page 702
Sign Language Recognition using Machine Learning
Ms. Ruchika Gaidhani
1
, Ms. Payal Pagariya
2
, Ms. Aashlesha Patil
3
, Ms. Tejaswini Phad
4
5
Mr. Dhiraj Birari, Dept. of Information Technology, MVPs KBT College of Engineering, Maharashtra, India
---------------------------------------------------------------------***----------------------------------------------------------------------
Abstract - Our goal is to develop a model that can detect hand
movements and signs. We'll train a simple gesture detecting
model for sign language conversion, which will allow people to
communicate with persons who are deaf and mentally
challenged. This project can be performed using a variety of
methods, including KNN, Logistic Regression, Nave Bayes
Classification, Support Vector Machine, and CNN. The method
we have chosen is CNN because it has a higher level of
accuracy than other methods. A computer program written in
the programming language Python is used for model training
based on the CNN system. By comparing the input with a pre-
existing dataset created using Indian sign language, the
algorithm will be able to understand hand gestures. Users will
be able to recognize the signs offered by converting Sign
Language into text as an output. by a sign language
interpreter This approach is implemented in Jupyter Lab,
which is an add-on to the Anaconda documentation platform.
To improve even more, we'll convert the inputs to black and
white and accept input from the camera after applying the
Background subtraction approach. Because the mask is
configured to identify human skin, this model doesn't need a
simple background to work and can be constructed with just a
camera and a computer.
Key Words: ISL, Data set, Convolutional Neural Network,
Accuracy, Haar cascade.
1.INTRODUCTION
In today's world, we have computers that operate at
extremely high rates, performing massive amounts of
calculations in a fraction of a second. We humans are now
attempting to accomplish a goal in which the computer will
begin to think and work like a human person. This
necessitates the most fundamental quality, 'learning.' This
brings us to the topic of artificial intelligence. AI states that a
computer may start or begin executing activities on its own,
without the need for human interaction. To achieve this, a
computer must first learn how to react to inputs and
variables. The computer should be taught with a large
amount of data, the data required to train the computer is
determined by the desired output and the machine's
operation. We create a computer model that can identify
human hand motions; there are numerous apps that function
with hand gestures that we observe in our daily lives. Look
at the console in our living room; connect it to a sensor, and
we'll be able to play tennis with our hands. We have created
a touch detection model that translates sign language into
speech. There are a number of devices that rely on touch
detection, whether for security or entertainment. Sign
language is a vision-based language that uses a combination
of material, body language, and gestures, fingers, and
orientation, posture, and hand and body movements, as well
as eyes, lips, and wholeness. facial expressions and speech. A
variety of signature circuits exist, just as there are regional
variations in spoken language. Through gestures and
gestures, facial expressions, and lip movements, we can spell
the letters of each word with our fingers and maintain a
certain vocabulary of Indian Sign Language (ISL), American
Sign Language (ASL), and Portuguese Signature (PSL). Sign
language can be separated or used continuously. People
interact with a single sign language by performing a single
word movement, while continuous sign language can be a
series of steps that form a coherent sentence. All methods of
identifying hand movements are often classified as based on
vision and based on measurements taken by sensors
embedded in gloves. This vision-based process uses human
computer interactions to detect touch with bare hands.
OpenCV is used to detect sign languages in this project. Uses
a webcam to detect user touch; Our model can distinguish
handmade touches with bare hands, so we will not need
gloves for this project. OpenCV is used to detect sign
languages in this project. Uses a webcam to detect the
movement of a user's hand, with words displayed on the
screen as output. Our project aims to help people who do not
know sign language well by identifying and converting man-
made symbols into legible characters. Machine learning,
especially CNN, can help us achieve this. We want to use an
image from the web camera/phone camera to train a model
that can predict text (using IP camera software and OpenCV).
Because the model was trained using a well-known dataset
(ASL), there will be enough data for the training algorithm to
produce a precise and accurate model. Because the model's
code is written in Python, the project can be completed on a
simple computer without the use of high-end processing
units or GPUs. The software Jypter Lab, where this model is
trained and implemented, is built on the Anaconda
Documentation platform. The many concepts involved, as
well as the technique for carrying out this project, are
addressed in detail in the following sections of the paper.
1.1 Sign Language Recognition
Without this notion, our entire model would be
impossible to materialize. Deep neural networks are one of
the classes of deep neural networks, and CNN is one of them.
CNN is used in a variety of fields, the majority of which
include visual images. A CNN is made up of many neurons
whose values, or weights, can be changed to achieve the
desired result. These biases and weights can be learned.
Non-linearity is caused by the dot products made between
the neurons. The key distinction between regular neural
networks and convolutional neural networks is CNN's