SlideShare a Scribd company logo
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 09 Issue: 01 | Jan 2022 www.irjet.net p-ISSN: 2395-0072
© 2022, IRJET | Impact Factor value: 7.529 | ISO 9001:2008 Certified Journal | Page 1643
SIGN LANGUAGE RECOGNITION USING MACHINE LEARNING
Sneha Prabhu1, Sriraksha Shetty2, Sushmitha P Suvarna3, Vindya Sanil4, Dr. Jagadisha N5
1,2,3,4Student, Dept. of Information Science and Engineering, CEC, Karnataka
5HOD & Associate Professor, Dept. of Information Science and Engineering, CEC, Karnataka
---------------------------------------------------------------------***----------------------------------------------------------------------
Abstract - Communication must happen in both ways.
Communicating with the people having hearing disability is a
difficult task so we developed the real time sign language
recognition system using machine learning. This system
contains the hand gestures performed by people which is
captured using the web camera. Out of allimagestheregionof
interest is identified and the system will predict and show the
name of the sign captured. The captured signal will undergo
some of the processing steps inclusive of data acquisition,
image processing, characteristic extraction and classification.
We will use the Convolutional Neural Network (CNN) to train
and build our model. Hence it is very useful for people with
hearing and speech disability.
Key Words: CNN, Sign Language, Machine Learning.
1.INTRODUCTION
Sign language is a language for the deaf and dumb which
uses simultaneous orientation and movement of hand
shapes instead of acousticallyconveyedsoundpatterns. Deaf
and Dumb human beings depend on sign language
translators for communications. However, finding
experienced and qualified translators for their day to day
affairs throughout life period is a very difficult task and also
overpriced. Sign language is the basic means of
communication for those with hearing and vocal disabilities.
Those disadvantaged have difficulty in their day to daylives.
In other scenario large group of people who had no problem
in communicating but are unable to speak because of injury
like accidents or disease are not able to speak. Sign language
translates to text or speech, would give them everyday
communication. In regional and rural areas, it is very
effortful for the deaf and dumb people to communicate to
normal people so to bridge the gap between them a
communication tool needs to be provided. The purpose is to
increase the ability of deaf and dumb people to acquire
services wherever they want without taking the help of any
sign language experts. This systems can reduce the urge for
professional translating services and can actually makesthe
life better. This can also be used for training sign language
and for people who are dumb and deaf due to sickness or
injury because of accidents. We aim todevelopa systemthat
would ease this difficulty in communication. Sign language
consists of making shapes or movements with your hands
with respect to the head or other body parts along with
certain facial cues. A recognition system would thus have to
identify specifically the head and hand orientation or
movements, facial expression and even body pose. We
propose the design for a basic yet extensible system that is
able to recognize Sign Language. American Sign Language
was chosen since it is utilized by a majorityofthosedisabled.
We believe that it will benefit for deaf and dumb people by
providing them a flexible translator. Sign Language
Recognition System (SLR) utilize computer vision and
machine learning method convolutional neural network
(CNN).
Fig 1: Sign Language for Alphabets A-Z
2. RELATED WORK
In[1],Citra Suradi et al., This paper implements a design of
sign language recognition using E-CNN. In this study SIBI
alphabetical images were used datasets. The total datasets
used is 38656 data which are then divided into three parts of
data, namely training data, validation data and testing data.
Handkey point was used forfilterandcoppingprocessbefore
the use of dataset. Since Handkey point will detect the hand
joints from the image and it will crop and filter image data so
that the dataset used focuses on hand object. In this study,
there were 5 CNN (Convolution Neural Network) model
combined with 3 Ensemble model were used. Also observed
that combining several algorithms can increase the accuracy
hence ensemble method was used along with CNN model.
In[2],Lionel Pigou et al., This paper signlanguagerecognition
system using CNN. For building the automated recognition
system two main steps are necessary. The first step is
extracting the features from the framesequence. The second
step is actions will be classified. In this they have used CNN
for extraction of frame sequence from the image and ANN is
used for the classification. Datasets had been used from the
Chalearn Looking at human beings. Data units consists of 20
Italian gestures achieved from the 27 users with variant in
surrounding, garb, mild.. There was 6600 gestures in
development set used, out of which 4600 fortraining set and
2000 for validation set. The methodology used are dataset
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 09 Issue: 01 | Jan 2022 www.irjet.net p-ISSN: 2395-0072
© 2022, IRJET | Impact Factor value: 7.529 | ISO 9001:2008 Certified Journal | Page 1644
collection, pre-processing, CNN, proposed architecture,
generalization and training, temporal segmentation and
results. In proposed architecturethereweretwofeaturesone
for extracting hand gestures and other one for upper body
features extraction.
In[3],Adithya V et al., In this paper a method for automation
recognition of fingerspelling in Indian sign language was
proposed. Digital Image processing strategies and artificial
neural community for recognizing special symptoms were
used to enforce the gadget.. This paper presents amethodfor
both English alphabets as well as numbers. This technique
had included 26 letters of English alphabets and zero-9
numbers.This approach had four foremost steps namely
Image Acquistion and pre-processing, hand gestures,
characteristic extraction, classification. Images have been
captured with black heritage.Data units were created for
education and checking out the device. Next extracting the
hand signs and symptoms from the captured photograph
usingskin colour-basedtotallysegmentationandproducethe
binary image with signs or skin pixels with white colour and
background become set to black colour. The form function
was derived from the space remodel the binary
picture.Euclidean,Cityblockandchessboardusedforlocating
the distance transform of an photograph. ANN turned into
used for classification. The feature vector acquired from
inside the feature extraction become used as enter for the
classifier to apprehend the Sign.
In[4],Muthu Mariappan H et al., This system is implemented
for recognising the gestures of Indian Sign Language (ISL).
System uses non-invasive vision-based recognition method.
Using skin segmentation feature od OpenCV the region of
interest(ROI) are identified and tracked. By applying the
Fuzzy C-meansClustering(FCM) machinelearningalgorithm
the training and prediction of hand gestures are performed.
The data samples of 80 words and 50 sentences of daily
usage terms of Indian sign language were collected and 10
volunteers videos were recorded therefore there was
80X10=800 total samples of words and 50X10=500 total
samples of sentences. The raw videos were taken in dynamic
background and given to the system as the input. Duringpre-
processing noise eliminatedand blurred imagewasobtained
by performingconvolutionoperationwithlow-passboxfilter.
BGR image is converted into HSV color space and
morphological operations are performed on HSV image and
gives the binary image. Areaofallcontourwascalculatedand
using them three were selected and the orientation between
the contours is also kept track. Those three contour are face
,right and left hand which contributes gestures. For each
frame in video required features are extracted from these
regions as vector feature. During training, extracted features
are given to the c-mean algorithm and during testing, it
matches the text file with existing cluster.
In[5],Soma Shrenika et al., This system implemented to
overcome the barriers faced by the impaired hearing people
using sign languages. ASL (American sign language) dataset
was used to identify the sign made by gesture. The dataset
was consist of numbers zero - nine and alphabets a - z. There
were 70 samples for each 36 symbols. This system uses
camera as there source which captures various gestures of
hand. The captured image changed intoinRGBformatandby
using the common approach the RGB image turned into
converted into grayscale photo. Then pr-processing of the
image takes region wherein edges are detected the usage of
Canny facet detection set of rules. Canny area detection
algorithm had 5 extraordinary stages specifically, Image
smoothing,GradientMagnitude,Non-maximumSuppression,
Double threshold , Hysteresis .Finally template matching
algorithm turned into used to become aware of the sign and
show the textual content. Sum of Absolute Difference (SAD)
method was used in comparison of template image with the
images present in the data set.
In[6],Divya Deroa et al., In this paper simultaneously
combining hand shapes, orientation andmovementofhands,
fingers or frame,and facial features to fluidly specific a
speaker ’ s thought.. Here data is acquired through a 3 Mega
pixel cameraand person wears red and blue colouredgloves.
system is made only to recognise alphabets and numbers
Therefore total number of images in the database is 510
which include 25 alphabets and 9 numbers. The steps
included in this experiment are (1) segmentation, (2)
fingertip finding algorithm and (3) PCA. Recognition is done
using PCA for live data frames where every 20th frame is
given as input to be recognized. The system can recognise
94% of the signs made correctly. Fingertip algorithm is used
to make database on the basis of number of finger tips. PCA
compares the input image with the new database formed
with the help of finger tips.
In[7],G. Anantha Rao et al., This paper proposes the
recognition of Indiansign language using CNN. Thecaptured
method usedin this paper was Selfie mode continuous sign
language video. Datasets were created with 5 different
subjects performing 200 signs in 5 different viewing angles
under various circumstances. Out of which 3 different
samples size was used forperforming CNN training whereas
remaining 2 samples were used for testing the trained CNN.
This model was constructedwithinputlayerwheredata’sare
given as input ,in feature extraction stage 4 convolution
neural network of different window size, five rectified linear
unit for non-linearities , 2 stochastic pooling layer to avoid a
substantial information loss, in classificationstageonedense
and one fully connected layers. In classification Softmax
regression was adopted.
In[8],Surejya Suresh et al., In this paper the proposed
architecture is analysed using two models Stochastic
Gradient Descent optimizer and Adam optimizer. The
algorithm used is Convolution Neuralnetwork.CNNcontains
several layers such as pooling, dropout and fully connected
layer. All the important features are extracted from the
bottom layer and abstract features are computed from high
level layers of CNN. The proposed modelarchitecturehasthe
input layer which receives the images of size 64x64 pixels.
Later two convolutionoperationsareappliedtotheimagesto
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 09 Issue: 01 | Jan 2022 www.irjet.net p-ISSN: 2395-0072
© 2022, IRJET | Impact Factor value: 7.529 | ISO 9001:2008 Certified Journal | Page 1645
generatefeaturemaps. The featuremaps are flattenedwhich
results in feature vector. Python is the programming
language used. Some of the libraries used are keras, scikit-
learn, matplotlib and numpy.
In[9],Kusumika Krori Dutta et al.,In this paper Indian Sign
Language is communicated using hand gestures made both
Single hand and Double hands . In this, only main words are
taken into consideration grammar usage is not considered.
Articles such as ‘a’, ‘an’, ‘the’ are omitted and also doesn’t
include tense form .This paper deals with classification of
Indian sign language using machine learning. The system is
trained with double handed sign language by using a
Principal Component Analysis(PCA) and Artificial neural
network (ANN) algorithm in MATLAB. Thehandgesturesare
acquired from different sized and complexion hands and
made an image database of having ease image of size 768 by
1024. Once the images were preprocessed the size of the
image was reduced to 256 by256.Signlanguagepatternsare
trained separately using two machinelearningtechniques.A.
K Nearest Neighbours Algorithm B. Back Propogation
Algorithm. This experiment was conducted over 220 images
of double handed Indian sign language alphabets and 800
images of single handed Indian sign language alphabets.
EnglishlettersandnumbersusingKNNandBackpropagation
techniques where PCA is used for dimension reductionality
.For K-NN techniques with K=1, achieved 100% recognition
rate whereas using back propagation techniqueachieved94-
96.
In[10],LihongZheng et al., Inthis paperAuslanistheprimary
language used by the Australian Deaf Community, with an
estimated 30,000 usersinAustraliaSignlanguagetranslation
to text or speech would give them everyday communication.
The fashionable framework of imaginative and prescient
based SLR includes pre-processing, sign gesture illustration,
function extraction and type. The widely used 2D features
include SIFT, HOG, HOF, STIP,and kerneldescriptors. Forthe
purpose of dealing with occlusion 3D/4D space of time
(HON4D) and Random Occupancy Pattern(ROP) features
were proposed. Sign gesture recognition system were
divided into two types based on image acquisition: touched
based or untouched/vision based There are about seven
main large datasets of sign language.TheyareAmericanSign
LanguageLexiconVideoDatasetMSRGesture3D,Auslandata
set, LTI-Gesture Database, RWTH German Fingerspelling
Database, DEVISIGN Chinese Sign Language dataset, and
Indian Sign Language dataset.
3. PROPOSED SYSTEM
The system uses the web camera for capturing the hand
gestures of the hearing and speechimpairedpeople. Theraw
images taken is given as an input to the system. The image
frames are resized to maintain the equality among all the
images. The proposed system includes several stepsnamely,
Image Acquisition, Image Pre-Processing, Segmentation
,Feature Extraction, Recognition and Text Output.
Fig 2: Schematic Diagram of Sign Language Recognition
3.1 Image Acquisition:
This is the first step in the workflow as no processing can be
done without getting an image. Image Acquisition is the
process of extracting an image from a source, here the
source is the hardware-based source that is thewebcamera.
This image is not processed in any way so itistherawimage.
3.2 Segmentation:
The method of separating objects or signs from the context
of a captured image is known as segmentation. This process
includes Con-text subtracting, skin-colordetection,andedge
detection. The location of the hand must be detected and
separated in order to recognize the hand gestures.
3.3 Features Extraction:
Predefined features such as form, contour, geometrical
feature like position, angle, distance, etc, colour feature,
histogram, and others are extracted from the pre-processed
images. Later it is used for sign language recognition.
Feature extraction is a stepthatdividesand organizes a large
collection of raw data. It reduces to smaller and easier-to-
manage size so that processing would be simpler. There is
massive amount of data sets that have a lot of variable
features in it. To process these variable features, a large
amount of computation is needed. By Reducing the size of
the data these features are simple to use.
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 09 Issue: 01 | Jan 2022 www.irjet.net p-ISSN: 2395-0072
© 2022, IRJET | Impact Factor value: 7.529 | ISO 9001:2008 Certified Journal | Page 1646
3.4 Pre-Processing:
Pre-processing is the process of eliminating high-intensity
noises from the image frames. The first step is smoothing or
blurring. Then the image is converted to gray scale from
RGB. The main aim of converting the image into grey scale is
to reduce the size and amount of processing.
The different phases of pre-processing are:
3.4.1 Morphological Transform
It is a simple operation based on the shape of the image
which are performed on binary images. There are forms of
morphological transforms specifically dilation and erosion.
3.4.2 Blurring
Blurring is the process of eliminating the noice in an image
and keeping required feature intact.
3.4.3 Thresholding
The process of convertinggreyscaleimageintobinaryimage
is called Thresholding. Thresholding is used to pick thearea
of interest ignoring the area that is not required.
3.4.4 Recognition
Classifiers are used in this section Classifiers are the
algorithm used to interpret the alerts. The classifier used tin
this challenge is Convolution Neural Network. Convolution
neural network is a deep mastering algorithm which takes
the photograph as input, images the critical aspects fromthe
photo and differentiates from one another. CNN algorithm
offers better precision compared toother algorithminphoto
type and reputation.
3.4.5 Text output
The hand gestures are recognized and the system gives an
output in the human understandable form that is in text.
4. CONCLUSION
The Sign Language Recognition (SLR) systemisa method for
recognizing a collectionofformedsignsandtranslatingthem
into text with the appropriate context. The significance of
gesture recognition can be seen in the development of
effective human-machine interactions. We attempted to
build a model using a Convolutional Neural Network in this
project.
5. REFERENCES
[1] Citra Suradi , Anik Nur Handayani, Rosa Andrie Asmara,
Aji Prasetya Wibawa ,Lilis Nur Hayati,Huzain Azis(2021).
“Design Of Sign Language Recognition Using E-CNN”.
[2] Lionel Pigou; Sander Dieleman; Pieter-Jan Kindermans;
Benjamin Schrauwen(2014). “Sign Language Recognition
Using Convolution Neural Network”.
[3] Adithya V, Vinod P.R, Usha Gopalakrishnan(2013).
“Artificial Neural Network Based Method For Indian Sign
Language Recognition”.
[4] Muthu Mariappan H, Dr. Gomathi V (2019). “Real Time
Recognition of Indian Sign Language”.
[5] Soma Shrenika , Myneni Madhu Bala(2020). “Sign
Language Recognition UsingTemplateMatchingTechnique”.
[6] Divya Deroa, Nikesh Bajaj(2012). “Indian Sign Language
recognition”.
[7] G. Anantha Rao,K Syamala, P V V Kishore, A S C S Sastry
(2018).“DeepConvolutionNeuralNetworkforSignLanguage
Recognition”.
[8] Surejya Suresh,Mithun Haridas T P, Supriya M H(2019).
“Sign Language Recognition using Deep Neural Network”.
[9] Kusumika Krori Dutta, Sunny Arokia Swamy
Bellary(2017). “MachineLearningTechniquesforIndianSign
Language Recognition”.
[10] Lihong Zheng, Bin Liang, Ailian Jiang (2017). “ Recent
Advances of Deep Learning for Sign Language Recognition”.
[11] Rachana Patil, Vivek Patil1, Abhishek Bahuguna1, Mr.
Gaurav Datkhile(2021). “Indian Sign Language Recognition
Using Convolution Neural Network”.

More Related Content

PDF
IRJET- Sign Language Interpreter using Image Processing and Machine Learning
PDF
IRJET- Tamil Sign Language Recognition Using Machine Learning to Aid Deaf and...
PDF
IRJET- Vision Based Sign Language by using Matlab
PDF
Ay4102371374
PDF
SIGN LANGUAGE INTERFACE SYSTEM FOR HEARING IMPAIRED PEOPLE
PDF
Real Time Sign Language Detection
PDF
IRJET - Sign Language Text to Speech Converter using Image Processing and...
PDF
Gesture Acquisition and Recognition of Sign Language
IRJET- Sign Language Interpreter using Image Processing and Machine Learning
IRJET- Tamil Sign Language Recognition Using Machine Learning to Aid Deaf and...
IRJET- Vision Based Sign Language by using Matlab
Ay4102371374
SIGN LANGUAGE INTERFACE SYSTEM FOR HEARING IMPAIRED PEOPLE
Real Time Sign Language Detection
IRJET - Sign Language Text to Speech Converter using Image Processing and...
Gesture Acquisition and Recognition of Sign Language

Similar to SIGN LANGUAGE RECOGNITION USING MACHINE LEARNING (20)

PDF
GRS '“ Gesture based Recognition System for Indian Sign Language Recognition ...
PDF
Sign Language Detection using Action Recognition
PDF
Design of a Communication System using Sign Language aid for Differently Able...
PDF
KANNADA SIGN LANGUAGE RECOGNITION USINGMACHINE LEARNING
PDF
SignReco: Sign Language Translator
PDF
IRJET- ASL Language Translation using ML
PDF
IRJET- Gesture Recognition for Indian Sign Language using HOG and SVM
PDF
Sign Language Identification based on Hand Gestures
PDF
Sign Language Detector Using Cloud
PDF
Design and Development of Motion Based Continuous Sign Language Detection
PDF
INDIAN SIGN LANGUAGE TRANSLATION FOR HARD-OF-HEARING AND HARD-OF-SPEAKING COM...
PDF
IRJET- Survey on Sign Language and Gesture Recognition System
PDF
Translation of sign language using generic fourier descriptor and nearest nei...
PDF
IRJET - Mutecom using Tensorflow-Keras Model
PDF
IRJET- Hand Gesture Recognition System using Convolutional Neural Networks
PDF
IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...
PDF
Live Sign Language Translation: A Survey
PPTX
finalyear_projecGHHHHHHHHHHHHHHHDYTDYTRTRTD
PDF
Sign Language Recognition using Facial Gesture and Skeleton Keypoints
PDF
Real Time Sign Language Detection
GRS '“ Gesture based Recognition System for Indian Sign Language Recognition ...
Sign Language Detection using Action Recognition
Design of a Communication System using Sign Language aid for Differently Able...
KANNADA SIGN LANGUAGE RECOGNITION USINGMACHINE LEARNING
SignReco: Sign Language Translator
IRJET- ASL Language Translation using ML
IRJET- Gesture Recognition for Indian Sign Language using HOG and SVM
Sign Language Identification based on Hand Gestures
Sign Language Detector Using Cloud
Design and Development of Motion Based Continuous Sign Language Detection
INDIAN SIGN LANGUAGE TRANSLATION FOR HARD-OF-HEARING AND HARD-OF-SPEAKING COM...
IRJET- Survey on Sign Language and Gesture Recognition System
Translation of sign language using generic fourier descriptor and nearest nei...
IRJET - Mutecom using Tensorflow-Keras Model
IRJET- Hand Gesture Recognition System using Convolutional Neural Networks
IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...
Live Sign Language Translation: A Survey
finalyear_projecGHHHHHHHHHHHHHHHDYTDYTRTRTD
Sign Language Recognition using Facial Gesture and Skeleton Keypoints
Real Time Sign Language Detection
Ad

More from IRJET Journal (20)

PDF
Enhanced heart disease prediction using SKNDGR ensemble Machine Learning Model
PDF
Utilizing Biomedical Waste for Sustainable Brick Manufacturing: A Novel Appro...
PDF
Kiona – A Smart Society Automation Project
PDF
DESIGN AND DEVELOPMENT OF BATTERY THERMAL MANAGEMENT SYSTEM USING PHASE CHANG...
PDF
Invest in Innovation: Empowering Ideas through Blockchain Based Crowdfunding
PDF
SPACE WATCH YOUR REAL-TIME SPACE INFORMATION HUB
PDF
A Review on Influence of Fluid Viscous Damper on The Behaviour of Multi-store...
PDF
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...
PDF
Explainable AI(XAI) using LIME and Disease Detection in Mango Leaf by Transfe...
PDF
BRAIN TUMOUR DETECTION AND CLASSIFICATION
PDF
The Project Manager as an ambassador of the contract. The case of NEC4 ECC co...
PDF
"Enhanced Heat Transfer Performance in Shell and Tube Heat Exchangers: A CFD ...
PDF
Advancements in CFD Analysis of Shell and Tube Heat Exchangers with Nanofluid...
PDF
Breast Cancer Detection using Computer Vision
PDF
Auto-Charging E-Vehicle with its battery Management.
PDF
Analysis of high energy charge particle in the Heliosphere
PDF
A Novel System for Recommending Agricultural Crops Using Machine Learning App...
PDF
Auto-Charging E-Vehicle with its battery Management.
PDF
Analysis of high energy charge particle in the Heliosphere
PDF
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...
Enhanced heart disease prediction using SKNDGR ensemble Machine Learning Model
Utilizing Biomedical Waste for Sustainable Brick Manufacturing: A Novel Appro...
Kiona – A Smart Society Automation Project
DESIGN AND DEVELOPMENT OF BATTERY THERMAL MANAGEMENT SYSTEM USING PHASE CHANG...
Invest in Innovation: Empowering Ideas through Blockchain Based Crowdfunding
SPACE WATCH YOUR REAL-TIME SPACE INFORMATION HUB
A Review on Influence of Fluid Viscous Damper on The Behaviour of Multi-store...
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...
Explainable AI(XAI) using LIME and Disease Detection in Mango Leaf by Transfe...
BRAIN TUMOUR DETECTION AND CLASSIFICATION
The Project Manager as an ambassador of the contract. The case of NEC4 ECC co...
"Enhanced Heat Transfer Performance in Shell and Tube Heat Exchangers: A CFD ...
Advancements in CFD Analysis of Shell and Tube Heat Exchangers with Nanofluid...
Breast Cancer Detection using Computer Vision
Auto-Charging E-Vehicle with its battery Management.
Analysis of high energy charge particle in the Heliosphere
A Novel System for Recommending Agricultural Crops Using Machine Learning App...
Auto-Charging E-Vehicle with its battery Management.
Analysis of high energy charge particle in the Heliosphere
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...
Ad

Recently uploaded (20)

PPTX
web development for engineering and engineering
PDF
R24 SURVEYING LAB MANUAL for civil enggi
PPTX
MCN 401 KTU-2019-PPE KITS-MODULE 2.pptx
PDF
Operating System & Kernel Study Guide-1 - converted.pdf
PPTX
KTU 2019 -S7-MCN 401 MODULE 2-VINAY.pptx
PPTX
additive manufacturing of ss316l using mig welding
PPTX
CARTOGRAPHY AND GEOINFORMATION VISUALIZATION chapter1 NPTE (2).pptx
PPTX
Recipes for Real Time Voice AI WebRTC, SLMs and Open Source Software.pptx
PDF
Mohammad Mahdi Farshadian CV - Prospective PhD Student 2026
PDF
BMEC211 - INTRODUCTION TO MECHATRONICS-1.pdf
DOCX
ASol_English-Language-Literature-Set-1-27-02-2023-converted.docx
PDF
Automation-in-Manufacturing-Chapter-Introduction.pdf
PDF
July 2025 - Top 10 Read Articles in International Journal of Software Enginee...
PPT
CRASH COURSE IN ALTERNATIVE PLUMBING CLASS
PPT
Mechanical Engineering MATERIALS Selection
PDF
SM_6th-Sem__Cse_Internet-of-Things.pdf IOT
PPTX
Lecture Notes Electrical Wiring System Components
PPTX
UNIT 4 Total Quality Management .pptx
PDF
PPT on Performance Review to get promotions
PPTX
Sustainable Sites - Green Building Construction
web development for engineering and engineering
R24 SURVEYING LAB MANUAL for civil enggi
MCN 401 KTU-2019-PPE KITS-MODULE 2.pptx
Operating System & Kernel Study Guide-1 - converted.pdf
KTU 2019 -S7-MCN 401 MODULE 2-VINAY.pptx
additive manufacturing of ss316l using mig welding
CARTOGRAPHY AND GEOINFORMATION VISUALIZATION chapter1 NPTE (2).pptx
Recipes for Real Time Voice AI WebRTC, SLMs and Open Source Software.pptx
Mohammad Mahdi Farshadian CV - Prospective PhD Student 2026
BMEC211 - INTRODUCTION TO MECHATRONICS-1.pdf
ASol_English-Language-Literature-Set-1-27-02-2023-converted.docx
Automation-in-Manufacturing-Chapter-Introduction.pdf
July 2025 - Top 10 Read Articles in International Journal of Software Enginee...
CRASH COURSE IN ALTERNATIVE PLUMBING CLASS
Mechanical Engineering MATERIALS Selection
SM_6th-Sem__Cse_Internet-of-Things.pdf IOT
Lecture Notes Electrical Wiring System Components
UNIT 4 Total Quality Management .pptx
PPT on Performance Review to get promotions
Sustainable Sites - Green Building Construction

SIGN LANGUAGE RECOGNITION USING MACHINE LEARNING

  • 1. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 09 Issue: 01 | Jan 2022 www.irjet.net p-ISSN: 2395-0072 © 2022, IRJET | Impact Factor value: 7.529 | ISO 9001:2008 Certified Journal | Page 1643 SIGN LANGUAGE RECOGNITION USING MACHINE LEARNING Sneha Prabhu1, Sriraksha Shetty2, Sushmitha P Suvarna3, Vindya Sanil4, Dr. Jagadisha N5 1,2,3,4Student, Dept. of Information Science and Engineering, CEC, Karnataka 5HOD & Associate Professor, Dept. of Information Science and Engineering, CEC, Karnataka ---------------------------------------------------------------------***---------------------------------------------------------------------- Abstract - Communication must happen in both ways. Communicating with the people having hearing disability is a difficult task so we developed the real time sign language recognition system using machine learning. This system contains the hand gestures performed by people which is captured using the web camera. Out of allimagestheregionof interest is identified and the system will predict and show the name of the sign captured. The captured signal will undergo some of the processing steps inclusive of data acquisition, image processing, characteristic extraction and classification. We will use the Convolutional Neural Network (CNN) to train and build our model. Hence it is very useful for people with hearing and speech disability. Key Words: CNN, Sign Language, Machine Learning. 1.INTRODUCTION Sign language is a language for the deaf and dumb which uses simultaneous orientation and movement of hand shapes instead of acousticallyconveyedsoundpatterns. Deaf and Dumb human beings depend on sign language translators for communications. However, finding experienced and qualified translators for their day to day affairs throughout life period is a very difficult task and also overpriced. Sign language is the basic means of communication for those with hearing and vocal disabilities. Those disadvantaged have difficulty in their day to daylives. In other scenario large group of people who had no problem in communicating but are unable to speak because of injury like accidents or disease are not able to speak. Sign language translates to text or speech, would give them everyday communication. In regional and rural areas, it is very effortful for the deaf and dumb people to communicate to normal people so to bridge the gap between them a communication tool needs to be provided. The purpose is to increase the ability of deaf and dumb people to acquire services wherever they want without taking the help of any sign language experts. This systems can reduce the urge for professional translating services and can actually makesthe life better. This can also be used for training sign language and for people who are dumb and deaf due to sickness or injury because of accidents. We aim todevelopa systemthat would ease this difficulty in communication. Sign language consists of making shapes or movements with your hands with respect to the head or other body parts along with certain facial cues. A recognition system would thus have to identify specifically the head and hand orientation or movements, facial expression and even body pose. We propose the design for a basic yet extensible system that is able to recognize Sign Language. American Sign Language was chosen since it is utilized by a majorityofthosedisabled. We believe that it will benefit for deaf and dumb people by providing them a flexible translator. Sign Language Recognition System (SLR) utilize computer vision and machine learning method convolutional neural network (CNN). Fig 1: Sign Language for Alphabets A-Z 2. RELATED WORK In[1],Citra Suradi et al., This paper implements a design of sign language recognition using E-CNN. In this study SIBI alphabetical images were used datasets. The total datasets used is 38656 data which are then divided into three parts of data, namely training data, validation data and testing data. Handkey point was used forfilterandcoppingprocessbefore the use of dataset. Since Handkey point will detect the hand joints from the image and it will crop and filter image data so that the dataset used focuses on hand object. In this study, there were 5 CNN (Convolution Neural Network) model combined with 3 Ensemble model were used. Also observed that combining several algorithms can increase the accuracy hence ensemble method was used along with CNN model. In[2],Lionel Pigou et al., This paper signlanguagerecognition system using CNN. For building the automated recognition system two main steps are necessary. The first step is extracting the features from the framesequence. The second step is actions will be classified. In this they have used CNN for extraction of frame sequence from the image and ANN is used for the classification. Datasets had been used from the Chalearn Looking at human beings. Data units consists of 20 Italian gestures achieved from the 27 users with variant in surrounding, garb, mild.. There was 6600 gestures in development set used, out of which 4600 fortraining set and 2000 for validation set. The methodology used are dataset
  • 2. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 09 Issue: 01 | Jan 2022 www.irjet.net p-ISSN: 2395-0072 © 2022, IRJET | Impact Factor value: 7.529 | ISO 9001:2008 Certified Journal | Page 1644 collection, pre-processing, CNN, proposed architecture, generalization and training, temporal segmentation and results. In proposed architecturethereweretwofeaturesone for extracting hand gestures and other one for upper body features extraction. In[3],Adithya V et al., In this paper a method for automation recognition of fingerspelling in Indian sign language was proposed. Digital Image processing strategies and artificial neural community for recognizing special symptoms were used to enforce the gadget.. This paper presents amethodfor both English alphabets as well as numbers. This technique had included 26 letters of English alphabets and zero-9 numbers.This approach had four foremost steps namely Image Acquistion and pre-processing, hand gestures, characteristic extraction, classification. Images have been captured with black heritage.Data units were created for education and checking out the device. Next extracting the hand signs and symptoms from the captured photograph usingskin colour-basedtotallysegmentationandproducethe binary image with signs or skin pixels with white colour and background become set to black colour. The form function was derived from the space remodel the binary picture.Euclidean,Cityblockandchessboardusedforlocating the distance transform of an photograph. ANN turned into used for classification. The feature vector acquired from inside the feature extraction become used as enter for the classifier to apprehend the Sign. In[4],Muthu Mariappan H et al., This system is implemented for recognising the gestures of Indian Sign Language (ISL). System uses non-invasive vision-based recognition method. Using skin segmentation feature od OpenCV the region of interest(ROI) are identified and tracked. By applying the Fuzzy C-meansClustering(FCM) machinelearningalgorithm the training and prediction of hand gestures are performed. The data samples of 80 words and 50 sentences of daily usage terms of Indian sign language were collected and 10 volunteers videos were recorded therefore there was 80X10=800 total samples of words and 50X10=500 total samples of sentences. The raw videos were taken in dynamic background and given to the system as the input. Duringpre- processing noise eliminatedand blurred imagewasobtained by performingconvolutionoperationwithlow-passboxfilter. BGR image is converted into HSV color space and morphological operations are performed on HSV image and gives the binary image. Areaofallcontourwascalculatedand using them three were selected and the orientation between the contours is also kept track. Those three contour are face ,right and left hand which contributes gestures. For each frame in video required features are extracted from these regions as vector feature. During training, extracted features are given to the c-mean algorithm and during testing, it matches the text file with existing cluster. In[5],Soma Shrenika et al., This system implemented to overcome the barriers faced by the impaired hearing people using sign languages. ASL (American sign language) dataset was used to identify the sign made by gesture. The dataset was consist of numbers zero - nine and alphabets a - z. There were 70 samples for each 36 symbols. This system uses camera as there source which captures various gestures of hand. The captured image changed intoinRGBformatandby using the common approach the RGB image turned into converted into grayscale photo. Then pr-processing of the image takes region wherein edges are detected the usage of Canny facet detection set of rules. Canny area detection algorithm had 5 extraordinary stages specifically, Image smoothing,GradientMagnitude,Non-maximumSuppression, Double threshold , Hysteresis .Finally template matching algorithm turned into used to become aware of the sign and show the textual content. Sum of Absolute Difference (SAD) method was used in comparison of template image with the images present in the data set. In[6],Divya Deroa et al., In this paper simultaneously combining hand shapes, orientation andmovementofhands, fingers or frame,and facial features to fluidly specific a speaker ’ s thought.. Here data is acquired through a 3 Mega pixel cameraand person wears red and blue colouredgloves. system is made only to recognise alphabets and numbers Therefore total number of images in the database is 510 which include 25 alphabets and 9 numbers. The steps included in this experiment are (1) segmentation, (2) fingertip finding algorithm and (3) PCA. Recognition is done using PCA for live data frames where every 20th frame is given as input to be recognized. The system can recognise 94% of the signs made correctly. Fingertip algorithm is used to make database on the basis of number of finger tips. PCA compares the input image with the new database formed with the help of finger tips. In[7],G. Anantha Rao et al., This paper proposes the recognition of Indiansign language using CNN. Thecaptured method usedin this paper was Selfie mode continuous sign language video. Datasets were created with 5 different subjects performing 200 signs in 5 different viewing angles under various circumstances. Out of which 3 different samples size was used forperforming CNN training whereas remaining 2 samples were used for testing the trained CNN. This model was constructedwithinputlayerwheredata’sare given as input ,in feature extraction stage 4 convolution neural network of different window size, five rectified linear unit for non-linearities , 2 stochastic pooling layer to avoid a substantial information loss, in classificationstageonedense and one fully connected layers. In classification Softmax regression was adopted. In[8],Surejya Suresh et al., In this paper the proposed architecture is analysed using two models Stochastic Gradient Descent optimizer and Adam optimizer. The algorithm used is Convolution Neuralnetwork.CNNcontains several layers such as pooling, dropout and fully connected layer. All the important features are extracted from the bottom layer and abstract features are computed from high level layers of CNN. The proposed modelarchitecturehasthe input layer which receives the images of size 64x64 pixels. Later two convolutionoperationsareappliedtotheimagesto
  • 3. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 09 Issue: 01 | Jan 2022 www.irjet.net p-ISSN: 2395-0072 © 2022, IRJET | Impact Factor value: 7.529 | ISO 9001:2008 Certified Journal | Page 1645 generatefeaturemaps. The featuremaps are flattenedwhich results in feature vector. Python is the programming language used. Some of the libraries used are keras, scikit- learn, matplotlib and numpy. In[9],Kusumika Krori Dutta et al.,In this paper Indian Sign Language is communicated using hand gestures made both Single hand and Double hands . In this, only main words are taken into consideration grammar usage is not considered. Articles such as ‘a’, ‘an’, ‘the’ are omitted and also doesn’t include tense form .This paper deals with classification of Indian sign language using machine learning. The system is trained with double handed sign language by using a Principal Component Analysis(PCA) and Artificial neural network (ANN) algorithm in MATLAB. Thehandgesturesare acquired from different sized and complexion hands and made an image database of having ease image of size 768 by 1024. Once the images were preprocessed the size of the image was reduced to 256 by256.Signlanguagepatternsare trained separately using two machinelearningtechniques.A. K Nearest Neighbours Algorithm B. Back Propogation Algorithm. This experiment was conducted over 220 images of double handed Indian sign language alphabets and 800 images of single handed Indian sign language alphabets. EnglishlettersandnumbersusingKNNandBackpropagation techniques where PCA is used for dimension reductionality .For K-NN techniques with K=1, achieved 100% recognition rate whereas using back propagation techniqueachieved94- 96. In[10],LihongZheng et al., Inthis paperAuslanistheprimary language used by the Australian Deaf Community, with an estimated 30,000 usersinAustraliaSignlanguagetranslation to text or speech would give them everyday communication. The fashionable framework of imaginative and prescient based SLR includes pre-processing, sign gesture illustration, function extraction and type. The widely used 2D features include SIFT, HOG, HOF, STIP,and kerneldescriptors. Forthe purpose of dealing with occlusion 3D/4D space of time (HON4D) and Random Occupancy Pattern(ROP) features were proposed. Sign gesture recognition system were divided into two types based on image acquisition: touched based or untouched/vision based There are about seven main large datasets of sign language.TheyareAmericanSign LanguageLexiconVideoDatasetMSRGesture3D,Auslandata set, LTI-Gesture Database, RWTH German Fingerspelling Database, DEVISIGN Chinese Sign Language dataset, and Indian Sign Language dataset. 3. PROPOSED SYSTEM The system uses the web camera for capturing the hand gestures of the hearing and speechimpairedpeople. Theraw images taken is given as an input to the system. The image frames are resized to maintain the equality among all the images. The proposed system includes several stepsnamely, Image Acquisition, Image Pre-Processing, Segmentation ,Feature Extraction, Recognition and Text Output. Fig 2: Schematic Diagram of Sign Language Recognition 3.1 Image Acquisition: This is the first step in the workflow as no processing can be done without getting an image. Image Acquisition is the process of extracting an image from a source, here the source is the hardware-based source that is thewebcamera. This image is not processed in any way so itistherawimage. 3.2 Segmentation: The method of separating objects or signs from the context of a captured image is known as segmentation. This process includes Con-text subtracting, skin-colordetection,andedge detection. The location of the hand must be detected and separated in order to recognize the hand gestures. 3.3 Features Extraction: Predefined features such as form, contour, geometrical feature like position, angle, distance, etc, colour feature, histogram, and others are extracted from the pre-processed images. Later it is used for sign language recognition. Feature extraction is a stepthatdividesand organizes a large collection of raw data. It reduces to smaller and easier-to- manage size so that processing would be simpler. There is massive amount of data sets that have a lot of variable features in it. To process these variable features, a large amount of computation is needed. By Reducing the size of the data these features are simple to use.
  • 4. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 09 Issue: 01 | Jan 2022 www.irjet.net p-ISSN: 2395-0072 © 2022, IRJET | Impact Factor value: 7.529 | ISO 9001:2008 Certified Journal | Page 1646 3.4 Pre-Processing: Pre-processing is the process of eliminating high-intensity noises from the image frames. The first step is smoothing or blurring. Then the image is converted to gray scale from RGB. The main aim of converting the image into grey scale is to reduce the size and amount of processing. The different phases of pre-processing are: 3.4.1 Morphological Transform It is a simple operation based on the shape of the image which are performed on binary images. There are forms of morphological transforms specifically dilation and erosion. 3.4.2 Blurring Blurring is the process of eliminating the noice in an image and keeping required feature intact. 3.4.3 Thresholding The process of convertinggreyscaleimageintobinaryimage is called Thresholding. Thresholding is used to pick thearea of interest ignoring the area that is not required. 3.4.4 Recognition Classifiers are used in this section Classifiers are the algorithm used to interpret the alerts. The classifier used tin this challenge is Convolution Neural Network. Convolution neural network is a deep mastering algorithm which takes the photograph as input, images the critical aspects fromthe photo and differentiates from one another. CNN algorithm offers better precision compared toother algorithminphoto type and reputation. 3.4.5 Text output The hand gestures are recognized and the system gives an output in the human understandable form that is in text. 4. CONCLUSION The Sign Language Recognition (SLR) systemisa method for recognizing a collectionofformedsignsandtranslatingthem into text with the appropriate context. The significance of gesture recognition can be seen in the development of effective human-machine interactions. We attempted to build a model using a Convolutional Neural Network in this project. 5. REFERENCES [1] Citra Suradi , Anik Nur Handayani, Rosa Andrie Asmara, Aji Prasetya Wibawa ,Lilis Nur Hayati,Huzain Azis(2021). “Design Of Sign Language Recognition Using E-CNN”. [2] Lionel Pigou; Sander Dieleman; Pieter-Jan Kindermans; Benjamin Schrauwen(2014). “Sign Language Recognition Using Convolution Neural Network”. [3] Adithya V, Vinod P.R, Usha Gopalakrishnan(2013). “Artificial Neural Network Based Method For Indian Sign Language Recognition”. [4] Muthu Mariappan H, Dr. Gomathi V (2019). “Real Time Recognition of Indian Sign Language”. [5] Soma Shrenika , Myneni Madhu Bala(2020). “Sign Language Recognition UsingTemplateMatchingTechnique”. [6] Divya Deroa, Nikesh Bajaj(2012). “Indian Sign Language recognition”. [7] G. Anantha Rao,K Syamala, P V V Kishore, A S C S Sastry (2018).“DeepConvolutionNeuralNetworkforSignLanguage Recognition”. [8] Surejya Suresh,Mithun Haridas T P, Supriya M H(2019). “Sign Language Recognition using Deep Neural Network”. [9] Kusumika Krori Dutta, Sunny Arokia Swamy Bellary(2017). “MachineLearningTechniquesforIndianSign Language Recognition”. [10] Lihong Zheng, Bin Liang, Ailian Jiang (2017). “ Recent Advances of Deep Learning for Sign Language Recognition”. [11] Rachana Patil, Vivek Patil1, Abhishek Bahuguna1, Mr. Gaurav Datkhile(2021). “Indian Sign Language Recognition Using Convolution Neural Network”.