Sign Language Detection Report
Sign Language Detection Report
A Project Report
Submitted By
S.SHARATH CHANDRA
210304124344
JYOTHI PRABHAS
210303105452
SANDHYA KOIRALA
210303105550
RANADHEER
210304124292
the Degree of
BACHELOR OF TECHNOLOGY
VADODARA
October - 2024
PARUL UNIVERSITY
C ERTIFICATE
This is to Certify that Project - 1 (203105499) of 7th Semester entitled “SIGN LANGUAGE
DETECTOR” of Group No. PUCSE 492 has been successfully completed by
• RANADHEER- 210304124292
Date of Submission :
CSE, PIET,
ii
Acknowledgements
-Auliq-Ice
The satisfaction and euphoria that accompany the successful completion of any task would
be incomplete without the mention of the people who made it possible, whose constant guidance
and encouragement crowned our efforts with success. It is a pleasant aspect that we have now the
opportunity to express our guidance to all of them
We are thankful to ”Dr. DEVANSHU PATEL” chairman of “PARUL UNIVERSITY” for
accepting us and providing us with an opportunity to do a project in their esteemed organization.
We are thankful to Principal ”Dr. VIPUL VEKARIYA” helped us to undergo project work as a
part of the university curriculum.
Our special thanks to ”Mrs.BHAGYESHA PANDHI” Asst. Professor in Computer Science
Engineering Department for guiding us in the right way to complete our project at the right time
And, we would like to express our heartfelt thanks Faculty Members of the Computer
Department, Lab Technicians and Friends, one and all who has helped us directly or indirectly
in successful completion of the Mini Project.
Sign Language Recognition is a breakthrough for communication among deaf-mute society and
has been a critical research topic for years. Although some of the previous studies have successfully
recognized sign language, it requires many costly instruments including sensors, devices, and high-
end processing power. However, such drawbacks can be easily overcome by employing artificial
intelligence-based techniques. Since, in this modern era of advanced mobile technology, using a
camera to take video or images is much easier, this study demonstrates a cost-effective technique to
detect American Sign Language (ASL) using an image dataset. Here, “Finger Spelling, A” dataset
has been used, with 24 letters (except j and z as they contain motion). The main reason for using
this dataset is that these images have a complex background with different environments and scene
colors
A real time sign language detector is a significant step forward in improving communication
between the deaf and the general population. We are pleased to showcase the creation and
implementation of sign language recognition model based on a Convolutional Neural
Network(CNN).We utilized a Pre-Trained SSD Mobile net V2 architecture trained on our own
dataset in order to apply Transfer learning to the task. We developed a robust model that
consistently classifies Sign language in majority of cases. Additionally, this strategy will be
extremely beneficial to sign language learners in terms of practising sign language. Various
human-computer interface methodologies for posture recognition were explored and assessed
during the project. A series of image processing techniques with Human movement classification
was identified as the best approach. The system is able to recognize selected Sign Language signs
with the accuracy of 70-80
There are several ways for recognising gestures which includes sensor-based and vision-based
systems. Sensor-equipped devices capture numerous parameters such as the trajectory, location,
and velocity of the hand in the sensor-based approach. On the other hand,vision-based approaches
are those in which images of video footages of the hand gestures are used.
The main purpose of sign language detection system is providing a feasible way of
communication between a normal and dumb people by using hand gesture. The proposed system
can be accessed by using webcam or any in-built camera that detects the signs and processes them
for recognition.
Table of Contents
Acknowledgements iii
Abstract iv
List of Tables ix
List of Figures ix
1 Introduction 1
2 Literature Survey 3
2.1 Paper 1: Automation sign language analysis : Asurvey and the future beyond lexical
meaning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
2.2 Paper 2: Indian sign language to text conversion in real time using machine learning 4
2.3 Paper 3: Sign language recognition system for deaf and dumb people . . . . . . . . 4
v
TABLE OF CONTENTS
2.14 Paper 14:Deep learning-based sign language recognition system for static signs . . 9
2.15 Paper 15:American sign language recognition technique using 3-D convolutions
neural networks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
2.16 Paper 16:Sign Language Translation Using Deep Convolutional Neural Networks . 10
NEURAL NETWORKS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
2.18 Paper 18:Simulation and Analysis of Hand Gesture Recognition for Indian Sign
2.19 Paper 19:Deep convolutional neural network for hand sign language recognition
using model E . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
3.1 INTRODUCTION . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
3.2 PURPOSE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
vi
TABLE OF CONTENTS
4 System Design 16
5 Methodology 21
5.4 DEVELOPMENT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
6 Implementation 23
6.1 INTRODUCTION . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
7 Conclusion 27
8 Future Work 28
vii
TABLE OF CONTENTS
viii
List of Figures
6.1 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
6.2 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
6.3 A . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
6.4 B . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
ix
Chapter 1
Introduction
1
CHAPTER 1. INTRODUCTION
2
Chapter 2
Literature Survey
2.1 Paper 1: Automation sign language analysis : Asurvey and the future
3
CHAPTER 2. LITERATURE SURVEY
2.2 Paper 2: Indian sign language to text conversion in real time using
machine learning
Author :-Yash Narang, Aditya Sharma
ABSTRACT:Before language as we know it today existed, correspondence between individual
people was emblematic and comprised of the utilization of hand signs. Sign language is one of the
oldest and most natural form of language for communication. In our community today, contact with
hearing impaired (deaf/mute) persons is a major challenge. This can be attributed to the fact that
their method of communication (sign language or hand gestures at community level) requires an
intermediary at all levels. We have thought of a constant strategy utilizing neural organizations for
finger spelling dependent on Indian gesture-based communication. In our technique, the hand is
first gone through a filter and after the filter is applied the hand is passed through a classifier that
predicts the class of the hand motions. Our strategy provides 95.7for the 26 letters of the alphabet.
2.3 Paper 3: Sign language recognition system for deaf and dumb people
Author :-Sakshi Goyal, Ishita Sharma, Shanu Sharma
ABSTRACT:The Sign language is very important for people who have hearing and speaking
deficiency generally called Deaf And Mute. It is the only mode of communication for such people
to convey their messages and it becomes very important for people to understand their language.
This paper proposes the method or algorithm for an application which would help in recognising
the different signs which is called Indian Sign Language. The images are of the palm side of right
and left hand and are loaded at runtime. The method has been developed with respect to single user.
The real time images will be captured first and then stored in directory and on recently captured
image and feature extraction will take place to identify which sign has been articulated by the user
through SIFT (scale in variance Fourier transform) algorithm. The comparisons will be performed
in arrears and then after comparison the result will be produced in accordance through matched
keypoints from the input image to the image stored for a specific letter already in the directory or
the database the outputs for the following can be seen in below sections. There are 26 signs in
Indian Sign Language corresponding to each alphabet out which the proposed algorithm provided
with 95results for 9 alphabets with their images captured at every possible angle and distance i.e.
for every alphabet even if have approximately 5 images at different angles and distances then the
algorithm is working accurately for 45 types of inputs.
4
CHAPTER 2. LITERATURE SURVEY
5
CHAPTER 2. LITERATURE SURVEY
information which is then input into various supervised learning techniques. Currently, we have
reported four fold cross validated results for the different approaches, and the difference from the
previous work done can be attributed to the fact that in our four fold i cross validation, the
validation set Correspond to images of a iperson different from the persons in the training set.
Networks
Author:Lionel Pigou, Sander Dieleman, Pieter-Jan Kindermans, Benjamin Schrauwen Ghent
University, ELIS, Belgium
ABSTRACT:There is an undeniable communication problem between the Deaf community and
the hearing majority. Innovations in automatic sign language recognition try to tear down this
communication barrier. Our contribution considers a recognition system using the Microsoft Kinect,
convolutional neural networks (CNNs) and GPU acceleration. Instead of constructing complex
handcrafted features, CNNs are able to automate the process of feature construction. We are able
to recognize 20 Italian gestures with high accuracy. The predictive model is able to generalize on
users and surroundings not occurring during training with a cross-validation accuracy of 91.7Index
of 0.789 in the ChaLearn 2014 Looking at People gesture spotting competition.
6
CHAPTER 2. LITERATURE SURVEY
7
CHAPTER 2. LITERATURE SURVEY
began to show that these languages are bona fide linguistic systems, with structures and rules and
the full range of expressive power that characterize spoken languages. Researchers have spent
most of that time demonstrating, with increasing rigor and formality, the sometimes surprising
similarities between languages in the two modalities, spoken and signed. Concomitantly, scholars in
the related disciplines of language acquisition and neurolinguistics have been discovering significant
similarities between spoken and signed languages in these domains as well. It is safe to say that the
academic world is now convinced that sign languages are real languages in every sense of the term.
If this were the whole story, however, there would be no need for a chapter on sign languages in
this volume. Each sign language would be seen as a language like any other, English, Hungarian,
Central Alaskan Yupik Eskimo, or Mandarin Chinese, each with its own contribution to make
toward understanding the general language faculty of humans. But this is not the whole story.
recognition
Author:Tolga Yirtici1 · Kamil Yurtkan
ABSTRACT:Hearing impaired people use sign language for communication. Considering the
spoken languages, the most widely spread ones have their own sign languages. Thus, sign languages
are spoken language dependent. In this paper, the focus is on Turkish Sign Language (TSL). In
parallel with the developments in computer science and artificial intelligence, the human computer
interaction is now possible and it is a definite field of computer science. Automatically detecting
and recognizing sign languages by computers are also possible applications. The improvements in
machine learning field in the last decade made it possible to improve systems that can automatically
detect and recognize sign language from still images. On the other hand, using a robust algorithm
that is recognizing a sign language may not be applicable to another language as signs and linguistic
properties may differ between languages. In this paper, we propose a novel method of recognizing
characters of Turkish Sign Language (TSL). The method is tested on captured images containing
signs, which are extracted from video files. Alexnet is employed as a pre-trained network in
this system. Region-based Convolutional Neural Network (R-CNN) object detector is used to
train the new model. The knowledge of the neural network is transferred using transfer learning
method and tuned for recognizing TSL. The system achieves 99.7conventional methods applied
on TSL. Furthermore, the study is significant since it is one of the first successful application of
transfer learning approach to TSL. The system is open to further improvements by improving the
8
CHAPTER 2. LITERATURE SURVEY
static signs
Author:Ankita Wadhawan • Parteek Kumar
ABSTRACT:Sign language for communication is efficacious for humans, and vital research is in
progress in computer vision systems. The earliest work in Indian Sign Language (ISL) recognition
considers the recognition of significant differentiable hand signs and therefore often selecting a
few signs from the ISL for recognition. This paper deals with robust modeling of static signs in
the context of sign language recognition using deep learning-based convolutional neural networks
(CNN). In this research, total 35,000 sign images of 100 static signs are collected from different
users. The efficiency of the proposed system is evaluated on approximately 50 CNN models. The
results are also evaluated on the basis of different optimizers, and it has been observed that the
proposed approach has achieved the highest training accuracy of 99.72colored and grayscale images,
respectively. The performance of the proposed system has also been evaluated on the basis of
precision, recall and F-score. The system also demonstrates its effectiveness over the earlier works
in which only a few hand signs are considered for recognition.
9
CHAPTER 2. LITERATURE SURVEY
Networks
Author:Rahib H.Abiyev1, Murat Arslan1 and John Bush Idoko1
ABSTRACT:Sign language is a natural, visually oriented and non-verbal communication channel
between people that facilitates communication through facial/bodily expressions, postures and a set
of gestures. It is basically used for communication with people who are deaf or hard of hearing.
In order to understand such communication quickly and accurately, the design of a successful
sign language translation system is considered in this paper. The proposed system includes object
detection and classification stages. Firstly, Single Shot Multi Box Detection (SSD) architecture is
utilized for hand detection, then a deep learning structure based on the Inception v3 plus Support
Vector Machine (SVM) that combines feature extraction and classification stages is proposed to
constructively translate the detected hand gestures. A sign language fingerspelling dataset is used
for the design of the proposed model. The obtained results and comparative analysis demonstrate
the efficiency of using the proposed hybrid structure in sign language translation.
10
CHAPTER 2. LITERATURE SURVEY
11
CHAPTER 2. LITERATURE SURVEY
2.19 Paper 19:Deep convolutional neural network for hand sign language
12
Chapter 3
3.1 INTRODUCTION
Sign language detection using AI/ML(Artificial intelligence/ Machine Learning) system aims to
assist deaf and dumb people to maintain communication it converts the sign language into subtitles
which makes it easier for any user to communicate This document outlines the requirements for the
development of this software system
3.2 PURPOSE
The purpose of this software to provide a platform for normal people able to communicate with deaf
and dumb people using advanced AI and ML algorithms . by analysing the subject signs played on
video it detects the sings and convert it into subtitles which makes it easier for users to understand
the system will classify and provide information about the sign which helps in aiding in timely and
effective management practices
13
CHAPTER 3. ANALYSIS / SOFTWARE REQUIREMENTS SPECIFICATION (SRS)
” Text generated from this application can be converted to speech for better communication. ”
Use hand gestures to control and automate other devices.
14
CHAPTER 3. ANALYSIS / SOFTWARE REQUIREMENTS SPECIFICATION (SRS)
15
Chapter 4
System Design
16
CHAPTER 4. SYSTEM DESIGN
17
CHAPTER 4. SYSTEM DESIGN
18
CHAPTER 4. SYSTEM DESIGN
19
CHAPTER 4. SYSTEM DESIGN
20
Chapter 5
Methodology
21
CHAPTER 5. METHODOLOGY
5.4 DEVELOPMENT
-Design system architecture, considering input sources, preprocessing techniques, feature extraction
methods, classification algorithms, translation mechanisms, and output formats.
Our project aims to develop a robust system for identifying sign language signs then turning into
text using advanced AI/ML we seek to provide users and experts with a reliable tool to understand
deaf and dumb people efficiently.
22
Chapter 6
Implementation
6.1 INTRODUCTION
The Sign Language Detector aims to bridge communication barriers by interpreting sign language
gestures and translating them into text or spoken language. This implementation involves several
key stages, including data collection, data preprocessing, machine learning model selection, training,
and evaluation.
23
CHAPTER 6. IMPLEMENTATION
3.Color Features: Analyze variations in color intensity and distribution across the video area of
the subject
24
CHAPTER 6. IMPLEMENTATION
Figure 6.1: 1
Figure 6.2: 2
25
CHAPTER 6. IMPLEMENTATION
Figure 6.3: A
Figure 6.4: B
26
Chapter 7
Conclusion
This work proposes a methodology for perceiving the classification of sign language recognition.
Sign language is the core medium of communication between deaf-mute and everyday people. It
is highly implacable in real-world scenarios like communication, human–computer interaction,
security, advanced AI, and much more. For a long time, researchers have been working in this field
to make a reliable, low cost and publicly available SRL system using different sensors, images,
videos, and many more techniques. Detecting sign language has become an important research
field to improve communication with deaf and dumb people. It is also important to understand that
different sign languages are developed in different language communities, and research on sign
language detection is also language-specific. Even though English is a mainstream language with a
large deaf and dumb community, there has been very little research conducted on sign language
detection in English. In this paper, we propose a new English sign language detection scheme that
relies on fingertip position as a training tool for a CNN classifier. Several methods have been tested
and compared against a large dataset of images. Based on test set accuracy, the proposed method
outperforms all existing methods. In addition, the proposed scheme appears to be compact and
efficient in terms of computation and size.
This Project is a working system for the conversion of Indian Sign Language to text in real-time
for deaf and mute people. For now, this project is focused on the single spelling-based signs where
each alphabet is represented by a particular hand sign. The accuracy achieved is 97.3percent on our
testing dataset. The approach we have used is that the training data has some background noise like
patterned clothes other objects in the background which might be present in real-world scenarios.
However, the distance dataset has negligible background noise adequate lighting. This way we can
train the model to ignore the background noise focus only on the hand signs and gestures, therefore
getting better results.
27
Chapter 8
Future Work
28
CHAPTER 8. FUTURE WORK
29
References
1. S. Akyol and P. Alvarado, “Finding Relevant Image Content for mobile Sign Language
Recognition,” Proc. IASTED Int’l Conf. Signal Processing, Pattern Recognition and
Application, pp. 48-52, 2001
2. Lin, H.-I., Hsu, M.-H., and Chen, W.-K. (2014). “Human hand gesture recognition using
a convolution neural network.” IEEE International Conference on Automation Science and
Engineering
3. Pallavi Gurjal, Kiran Kunnur, “Real Time Hand Gesture Recognition using SIFT”,
International Journal for Electronics and Engineering, 2012,pp 19-33.
4. Bahl, L., Jelinek, F., Mercer, R.: A maximum likelihood approach to continuous speech
recognition. IEEE Trans. Pattern Anal. Mach. Intell
5. Brien, David (ed.). 1992. Dictionary of British Sign Language/English. London: Faber Faber.
6. XIAOLONG TENG, BIANI WU, W. Y., AND LIU, C. A hand gesture recognition system
based on local linear embedding, April 2005.
7. . Bergstra, J., Breuleux, O., Bastien, F., Lamblin, P., Pascanu, R., Desjardins, G., Turian,
J., Warde-Farley, D., Bengio, Y.: Theano: a CPU and GPU math expression compiler. In:
Proceedings of the Python for Scientific Computing Conference (SciPy) (Jun 2010), oral
Presentation
8. Chai, X., Li, G., Lin, Y., Xu, Z., Tang, Y., Chen, X., Zhou, M.: Sign Language Recognition
and Translation with Kinect (2013), http://vipl.ict.ac.cn/sites/default/files/papers/files/2013
FG xjchai Sign Language Recognition and Translation with Kinect.pdf
9. Cire¸san, D., Meier, U., Schmidhuber, J.: Multi-column deep neural networks for image
classification. In: Computer Vision and Pattern Recognition (CVPR),
30
CHAPTER 8. FUTURE WORK
10. Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton, “Imagenet classification with deep
convolutional neural networks,
11. Andrej Karpathy, George Toderici, Sanketh Shetty, Thomas Leung, Rahul Sukthankar, and Li
Fei-Fei, “Large-scale video classification with convolutional neural networks,”
12. Yann LeCun, Leon Bottou, Yoshua Bengio, and Patrick ´ Haffner, “Gradient-based learning
applied to document recognition,”
13. Hueihan Jhuang, Thomas Serre, Lior Wolf, and Tomaso Poggio, “A biologically inspired
system for action recognition,”
14. Shuiwang Ji, Wei Xu, Ming Yang, and Kai Yu, “3D convolutional neural networks for human
action recognition,”
15. Kirsti Grobel and Marcell Assan, “Isolated sign language recognition using hidden markov
models,” in Systems, Man, and Cybernetics
16. Christian Vogler and Dimitris Metaxas, “Parallel hidden markov models for american sign
language recognition,”
17. Kouichi Murakami and Hitomi Taguchi, “Gesture recognition using recurrent neural
networks,”
18. Jong-Sung Kim, Won Jang, and Zeungnam Bien, “A dynamic gesture recognition system for
the korean sign language (ksl),”
19. Ross Girshick, Jeff Donahue, Trevor Darrell, and Jitendra Malik, “Rich feature hierarchies for
accurate object detection and semantic segmentation,”
20. M.A. Jalal, R.Chen, R.Moore, et al., “American sign language posture understanding with
deep neural networks,”
31