[go: up one dir, main page]

100% found this document useful (1 vote)
1K views40 pages

Sign Language Detection Report

Latex sign language detection
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
100% found this document useful (1 vote)
1K views40 pages

Sign Language Detection Report

Latex sign language detection
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 40

SIGN LANGUAGE DETECTION

A Project Report

Submitted By

S.SHARATH CHANDRA

210304124344

JYOTHI PRABHAS

210303105452

SANDHYA KOIRALA

210303105550

RANADHEER

210304124292

in Partial Fulfilment For the Award of

the Degree of

BACHELOR OF TECHNOLOGY

COMPUTER SCIENCE & ENGINEERING

Under the Guidance of

ASST PROF. SUDHANSHU SINGH

ASST PROF. NIGAM KUMARI

VADODARA

October - 2024
PARUL UNIVERSITY
C ERTIFICATE
This is to Certify that Project - 1 (203105499) of 7th Semester entitled “SIGN LANGUAGE
DETECTOR” of Group No. PUCSE 492 has been successfully completed by

• SHARATH CHANDRA - 210304124344

• JYOTHI PRABHAS- 210303105452

• SANDHYA KOIRALA- 210303105550

• RANADHEER- 210304124292

under my guidance in partial fulfillment of the Bachelor of Technology (B.Tech) in Computer


Science & Engineering of Parul University in Academic Year 2024- 2025.

Date of Submission :

ASST PROF. SUDHANSHU SINGH ,

ASST PROF, NIGAM KUMARI Dr. Amit Barve,

Project Guide Head of Department,

CSE, PIET,

Project Coordinator:-Dr.Kruti sutaria,

Prof.Yatin shukla Parul University.

ii
Acknowledgements

“The single greatest cause of happiness is gratitude.”

-Auliq-Ice

The satisfaction and euphoria that accompany the successful completion of any task would
be incomplete without the mention of the people who made it possible, whose constant guidance
and encouragement crowned our efforts with success. It is a pleasant aspect that we have now the
opportunity to express our guidance to all of them
We are thankful to ”Dr. DEVANSHU PATEL” chairman of “PARUL UNIVERSITY” for
accepting us and providing us with an opportunity to do a project in their esteemed organization.
We are thankful to Principal ”Dr. VIPUL VEKARIYA” helped us to undergo project work as a
part of the university curriculum.
Our special thanks to ”Mrs.BHAGYESHA PANDHI” Asst. Professor in Computer Science
Engineering Department for guiding us in the right way to complete our project at the right time
And, we would like to express our heartfelt thanks Faculty Members of the Computer
Department, Lab Technicians and Friends, one and all who has helped us directly or indirectly
in successful completion of the Mini Project.

Student name- Enrollment no:


SHARATH CHANADRA-210304124344
JYOTHI PRABHAS-210303104552
SANDHYA KOIRALA-210303105550
RANADHEER-210304124292
CSE, PIET
Parul University,
Vadodara
Abstract

Sign Language Recognition is a breakthrough for communication among deaf-mute society and
has been a critical research topic for years. Although some of the previous studies have successfully
recognized sign language, it requires many costly instruments including sensors, devices, and high-
end processing power. However, such drawbacks can be easily overcome by employing artificial
intelligence-based techniques. Since, in this modern era of advanced mobile technology, using a
camera to take video or images is much easier, this study demonstrates a cost-effective technique to
detect American Sign Language (ASL) using an image dataset. Here, “Finger Spelling, A” dataset
has been used, with 24 letters (except j and z as they contain motion). The main reason for using
this dataset is that these images have a complex background with different environments and scene
colors
A real time sign language detector is a significant step forward in improving communication
between the deaf and the general population. We are pleased to showcase the creation and
implementation of sign language recognition model based on a Convolutional Neural
Network(CNN).We utilized a Pre-Trained SSD Mobile net V2 architecture trained on our own
dataset in order to apply Transfer learning to the task. We developed a robust model that
consistently classifies Sign language in majority of cases. Additionally, this strategy will be
extremely beneficial to sign language learners in terms of practising sign language. Various
human-computer interface methodologies for posture recognition were explored and assessed
during the project. A series of image processing techniques with Human movement classification
was identified as the best approach. The system is able to recognize selected Sign Language signs
with the accuracy of 70-80
There are several ways for recognising gestures which includes sensor-based and vision-based
systems. Sensor-equipped devices capture numerous parameters such as the trajectory, location,
and velocity of the hand in the sensor-based approach. On the other hand,vision-based approaches
are those in which images of video footages of the hand gestures are used.
The main purpose of sign language detection system is providing a feasible way of
communication between a normal and dumb people by using hand gesture. The proposed system
can be accessed by using webcam or any in-built camera that detects the signs and processes them
for recognition.
Table of Contents

Acknowledgements iii

Abstract iv

List of Tables ix

List of Figures ix

1 Introduction 1

1.1 INTRODUCTION OF PROJECT . . . . . . . . . . . . . . . . . . . . . . . . . . 1

1.2 PROBLEM STATEMENT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1

1.3 SCOPE OF THE PROJECT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2

1.4 PURPOSE OF THE PROJECT . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2

2 Literature Survey 3

2.1 Paper 1: Automation sign language analysis : Asurvey and the future beyond lexical

meaning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3

2.2 Paper 2: Indian sign language to text conversion in real time using machine learning 4

2.3 Paper 3: Sign language recognition system for deaf and dumb people . . . . . . . . 4

2.4 Paper 4:Recent developments in visual sign language recognition . . . . . . . . . . 5

2.5 Paper 5 :Documenting sign language detection system . . . . . . . . . . . . . . . 5

2.6 Paper 6: Sign language using machine learning algorithm . . . . . . . . . . . . . . 5

v
TABLE OF CONTENTS

2.7 Paper 7: Sign Language Recognition using Convolutions Neural Networks . . . . . 6

2.8 Paper 8:Real Time Sign Language Detection . . . . . . . . . . . . . . . . . . . . . 6

2.9 Paper 9:Indian Sign Language Recognition System . . . . . . . . . . . . . . . . . 7

2.10 Paper 10:Sign Language Production: A Review . . . . . . . . . . . . . . . . . . . 7

2.11 Paper 11:Natural Sign Languages . . . . . . . . . . . . . . . . . . . . . . . . . . 7

2.12 Paper 12:Regional-CNN-based enhanced Turkish sign language recognition . . . . 8

2.13 Paper 13:Recent developments in visual sign language recognition . . . . . . . . . 9

2.14 Paper 14:Deep learning-based sign language recognition system for static signs . . 9

2.15 Paper 15:American sign language recognition technique using 3-D convolutions

neural networks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10

2.16 Paper 16:Sign Language Translation Using Deep Convolutional Neural Networks . 10

2.17 Paper 17:SIGN LANGUAGE RECOGNITION USING 3D CONVOLUTIONAL

NEURAL NETWORKS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11

2.18 Paper 18:Simulation and Analysis of Hand Gesture Recognition for Indian Sign

Language using CNN . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11

2.19 Paper 19:Deep convolutional neural network for hand sign language recognition

using model E . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12

2.20 Paper 20:Sign Language Recognition . . . . . . . . . . . . . . . . . . . . . . . . 12

3 Analysis / Software Requirements Specification (SRS) 13

3.1 INTRODUCTION . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13

3.2 PURPOSE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13

3.3 PRODUCT SCOPE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13

3.4 DOCUMENT CONVENTIONS . . . . . . . . . . . . . . . . . . . . . . . . . . . 14

3.5 USER INTERFACE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14

3.6 HARDWARE INTERFACE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15

3.7 HARDWARE REQUIREMENTS . . . . . . . . . . . . . . . . . . . . . . . . . . 15

vi
TABLE OF CONTENTS

3.8 SOFTWARE REQUIREMENTS . . . . . . . . . . . . . . . . . . . . . . . . . . . 15

4 System Design 16

4.1 INTRODUCTION TO SYSTEM DESIGN . . . . . . . . . . . . . . . . . . . . . . 16

4.2 SYSTEM ARCHITECTURE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16

4.3 FUNCTIONAL COMPONENTS . . . . . . . . . . . . . . . . . . . . . . . . . . . 17

4.4 MONITORING AND MAINTENANCE . . . . . . . . . . . . . . . . . . . . . . . 17

4.5 TESTING AND QUALITY ASSURANCE . . . . . . . . . . . . . . . . . . . . . 17

5 Methodology 21

5.1 REASEARCH ANALYSIS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21

5.2 REQUIREMENT GATHERING . . . . . . . . . . . . . . . . . . . . . . . . . . . 21

5.3 DESIGN AND PLANNING . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21

5.4 DEVELOPMENT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22

5.5 TESTING AND DEPLOYMENT . . . . . . . . . . . . . . . . . . . . . . . . . . 22

5.6 DEPLOYMENT PLANNING . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22

6 Implementation 23

6.1 INTRODUCTION . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23

6.2 DATA COLLECTION . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23

6.3 DATA PREPROCESSING . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23

6.4 MACHINE LEARNING MODEL SELECTION AND TRAINING: . . . . . . . . 24

6.5 MODEL TRAINING . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24

7 Conclusion 27

8 Future Work 28

8.1 MULTIPLE SIGN LANGUAGES DETECTOR . . . . . . . . . . . . . . . . . . . 28

8.2 ENHANCED ACCURACY . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28

vii
TABLE OF CONTENTS

8.3 GESTURE RECOGNITION . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28

8.4 REAL-TIME FEEDBACK . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28

8.5 ETHICAL CONSIDERATION . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29

8.6 ADAPTIVE ASSISTANCE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29

8.7 3D SIGN LANGUAGE AVATAR . . . . . . . . . . . . . . . . . . . . . . . . . . 29

viii
List of Figures

4.1 USE CASE DIAGRAM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18


4.2 DATA FLOW DIAGRAM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
4.3 STATE DIAGRAM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20

6.1 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
6.2 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
6.3 A . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
6.4 B . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26

ix
Chapter 1

Introduction

1.1 INTRODUCTION OF PROJECT


This is an important problem because, according to the Communication Service for the Deaf (CSD)
over 1 million people communicate using American Sign Language (ASL), and 93percecnt of them
prefer to communicate using ASL, including in business settings. They are economically very
important, representing an (86) Billion dollar market (according to the CSD). However, businesses
struggle to serve deaf people because they often don’t have the people (who can sign) or the
technological signs to communicate. Our research showed that there are less than 10 (not-widely-
used) sign language detection systems available on the market. Another very important use case is
for deaf children. The CSD states that 98 f deaf people do not receive an education in sign language,
and 72percent of families do not sign with their deaf children. By building a real-time sign language
detection system, we can help to bridge this gap and help businesses communicate with their deaf
customers. Specifically, it is our goal to build this system that can take images, videos, and real-time
sign language inputs and return an English alphabet letter detection (i.e translation).

1.2 PROBLEM STATEMENT


”Develop a machine learning-based system capable of accurately detecting and interpreting sign
language gestures from real-time video feeds. The goal is to enable seamless communication
between individuals who are deaf or hard of hearing and those who do not know sign language.
The system should efficiently recognize a variety of gestures, including both static and dynamic
signs, and translate them into corresponding text or speech outputs. The model’s performance will
be evaluated based on accuracy, speed, scalability, and its ability to handle diverse backgrounds,
lighting conditions, and variations in signing style among different users.”

1
CHAPTER 1. INTRODUCTION

1.3 SCOPE OF THE PROJECT


More sensors will be inserted soon to detect the entire sign language with greater precision and
accuracy. Most units will be embedded on a single board, resulting in a small model. The technology
might also be programmed to translate words from one language to another. This technology can
be used in any application or software and can be used in multiple place It can be made as handy
as google translate, but the only difference will be a translator for Sign language. It can be highly
scalable and with not require any material support. It can be fitted into any kiosk. It can be utilized
in a variety of settings, such as hospitals, banks, and airports, to help disabled persons communicate
more easily.

1.4 PURPOSE OF THE PROJECT


The main purpose of sign language detection system is providing a feasible way of communication
between a normal and dumb people by using hand gesture. The proposed system can be accessed
by using webcam or any in-built camera that detects the signs and processes them for recognition.
From the result of the model, we can conclude that the proposed system can give accurate results
under controlled light and intensity. Furthermore, custom gestures can easily be added and more
the images taken at different angle and frame will provide more accuracy to the model. Thus, the
model can easily be extended on a large scale by increasing the dataset.

2
Chapter 2

Literature Survey

2.1 Paper 1: Automation sign language analysis : Asurvey and the future

beyond lexical meaning


Author:-Sylvie C.W. Ong and Surendra Ranganath
ABSTRACT:Research in automatic analysis of sign language has largely focused on recognizing
the lexical (or citation) form of sign gestures as they appear in continuous signing, and developing
algorithms that scale well to large vocabularies. However, successful recognition of lexical signs is
not sufficient for a full understanding of sign language communication. Nonmanual signals and
grammatical processes which result in systematic variations in sign appearance are integral aspects
of this communication but have received comparatively little attention in the literature. In this survey,
we examine data acquisition, feature extraction and classification methods employed for the analysis
of sign language gestures. These are discussed with respect to issues such as modeling transitions
between signs in continuous signing, modeling inflectional processes, signer independence, and
adaptation. We further examine works that attempt to analyze nonmanual signals and discuss issues
related to integrating these with (hand) sign gestures. We also discuss the overall progress toward a
true test of sign recognition systems—dealing with natural signing by native signers. We suggest
some future directions for this research and also point to contributions it can make to other fields
of research. Web-based supplemental materials (appendicies) which contain several illustrative
examples and videos of signing can be found at www.computer.org/publications/dlib.

3
CHAPTER 2. LITERATURE SURVEY

2.2 Paper 2: Indian sign language to text conversion in real time using

machine learning
Author :-Yash Narang, Aditya Sharma
ABSTRACT:Before language as we know it today existed, correspondence between individual
people was emblematic and comprised of the utilization of hand signs. Sign language is one of the
oldest and most natural form of language for communication. In our community today, contact with
hearing impaired (deaf/mute) persons is a major challenge. This can be attributed to the fact that
their method of communication (sign language or hand gestures at community level) requires an
intermediary at all levels. We have thought of a constant strategy utilizing neural organizations for
finger spelling dependent on Indian gesture-based communication. In our technique, the hand is
first gone through a filter and after the filter is applied the hand is passed through a classifier that
predicts the class of the hand motions. Our strategy provides 95.7for the 26 letters of the alphabet.

2.3 Paper 3: Sign language recognition system for deaf and dumb people
Author :-Sakshi Goyal, Ishita Sharma, Shanu Sharma
ABSTRACT:The Sign language is very important for people who have hearing and speaking
deficiency generally called Deaf And Mute. It is the only mode of communication for such people
to convey their messages and it becomes very important for people to understand their language.
This paper proposes the method or algorithm for an application which would help in recognising
the different signs which is called Indian Sign Language. The images are of the palm side of right
and left hand and are loaded at runtime. The method has been developed with respect to single user.
The real time images will be captured first and then stored in directory and on recently captured
image and feature extraction will take place to identify which sign has been articulated by the user
through SIFT (scale in variance Fourier transform) algorithm. The comparisons will be performed
in arrears and then after comparison the result will be produced in accordance through matched
keypoints from the input image to the image stored for a specific letter already in the directory or
the database the outputs for the following can be seen in below sections. There are 26 signs in
Indian Sign Language corresponding to each alphabet out which the proposed algorithm provided
with 95results for 9 alphabets with their images captured at every possible angle and distance i.e.
for every alphabet even if have approximately 5 images at different angles and distances then the
algorithm is working accurately for 45 types of inputs.

4
CHAPTER 2. LITERATURE SURVEY

2.4 Paper 4:Recent developments in visual sign language recognition


Author:Jo¨rg Zieren Æ Ulrich Canzler Æ Britta Bauer Æ Karl-Friedrich Kraiss
ABSTRACT:Research in the field of sign language recognition has made significant advances in
recent years. The present achievements provide the basis for future applications with the objective of
supporting the integration of deaf people into the hearing society. Translation systems, for example,
could facilitate communication between deaf and hearing people in public situations. Further
applications, such as user interfaces and automatic indexing of signed videos, become feasible. The
current state in sign language recognition is roughly 30 years behind speech recognition, which
corresponds to the gradual transition from isolated to continuous recognition for small vocabulary
tasks.

2.5 Paper 5 :Documenting sign language detection system


Author: Adam Schembri
ABSTRACT:As background, I have worked primarily on Australian Sign Language (Auslan); it
has been easy for me to move between British Sign Language (BSL) and Auslan because they are
arguably dialects of the same sign language. New Zealand Sign Language (NZSL) also forms part
of the BSL family, so the three sign languages might best be considered dialects of the same variety
(Johnston 2003). They are quite different from American Sign Language (ASL); BSL and ASL
have quite different histories and are not completely mutually intelligible

2.6 Paper 6: Sign language using machine learning algorithm


Author:Prof.Radha S.Shirbhate, Mr.Vedant D.Shinde, Ms.Sanam A.Metkari, Ms.Pooja U.
Borkar,Ms. Mayuri A.Khandge
ABSTRACT:An Sign Language is one of the way to communicate with deaf people. In this work
sets, included features and variation in the language with locality have been the major barriers
which has led to little research being done in ISL. One should learn sign language to interact with
them. Learning usually takes place in peer groups. There are very few study materials available for
sign learning. Because of this, the process of learning sign language learning is a very difficult task.
The initial stage of isign learning is Finger spelled signi learning and moreover, are used when no
corresponding sign exists or signer is not aware of it. Most of the existing tools for sign language
learning use external sensors which are costly. Our project aims at extending a step forward in this
field by collecting a dataset and then use various feature extraction techniques to extract useful

5
CHAPTER 2. LITERATURE SURVEY

information which is then input into various supervised learning techniques. Currently, we have
reported four fold cross validated results for the different approaches, and the difference from the
previous work done can be attributed to the fact that in our four fold i cross validation, the
validation set Correspond to images of a iperson different from the persons in the training set.

2.7 Paper 7: Sign Language Recognition using Convolutions Neural

Networks
Author:Lionel Pigou, Sander Dieleman, Pieter-Jan Kindermans, Benjamin Schrauwen Ghent
University, ELIS, Belgium
ABSTRACT:There is an undeniable communication problem between the Deaf community and
the hearing majority. Innovations in automatic sign language recognition try to tear down this
communication barrier. Our contribution considers a recognition system using the Microsoft Kinect,
convolutional neural networks (CNNs) and GPU acceleration. Instead of constructing complex
handcrafted features, CNNs are able to automate the process of feature construction. We are able
to recognize 20 Italian gestures with high accuracy. The predictive model is able to generalize on
users and surroundings not occurring during training with a cross-validation accuracy of 91.7Index
of 0.789 in the ChaLearn 2014 Looking at People gesture spotting competition.

2.8 Paper 8:Real Time Sign Language Detection


Author:Aman Pathak, Avinash Kumar,Priyam,Priyanshu Gupta,Gunjan Chugh
ABSTRACT:A real time sign language detector is a significant step forward in improving
communication between the deaf and the general population. We are pleased to showcase the
creation and implementation of sign language recognition model based on a Convolutional Neural
Network(CNN).We utilized a Pre-Trained SSD Mobile net V2 architecture trained on our own
dataset in order to apply Transfer learning to the task. We developed a robust model that
consistently classifies Sign language in majority of cases. Additionally, this strategy will be
extremely beneficial to sign language learners in terms of practising sign language. Various
human-computer interface methodologies for posture recognition were explored and assessed
during the project. A series of image processing techniques with Human movement classification
was identified as the best approach. The system is able to recognize selected Sign Language signs
with the accuracy of 70-80PERCENT without a controlled background with small light.

6
CHAPTER 2. LITERATURE SURVEY

2.9 Paper 9:Indian Sign Language Recognition System


Author:Yogeshwar I. Rokade , Prashant M. Jadav
ABSTRACT:—Normal humans can easily interact and communicate with one another, but the
person with hearing and speaking disabilities face problems in communicating with other hearing
people without a translator. The Sign Language is a barrier of communication for deaf and dumb
people. People with hearing and speaking disability are highly dependent on non-verbal form of
communication that involves hand gesture. This is the reason that the implementation of a system
that recognize the sign language would have a significant benefit impact on dumb - deaf people. In
this paper, a method is proposed for the automatic recognition of the finger spelling in the Indian
sign language. Here, the sign in the form of gestures is given as an input to the system. Further
various steps are performed on the input sign image. Firstly segmentation phase is performed based
on the skin color so as to detect the shape of the sign. The detected region is then transformed into
binary image. Later, the Euclidean distance transformation is applied on the obtained binary image.
Row and column projection is applied on the distance transformed image. For feature extraction
central moments along with HU’s moments are used. For classification, neural network and SVM
are used.

2.10 Paper 10:Sign Language Production: A Review


Author:Razieh Rastgoo, Kourosh Kiani1, Sergio Escalera, Mohammad Sabokrou
ABSTRACT:Sign Language is the dominant yet non-primary form of communication language
used in the deaf and hearingimpaired community. To make an easy and mutual communication
between the hearing-impaired and the hearing communities, building a robust system capable of
translating the spoken language into sign language and vice versa is fundamental. To this end, sign
language recognition and production are two necessary parts for making such a two-way system.
Sign language recognition and production need to cope with some critical challenges. In this survey,
we review recent advances in Sign Language Production (SLP) and related areas using deep learning.
This survey aims to briefly summarize recent achievements in SLP, discussing their advantages,
limitations, and future directions of research.

2.11 Paper 11:Natural Sign Languages


Author:Wendy Sandler,Diane Lillo-Martin
ABSTRACT:It has been nearly forty years since serious investigation of natural sign languages

7
CHAPTER 2. LITERATURE SURVEY

began to show that these languages are bona fide linguistic systems, with structures and rules and
the full range of expressive power that characterize spoken languages. Researchers have spent
most of that time demonstrating, with increasing rigor and formality, the sometimes surprising
similarities between languages in the two modalities, spoken and signed. Concomitantly, scholars in
the related disciplines of language acquisition and neurolinguistics have been discovering significant
similarities between spoken and signed languages in these domains as well. It is safe to say that the
academic world is now convinced that sign languages are real languages in every sense of the term.
If this were the whole story, however, there would be no need for a chapter on sign languages in
this volume. Each sign language would be seen as a language like any other, English, Hungarian,
Central Alaskan Yupik Eskimo, or Mandarin Chinese, each with its own contribution to make
toward understanding the general language faculty of humans. But this is not the whole story.

2.12 Paper 12:Regional-CNN-based enhanced Turkish sign language

recognition
Author:Tolga Yirtici1 · Kamil Yurtkan
ABSTRACT:Hearing impaired people use sign language for communication. Considering the
spoken languages, the most widely spread ones have their own sign languages. Thus, sign languages
are spoken language dependent. In this paper, the focus is on Turkish Sign Language (TSL). In
parallel with the developments in computer science and artificial intelligence, the human computer
interaction is now possible and it is a definite field of computer science. Automatically detecting
and recognizing sign languages by computers are also possible applications. The improvements in
machine learning field in the last decade made it possible to improve systems that can automatically
detect and recognize sign language from still images. On the other hand, using a robust algorithm
that is recognizing a sign language may not be applicable to another language as signs and linguistic
properties may differ between languages. In this paper, we propose a novel method of recognizing
characters of Turkish Sign Language (TSL). The method is tested on captured images containing
signs, which are extracted from video files. Alexnet is employed as a pre-trained network in
this system. Region-based Convolutional Neural Network (R-CNN) object detector is used to
train the new model. The knowledge of the neural network is transferred using transfer learning
method and tuned for recognizing TSL. The system achieves 99.7conventional methods applied
on TSL. Furthermore, the study is significant since it is one of the first successful application of
transfer learning approach to TSL. The system is open to further improvements by improving the

8
CHAPTER 2. LITERATURE SURVEY

representations of the sign images.

2.13 Paper 13:Recent developments in visual sign language recognition


Author:Ulrich von Agris Æ Jo¨rg Zieren Æ Ulrich Canzler
ABSTRACT:Research in the field of sign language recognition has made significant advances in
recent years. The present achievements provide the basis for future applications with the objective of
supporting the integration of deaf people into the hearing society. Translation systems, for example,
could facilitate communication between deaf and hearing people in public situations. Further
applications, such as user interfaces and automatic indexing of signed videos, become feasible. The
current state in sign language recognition is roughly 30 years behind speech recognition, which
corresponds to the gradual transition from isolated to continuous recognition for small vocabulary
tasks. Research efforts were mainly focused on robust feature extraction or statistical modeling of
signs. However, current recognition systems are still designed for signer-dependent operation under
laboratory conditions. This paper describes a comprehensive concept for robust visual sign language
recognition, which represents the recent developments in this field. The proposed recognition
system aims for signer-independent operation and utilizes a single video camera for data acquisition
to ensure userfriendliness

2.14 Paper 14:Deep learning-based sign language recognition system for

static signs
Author:Ankita Wadhawan • Parteek Kumar
ABSTRACT:Sign language for communication is efficacious for humans, and vital research is in
progress in computer vision systems. The earliest work in Indian Sign Language (ISL) recognition
considers the recognition of significant differentiable hand signs and therefore often selecting a
few signs from the ISL for recognition. This paper deals with robust modeling of static signs in
the context of sign language recognition using deep learning-based convolutional neural networks
(CNN). In this research, total 35,000 sign images of 100 static signs are collected from different
users. The efficiency of the proposed system is evaluated on approximately 50 CNN models. The
results are also evaluated on the basis of different optimizers, and it has been observed that the
proposed approach has achieved the highest training accuracy of 99.72colored and grayscale images,
respectively. The performance of the proposed system has also been evaluated on the basis of
precision, recall and F-score. The system also demonstrates its effectiveness over the earlier works
in which only a few hand signs are considered for recognition.

9
CHAPTER 2. LITERATURE SURVEY

2.15 Paper 15:American sign language recognition technique using 3-D

convolutions neural networks


Author:Shikhar Sharma · Krishan Kumar
ABSTRACT:The communication between a person from the impaired community with a person
who does not understand sign language could be a tedious task. Sign language is the art of conveying
messages using hand gestures. Recognition of dynamic hand gestures in American Sign Language
(ASL) became a very important challenge that is still unresolved. In order to resolve the challenges
of dynamic ASL recognition, a more advanced successor of the Convolutional Neural Networks
(CNNs) called 3-D CNNs is employed, which can recognize the patterns in volumetric data like
videos. The CNN is trained for classification of 100 words on Boston ASL (Lexicon Video Dataset)
LVD dataset with more than 3300 English words signed by 6 different signers. 70the remaining
30the existing state-of-art models in terms of precision (3.7(3.9proposal may be used in real-time
applications

2.16 Paper 16:Sign Language Translation Using Deep Convolutional Neural

Networks
Author:Rahib H.Abiyev1, Murat Arslan1 and John Bush Idoko1
ABSTRACT:Sign language is a natural, visually oriented and non-verbal communication channel
between people that facilitates communication through facial/bodily expressions, postures and a set
of gestures. It is basically used for communication with people who are deaf or hard of hearing.
In order to understand such communication quickly and accurately, the design of a successful
sign language translation system is considered in this paper. The proposed system includes object
detection and classification stages. Firstly, Single Shot Multi Box Detection (SSD) architecture is
utilized for hand detection, then a deep learning structure based on the Inception v3 plus Support
Vector Machine (SVM) that combines feature extraction and classification stages is proposed to
constructively translate the detected hand gestures. A sign language fingerspelling dataset is used
for the design of the proposed model. The obtained results and comparative analysis demonstrate
the efficiency of using the proposed hybrid structure in sign language translation.

10
CHAPTER 2. LITERATURE SURVEY

2.17 Paper 17:SIGN LANGUAGE RECOGNITION USING 3D

CONVOLUTIONAL NEURAL NETWORKS


Author:Jie Huang, Wengang Zhou, Houqiang Li, and Weiping Li
ABSTRACT:Sign Language Recognition (SLR) targets on interpreting the sign language into text
or speech, so as to facilitate the communication between deaf-mute people and ordinary people.
This task has broad social impact, but is still very challenging due to the complexity and large
variations in hand actions. Existing methods for SLR use hand-crafted features to describe sign
language motion and build classification models based on those features. However, it is difficult to
design reliable features to adapt to the large variations of hand gestures. To approach this problem,
we propose a novel 3D convolutional neural network (CNN) which extracts discriminative spatial-
temporal features from raw video stream automatically without any prior knowledge, avoiding
designing features. To boost the performance, multi-channels of video streams, including color
information, depth clue, and body joint positions, are used as input to the 3D CNN in order to
integrate color, depth and trajectory information. We validate the proposed model on a real dataset
collected with Microsoft Kinect and demonstrate its effectiveness over the traditional approaches
based on hand-crafted features.

2.18 Paper 18:Simulation and Analysis of Hand Gesture Recognition for

Indian Sign Language using CNN


Author:Satish Kumar Alaria, Ashish Raj, Vivek Sharma, Vijay Kumar
ABSTRACT:Sign Language Recognition is a device or program to help deaf and mute people.
However, communication has always been difficult for a person with verbal and physical disabilities.
Sign language recognition communication between the average person and the disabled using this
device easily communicates with people who cannot communicate with the average person, this
program reduces the communication gap between people. In total, the world has a population of
about 15 -20Awareness Program. Different methods are used to identify sign language but they are
not effective due to the economic and commercial situation so we use this cheap and affordable
method for people. Therefore, sign language recognition systems based on image processing
and sensory networks are preferred over gadget programs as they are more accurate and easier to
implement. This paper aims to create an easy-to-use and accurate sign language recognition system
trained in the neural network thus producing text and speech input.

11
CHAPTER 2. LITERATURE SURVEY

2.19 Paper 19:Deep convolutional neural network for hand sign language

recognition using model E


Author:Yohanssen Pratama, Ester Marbun, Yonatan Parapat, Anastasya Manullang
ABSTRACT:An image processing system that based computer vision has received many attentions
from science and technology expert. Research on image processing is needed in the development
of human-computer interactions such as hand recognition or gesture recognition for people with
hearing impairments and deaf people. In this research we try to collect the hand gesture data and
used a simple deep neural network architecture that we called model E to recognize the actual hand
gestured. The dataset that we used is collected from kaggle.com and in the form of ASL (American
Sign Language) datasets. We doing accuracy comparison with another existing model such as
AlexNet to see how robust our model. We find that by adjusting kernel size and number of epoch
for each model also give a different result. After comparing with AlexNet model we find that our
model E is perform better with 96.82

2.20 Paper 20:Sign Language Recognition


Author:Satwik Ram Kodandaram,N Pavan Kumar , Sunil G L
ABSTRACT:Sign Language is mainly used by deaf (hard hearing) and dumbpeople to exchange
information between their own community and with other people. It is a language where people use
theirhand gestures to communicate as they can’t speak or hear. Sign Language Recognition (SLR)
deals with recognizing the hand gestures acquisition and continues till text or speech is generated
for corresponding hand gestures. Here hand gestures for sign language can be classified as static and
dynamic.However, static hand gesture recognition is simpler than dynamic hand gesture recognition,
but both recognition is important to the human community.We can use Deep Learning Computer
Vision to recognize the hand gestures by building Deep Neural Network architectures (Convolution
Neural Network Architectures) where the model will learn to recognize the hand gestures images
over an epoch.Once the model Successfully recognizes the gesture the corresponding English text
is generated and then text can be converted to speech.This model will be more efficient and hence
communicate for the deaf (hard hearing) and dump people will beeasier. In this paper, we will
discuss how Sign Language Recognition is done using Deep Learning.

12
Chapter 3

Analysis / Software Requirements


Specification (SRS)

3.1 INTRODUCTION
Sign language detection using AI/ML(Artificial intelligence/ Machine Learning) system aims to
assist deaf and dumb people to maintain communication it converts the sign language into subtitles
which makes it easier for any user to communicate This document outlines the requirements for the
development of this software system

3.2 PURPOSE
The purpose of this software to provide a platform for normal people able to communicate with deaf
and dumb people using advanced AI and ML algorithms . by analysing the subject signs played on
video it detects the sings and convert it into subtitles which makes it easier for users to understand
the system will classify and provide information about the sign which helps in aiding in timely and
effective management practices

3.3 PRODUCT SCOPE


The applications can further be extended to security purposes, by developing a sign language of
your own. And even observing and analyzing any suspicious actions.
Some other applications and scopes of this project are:
” It can be used to provide live captions for the online meetings.
” It can be used to detect mistakes in sign languages.
” It can be used for learning and practicing sign languages.

13
CHAPTER 3. ANALYSIS / SOFTWARE REQUIREMENTS SPECIFICATION (SRS)

” Text generated from this application can be converted to speech for better communication. ”
Use hand gestures to control and automate other devices.

3.4 DOCUMENT CONVENTIONS


FEATURE: Features are individual measurable property or characteristic of a phenomenon being
observed. These required for action recognition.
” LABEL: Labels are the final output. We can also consider the output classes to be the labels.
” MODEL: A machine learning model is a mathematical portrayal of a real-life problem. There
are various algorithms that perform different tasks with different levels of accuracy.
” CLASSIFICATION: In classification, we will need to categorize data into a finite number of
predefined classes.
” LSTM: Long Short-Term Memory is a kind of recurrent neural network (RNN) which can
retain the information for a long period of time. It is used for processing, predicting, and classifying
based on time-series data.
” TRAINING SET: This is the data set over which LSTM model is trained. The predictions are
completely dependent on the training-data set.
” TESTING SET: The test dataset is a subset of the training dataset that is utilized to give an
objective evaluation of a final model.
TENSOR FLOW: TensorFlow is an open-source end-to-end platform for creating Machine
Learning applications. It is a symbolic math library that uses dataflow and differentiable
programming to perform various tasks focused on training and inference of deep neural networks.
” OPEN CV: Open CV(Open Source Computer Vision) is an open source library of
programming functions used for real-time computer-vision. It is mainly used for image processing,
video capture and analysis for features like face and object recognition.

3.5 USER INTERFACE


WEB INTERFACE:There will be an output screen where the video stream used for processing
will be displayed and on the bottom side of the video display window the predicted words will be
displayed.
OUTPUT:There will be three words displayed which will be arranged in order of high to low
possibility in a left to right manner. Word with highest possibility will be highlighted using a
coloured outline.

14
CHAPTER 3. ANALYSIS / SOFTWARE REQUIREMENTS SPECIFICATION (SRS)

3.6 HARDWARE INTERFACE


CAMERA SYSTEM: Integration with cameras or imaging devices for capturing high-quality
video
SENSORS:Integration with sensors for recognising the sign language symbol detected on the
screen then translating into text with most possible outcome

3.7 HARDWARE REQUIREMENTS


Processor : 2.5 gigahertz (GHz) frequency or above.
RAM: A minimum of 4 GB of RAM.
Hard disk : A minimum of 20 GB of available space.
Input Device : High resolution camera
Monitor : Minimum Resolution 1024 X 768.

3.8 SOFTWARE REQUIREMENTS


Operating System : Windows 10 and above.
Programming language : Python 2.7 and above.
Platform : Jet Brains PyCharm Community Edition 2018.3.5 x64
Supporting libraries : TensorFlow, Open CV, Media pipe etc.

15
Chapter 4

System Design

4.1 INTRODUCTION TO SYSTEM DESIGN


In a world striving for inclusively, accessibility to communication is paramount. Sign language
serves as a primary mode of communication for the deaf and hard of hearing community. However,
recognizing sign language poses unique challenges in technology. This necessitates the design of
a Sign Language Detector System (SLDS) capable of interpreting and translating sign language
gestures into text or spoken language.

4.2 SYSTEM ARCHITECTURE


Convolutional Neural Network (CNN): Employ a CNN-based architecture renowned for its
effectiveness in image classification tasks. Consider using pre-trained models like ResNet, Inception,
or MobileNet due to their proven performance and efficiency.
The SLDS architecture comprises several interconnected components:
1. Input Module: Captures sign language gestures through devices like cameras or motion
sensors.
2. Preprocessing Module: Cleans and enhances the input data, removing noise and irrelevant
information.
3. Feature Extraction Module: Extracts relevant features from the preprocessed data, such as
hand movements, gestures, and facial expressions.
4. Classification Module: Utilizes machine learning algorithms or deep neural networks to
classify extracted features into corresponding sign language symbols or words.
5. Translation Module: Translates classified symbols or words into text or spoken language.
6. Output Module: Presents the translated output in a user-friendly format.

16
CHAPTER 4. SYSTEM DESIGN

4.3 FUNCTIONAL COMPONENTS


Functional components in the system design for encompass the various modules or services
responsible for performing specific functions within the system
1. Input Processing: Receives input data and prepares it for further analysis.
2. Feature Extraction: Identifies and extracts relevant features from the input data.
3. Classification and Recognition: Classifies extracted features into corresponding sign
language symbols or words.
4. Translation: Translates recognized symbols or words into understandable text or speech.
5. Output Presentation: Presents the translated output to the user through appropriate channels.

4.4 MONITORING AND MAINTENANCE


1. Performance Monitoring: Regularly assesses the accuracy and efficiency of the system through
performance metrics.
2. Error Handling: Implements mechanisms to detect and handle errors, ensuring robustness
and reliability.
3. Data Quality Assurance: Maintains a high standard of input data quality to improve system
performance.

4.5 TESTING AND QUALITY ASSURANCE


1. Unit Testing: Validates the functionality of individual modules through isolated testing.
2. Integration Testing: Verifies the interaction and compatibility of interconnected modules.
3. User Acceptance Testing: Involves end-users to evaluate the system’s usability, accuracy,
and effectiveness.
4. Performance Testing: Assesses the system’s responsiveness, scalability, and resource
utilization under varying conditions.
5. Security Testing: Identifies and mitigates potential security vulnerabilities to protect user
privacy and data integrity.

17
CHAPTER 4. SYSTEM DESIGN

Figure 4.1: USE CASE DIAGRAM

18
CHAPTER 4. SYSTEM DESIGN

Figure 4.2: DATA FLOW DIAGRAM

19
CHAPTER 4. SYSTEM DESIGN

Figure 4.3: STATE DIAGRAM

20
Chapter 5

Methodology

5.1 REASEARCH ANALYSIS


Research analysis involves the systematic examination and interpretation of data collected through
research methods to draw meaningful conclusions and insights. This process typically includes
organizing and summarizing data, identifying patterns or trends, and evaluating the significance of
findings in relation to the research objectives or hypotheses. Researchers employ various analytical
techniques, such as statistical analysis, qualitative coding

5.2 REQUIREMENT GATHERING


User Research:
Engage with stakeholders, including the deaf and hard of hearing community, sign language
interpreters, and potential users.
User Interviews And Surveys: Identify functional requirements such as gesture recognition
accuracy, translation speed, and user interface accessibility.
Domain Analysis: - Gather non-functional requirements including system reliability, security,
and scalability. - Document user stories, use cases, and system specifications to guide development.

5.3 DESIGN AND PLANNING


Design and planning involve crafting a robust system architecture, intuitive user interface, and
effective machine learning algorithms, alongside infrastructure planning, project timelines, and risk
mitigation strategies to ensure successful implementation
System Architecture Design:Develop a high-level system architecture that outlines the
components, modules, and interactions within the system
Database Design:Design the database schema to efficiently store and retrieve Sign language

21
CHAPTER 5. METHODOLOGY

images, classification results, and user feedback.

5.4 DEVELOPMENT
-Design system architecture, considering input sources, preprocessing techniques, feature extraction
methods, classification algorithms, translation mechanisms, and output formats.
Our project aims to develop a robust system for identifying sign language signs then turning into
text using advanced AI/ML we seek to provide users and experts with a reliable tool to understand
deaf and dumb people efficiently.

5.5 TESTING AND DEPLOYMENT


Conduct unit tests for individual modules to verify their functionality and correctness.
Testing
Validation Dataset:Description of the dataset used for testing, including the number of samples,
distribution of classes, and sources of data.
Evaluation Metrics:Explanation of the metrics used to evaluate the performance of the disease
identification system, such as accuracy, precision, recall, F1-score, and confusion matrices.
Deployment
Infrastructure: Description of the infrastructure used to deploy the disease identification
system,including cloud services, on-premises servers, or edge devices.
Deployment Strategy:Explanation of the deployment strategy employed, such as
containerization using Docker or deployment through server less architectures

5.6 DEPLOYMENT PLANNING


Our deployment planning for the Sign Language Detector focuses on ensuring a smooth and reliable
transition from development to production
1.Infrastructure assessment and Data Management
2.Deployment Strategy-preparation,testing,training support,monitoring and optimization

22
Chapter 6

Implementation

6.1 INTRODUCTION
The Sign Language Detector aims to bridge communication barriers by interpreting sign language
gestures and translating them into text or spoken language. This implementation involves several
key stages, including data collection, data preprocessing, machine learning model selection, training,
and evaluation.

6.2 DATA COLLECTION


Data collection is the foundation for any AI/ML application,Gather a diverse dataset of sign language
gestures, including various hand movements, gestures, and facial expressions.
Here’s a detailed breakdown of the data collection process:
1.Define Scope:Determine the target signs,and sign language type you want to detect and the
desired accuracy of the model
2.Plan Logistics:select the type of sign language you want to use and train with the images on
proper image capture techniques
3.Quality Control:Implement measures to ensure the accuracy and consistency of your labled
data

6.3 DATA PREPROCESSING


Data preprocessing is crucial for preparing your images for AI/ML model training.Here is a quick
summary
1.Camera Quality:Ensure the video capturing is on high quality
2.Labelling:Assign labels to each sign languages sign with types for training supervision.

23
CHAPTER 6. IMPLEMENTATION

3.Color Features: Analyze variations in color intensity and distribution across the video area of
the subject

6.4 MACHINE LEARNING MODEL SELECTION AND TRAINING:


Model Selection
Choose an appropriate algorithm in ML for image classification.
Convolutional Neural Networks (CNNs): Particularly well-suited for image recognition tasks
due to their ability to learn features directly from the data

6.5 MODEL TRAINING


We used CNN because it gives very good accuracy on the classification tasks. We used the Conv2D
layer of CNN to detect sign language. In the first layers of our Conv2D, we set the input shape as
120x160x3 because our converted image has RGB color with a size of 120x160 pixels. We set the
filter size as 32, kernel size 4x4, and activation function as ReLu in the first Conv2D layer. After
that, we dropped 0.5 percent of neurons in the first layer using the dropout layer. In the next layer,
we added a pooling layer. In the pooling layer, we used max pooling with a pool size of 2x2. In the
next layer, we added another Conv2D layer with a filter size of 64, kernel size of 4x4, and activation
function as ReLu. Followed by the Conv2D layer we added the same dropout and pooling layer that
was already described above. After that, we flatten all the 2d layers into ld layers and add two dense
layers in the last of our model. In the last dense layer, we add 11 neurons and a soft max activation
function. We set 11 neurons in the last layers because we have only 11 labels on the dataset and the
soft max activation function is used here because our classification task is a multi-class classification
task. The soft max activation function is used when we need to classify more than two classes.
ReLU layer: Networks those utilizes the rectifier operation for the hidden layers are cited as
rectified linear unit (ReLU). This ReLU function is a simple computation that returns the value
given as input directly if the value of input is greater than zero else returns zero. This can be
represented as mathematically using the function over the set of 0 and the input as follows: G(x) =
max0,x
Principal component analysis: Principal component analysis is an approach of machine
learning which is utilized to reduce the dimensionality. It utilizes simple operations of matrices
from statistics and linear algebra to compute a projection of source data into the similar count or
lesser dimensions. PCA can be thought of a projection approach where data with m-columns or
features are projected into a subspace by m or

24
CHAPTER 6. IMPLEMENTATION

Figure 6.1: 1

Figure 6.2: 2

25
CHAPTER 6. IMPLEMENTATION

Figure 6.3: A

Figure 6.4: B

26
Chapter 7

Conclusion

This work proposes a methodology for perceiving the classification of sign language recognition.
Sign language is the core medium of communication between deaf-mute and everyday people. It
is highly implacable in real-world scenarios like communication, human–computer interaction,
security, advanced AI, and much more. For a long time, researchers have been working in this field
to make a reliable, low cost and publicly available SRL system using different sensors, images,
videos, and many more techniques. Detecting sign language has become an important research
field to improve communication with deaf and dumb people. It is also important to understand that
different sign languages are developed in different language communities, and research on sign
language detection is also language-specific. Even though English is a mainstream language with a
large deaf and dumb community, there has been very little research conducted on sign language
detection in English. In this paper, we propose a new English sign language detection scheme that
relies on fingertip position as a training tool for a CNN classifier. Several methods have been tested
and compared against a large dataset of images. Based on test set accuracy, the proposed method
outperforms all existing methods. In addition, the proposed scheme appears to be compact and
efficient in terms of computation and size.
This Project is a working system for the conversion of Indian Sign Language to text in real-time
for deaf and mute people. For now, this project is focused on the single spelling-based signs where
each alphabet is represented by a particular hand sign. The accuracy achieved is 97.3percent on our
testing dataset. The approach we have used is that the training data has some background noise like
patterned clothes other objects in the background which might be present in real-world scenarios.
However, the distance dataset has negligible background noise adequate lighting. This way we can
train the model to ignore the background noise focus only on the hand signs and gestures, therefore
getting better results.

27
Chapter 8

Future Work

8.1 MULTIPLE SIGN LANGUAGES DETECTOR


1. The implementation of our model for other sign languages such as Indian sign language or
American sign language.
2. we try to implement further more than multiple languages to expand the possibility of
communication

8.2 ENHANCED ACCURACY


1. Further training the neural network to efficiently recognise symbols.
2. Enhancement of model to recognise expressions.

8.3 GESTURE RECOGNITION


1.Integrate gesture recognition to identify non manual components such as facial expressions and
body movements, crucial in sign language comprehension
2.developing algorithms to analyze body movements and posture to enhance the accuracy of
interpreting signs,as certain signs rely on the positioning of the body or the orientation of the signer
relative to the listener

8.4 REAL-TIME FEEDBACK


1.Implementing real time feedback mechanism to assist users in improving their sign language
skills, providing instant corrections or suggestions.
2.introduce accessibility features like text-to-sign translation or enhanced communication
between sign languages users and non signers

28
CHAPTER 8. FUTURE WORK

8.5 ETHICAL CONSIDERATION


1.Address ethical consideration such as data privacy,bias mitigation,and inclusively to ensurer ad
respectful treatment of sign language users
2.Implement robust security measures to protect user data and prevent unauthorized access or
misuse of the sign language detector,especially in applications involving sensitive communication
or personal information

8.6 ADAPTIVE ASSISTANCE


1. Implement adaptive assistance mechanisms that adjust the level of support provided by the
detectors based on the users proficiency,preference and specific communication needs,fostering a
more personalized and effective user experience
2.Foster community engagement and collaboration among sign language users, educators,
researchers, and developers to co-create and co-evaluate the sign language detection product,
ensuring that it meets the real-world needs and preferences of its target users

8.7 3D SIGN LANGUAGE AVATAR


Explore the creation of personalized 3D avatars that mimic the user’s sign language gestures in real
time. This approach involves capturing the user’s movements using depth-sensing cameras or motion
capture technology and mapping them onto a virtual avatar, which then performs the corresponding
sign language gestures. Users could customize their avatars’ appearance and style to reflect
their identity, preferences, and cultural background, making sign language communication more
engaging and expressive. Additionally, these avatars could be integrated into virtual environments
or communication platforms to facilitate remote communication and collaboration among sign
language users and non-signers. Research in this area could focus on developing realistic and
expressive avatar animations, optimizing avatar creation pipelines, and evaluating the usability and
effectiveness of avatar-mediated sign language communication in various contexts.

29
References

1. S. Akyol and P. Alvarado, “Finding Relevant Image Content for mobile Sign Language
Recognition,” Proc. IASTED Int’l Conf. Signal Processing, Pattern Recognition and
Application, pp. 48-52, 2001

2. Lin, H.-I., Hsu, M.-H., and Chen, W.-K. (2014). “Human hand gesture recognition using
a convolution neural network.” IEEE International Conference on Automation Science and
Engineering

3. Pallavi Gurjal, Kiran Kunnur, “Real Time Hand Gesture Recognition using SIFT”,
International Journal for Electronics and Engineering, 2012,pp 19-33.

4. Bahl, L., Jelinek, F., Mercer, R.: A maximum likelihood approach to continuous speech
recognition. IEEE Trans. Pattern Anal. Mach. Intell

5. Brien, David (ed.). 1992. Dictionary of British Sign Language/English. London: Faber Faber.

6. XIAOLONG TENG, BIANI WU, W. Y., AND LIU, C. A hand gesture recognition system
based on local linear embedding, April 2005.

7. . Bergstra, J., Breuleux, O., Bastien, F., Lamblin, P., Pascanu, R., Desjardins, G., Turian,
J., Warde-Farley, D., Bengio, Y.: Theano: a CPU and GPU math expression compiler. In:
Proceedings of the Python for Scientific Computing Conference (SciPy) (Jun 2010), oral
Presentation

8. Chai, X., Li, G., Lin, Y., Xu, Z., Tang, Y., Chen, X., Zhou, M.: Sign Language Recognition
and Translation with Kinect (2013), http://vipl.ict.ac.cn/sites/default/files/papers/files/2013
FG xjchai Sign Language Recognition and Translation with Kinect.pdf

9. Cire¸san, D., Meier, U., Schmidhuber, J.: Multi-column deep neural networks for image
classification. In: Computer Vision and Pattern Recognition (CVPR),

30
CHAPTER 8. FUTURE WORK

10. Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton, “Imagenet classification with deep
convolutional neural networks,

11. Andrej Karpathy, George Toderici, Sanketh Shetty, Thomas Leung, Rahul Sukthankar, and Li
Fei-Fei, “Large-scale video classification with convolutional neural networks,”

12. Yann LeCun, Leon Bottou, Yoshua Bengio, and Patrick ´ Haffner, “Gradient-based learning
applied to document recognition,”

13. Hueihan Jhuang, Thomas Serre, Lior Wolf, and Tomaso Poggio, “A biologically inspired
system for action recognition,”

14. Shuiwang Ji, Wei Xu, Ming Yang, and Kai Yu, “3D convolutional neural networks for human
action recognition,”

15. Kirsti Grobel and Marcell Assan, “Isolated sign language recognition using hidden markov
models,” in Systems, Man, and Cybernetics

16. Christian Vogler and Dimitris Metaxas, “Parallel hidden markov models for american sign
language recognition,”

17. Kouichi Murakami and Hitomi Taguchi, “Gesture recognition using recurrent neural
networks,”

18. Jong-Sung Kim, Won Jang, and Zeungnam Bien, “A dynamic gesture recognition system for
the korean sign language (ksl),”

19. Ross Girshick, Jeff Donahue, Trevor Darrell, and Jitendra Malik, “Rich feature hierarchies for
accurate object detection and semantic segmentation,”

20. M.A. Jalal, R.Chen, R.Moore, et al., “American sign language posture understanding with
deep neural networks,”

31

You might also like