Methodologies For Sign Language Recognition A Survey
Methodologies For Sign Language Recognition A Survey
ISSN No:-2456-2165
Abstract:- Interpreting deaf-mute people has always mainly based on fingerspelling systems that transferred
been a problem for people as they primarily rely on sign spoken language to a sign language and vice versa. Sign
language for communicating. Active participation of the languages have evolved since then to develop more complex
deaf-mute community still remains at an elementary relations with the languages spoken in the land. Hence
stage, despite multiple nations providing resources for developing multiple dialects and variations from country to
the same, like a sign language interpreter and country.
communicator of news in the country of New Zealand.
Perturbing situations such as kidnapping, deception, fire Some notes about Sign Language that readers must be
breakout or any other situations of general agony could aware of are :
further exacerbate this barrier of communication, as the a.) Sign languages have an equally vibrant vocabulary as
mute people try their best to communicate, but the spoken languages and exhibit all fundamental structures that
majority remains oblivious to their language. Therefore, exist in all spoken languages.
bridging the gap between these two worlds is of utmost b.) Just like in spoken languages words do not have any
necessity. This paper aims to briefly acquaint the reader onomatopoeic relation with the referent they are describing,
with how sign language communication works and puts sign languages do not have any visual relevance with what
forward research conducted in this field that explains they convey.
how to capture and recognize sign language and also c.) Just like spoken languages use grammar to turn words
attempts to suggest a systemized solution. into meaningful sentences, Sign languages have semantics
that organize elementary meaningless units into meaningful
Keywords:- Hilbert Curve, Support Vector Machines, units/phrases.
Random Forests, Artificial Neural Network, Feed-forward d.) Unlike spoken languages, sign languages convey
Backpropagation, Hough Transform, Convolutional Neural meaning by simultaneous meanings by the main articulators
Networks, Stacked Deionized Decoders, Multilayer i.e. the head and the hands.
Perceptron Neural Network, Adaline Neural Network.
Given the significant percentage of people that rely
I. INTRODUCTION upon Sign Languages to be their primary mode of
communication, it is imperative for the wider public outside
Using statistics of the World Health organization, there the deaf-mute communities to be aware of sign languages to
are 466 Million hearing disabled people and a million some extent at least. However, normal individuals have
people that are speech impaired. This rounds up to over 5% very less incentive to learn even basic Sign language. For
of the World’s Population that cannot be communicated eg. In India there are only 250 certified sign language
with by using conventional speech based approaches. interpreters that translate for a community of up to 7 million
people. The current situation creates an overwhelmingly
Sign languages have been the most widespread method exclusionary society for the deaf-mute community . Given
of communicating with members of the deaf-mute the increasing impetus of communication skills in the
community throughout history, even being mentioned by workforce, the deaf mute community are presented with a
Socrates in Plato’s Cratylus. very high barrier of entry to participate in society as a fully
functioning member. These people are dependent on their
Multiple books and scholarly articles were written people close to them that have taken the effort to be able to
from the 16th to 18th century in European countries regarding understand and converse with them to be able to interact
instructions on how to communicate with and teach deaf- with society.
mute people. These books formed the basis for multiple sign
languages ,like the British Sign Language (BSL) and the It has been increasingly evident that a technological
French Sign Language (FSL) , American Sign Language solution is needed to bridge the communication gap that
(ASL) (based on the FSL),New Zealand Sign language and exists between the members of the speech impaired
the Sign Language used in Spain and Mexico. However, community and society as this community is most
until the 19th century most of these sign languages were
Camera-based image or video capturing has been one Brain-Computer Interfacing is an advanced approach
of the widest implemented and effective methods used in to identifying sign language. Electroencephalogram [6]
sign language interpretation systems. Using this technique, brain activities are obtained for the recognition of sign-
researchers have been successful in interpreting sign language. This approach goes one-step further by
language by capturing gestures of only one hand, both hands completely eliminating the need of any physical movements
and static or dynamic images. The signs could therefore be for detection of sign language. Here, brain waves are made
either isolated signs or continuous signs. In case of videos, use of, which are then directly transmitted to a computer
they are first captured and broken down into frames of with the help of Bluetooth. Other techniques like functional
images that can then be passed onto the system for further magnetic resonance imaging i.e. fRMI [7] and
analysis and interpretation. Hence, overall a stream of Electrocorticography [8] are also used in a similar fashion.
images is passed to the system, after which different They face a major problem of implementation complexity
techniques as per application are utilized to obtain results. and still rely on using devices connected to the head to
detect signals.
Using Kinect is another approach that has started
receiving recognition from the research community. III. LITERATURE SURVEY
Microsoft Kinect is a motion camera device that captures
users’ movements in real-time. Kinect has been primarily M. Qutaisht et al developed a system [9] for
used for gaming purposes in the recent past. It [1] provides a automatically translating the static gestures in the American
significant advantage over the camera-based approach as it Sign language(ASL). To facilitate natural interaction with
is not restricted to 2D image/video capturing, but can also the system, they performed recognition on hand images by
capture depth information such as color depth, etc. using neural networks and Hough transform. The vector
effectively. However, the maintenance and overall costs representation of the image was compared with the training
pose a higher overhead than the camera method and is hence set. Transformations such as shearing, rotation, scaling, etc.
not so commonly adopted for commercial purposes. helped by adding small noise to the model and made it
robust for the variations inherent in real-life input and also
The armband is a technique that depends on fostered flexibility. The system was implemented and tested
Electromyography (EMG) Signals. These signals are against 300 samples of hand gestures with 15 images for
generated in our muscles whenever there is any movement. each sign and an accuracy of 92.33% was achieved.
The data is collected [2] from the signers arm through a
band in the form of signals and then processed to interpret Hardik Rewari et al worked on directly processing
sign language. One of the greatest advantages this method video input to generate the relevant audio output. Their sign
assures over camera- and Kinect-based methods is zero language interpreter [10] essentially worked on the Indian
dependency on light. However, to detect signals effectively, Sign Language (ISL) to aid the deaf and dumb Indian
a lot of wires need to be connected to the band and the people. They harnessed the hardware ability of the system
portability is also an issue and proves to be a setback over by using various components like MPU6050, flex sensors,
the former two approaches. HC-05, etc. and worked on 90 words from the ISL.
A glove can also be used which primarily relies on the Microsoft Kinect was used by Rajaganapathy. S et al
path-breaking innovation [3] done in 1993 called a Cyber [11], who relied on motion capture and gesture
Glove. For getting data, the signers wear this glove which interpretation to recognize sign language and subsequently
comes with a number of sensors for each finger attached to converted it to audio. The device captured 20 human joins
it. A motion tracker [4] is also employed along with the and gestures. The device kept a track of the human gestures
glove to track the orientation and position of hands which is and eventually the data was matched with the user defined
then connected to a computer via serial ports. It provides an gestures to yield an outcome. The range of motion, which
easy way of detecting sign language, however, a lot of they could identify, was from 40 centimeters to 4 meters and
equipment needs to be appropriately set and configured for gestures of maximum 2 people at a time could be identified.
use. This is not feasible in real-world situations such as on Accuracy of upto 90% was achieved in this process.
roads, ships, shopping centers, etc. Moreover, it is also
unable to capture facial features and symbols which can be Sarbjeet Kaur et al provided a solution [12] to interpret
easily done in camera-based systems. the Indian Sign Language(ISL) which involves alphabet
Md. Mohiminul Islam et al proposed a real time hand The head subpart will be further categorized into pose
gesture recognition system [15] that worked on the and movements as well as facial expressions. Postures and
American Sign Language Dataset. It achieves higher gestures will be extracted from the movement of the hands.
accuracy by using a novel approach in the feature extraction All of the data will then be matched against the WLASL
step which includes combining K Curvature and Convex Dataset which would then be used for classification
Hull Algorithms allowing for better detection of fingertips purposes. The classification will result in generation of
in sign language gestures. This allows their Artificial Neural words.
network to recognize 37 signs of the ASL with a 94.32%
accuracy.