American Sign Language Detection System
American Sign Language Detection System
DETECTION SYSTEM
Machine Learning Internship
1. Introduction
American Sign Language (ASL) is a widely used visual language that enables communication among
individuals with hearing impairments. However, understanding ASL can be challenging for those
unfamiliar with the language. With advancements in deep learning and computer vision, real-time ASL
recognition has become feasible, providing a bridge between the deaf and hearing communities. This
project presents an ASL Detection System that recognizes hand gestures representing ASL alphabets
using TensorFlow.js and a webcam.
2. Problem Statement
Communication barriers exist between people who are deaf or hard of hearing and those who do not
understand ASL. Traditional methods such as text-based communication or interpreters may not
always be available. The need for an automated, real-time ASL detection system arises to assist in
recognizing and translating ASL signs into readable text.
3. Objectives
• To develop a real-time ASL recognition system using TensorFlow.js and Handpose models.
• To accurately classify hand gestures into 29 categories (26 letters A-Z and 3 additional signs:
SPACE, DELETE, NOTHING).
• To integrate a webcam-based gesture recognition system that detects and translates ASL
signs.
• To enhance accessibility and bridge communication gaps between ASL users and non-ASL
users.
4. Outcomes
• Accurately recognize ASL signs and translate them into corresponding letters.
• Offer an interactive and engaging user experience using React.js and Chakra UI.
5. The Dataset
• Each class folder contains multiple images of hand gestures representing the respective ASL
sign.
• The model is trained to detect hand positions and finger shapes to classify signs accurately.
6. Implementation
• The system uses Handpose, a pre-trained model from TensorFlow.js, to detect hand
landmarks. The Fingerpose library helps classify ASL gestures into predefined categories.
• The model processes real-time webcam input and estimates hand positions. The detected
hand landmarks are mapped to predefined sign language gestures, which are then classified
into their respective letters.
The detect() function captures webcam input, processes hand positions, and predicts gestures using
the Fingerpose library.
• Sets the detected sign to setSign(), updating the UI with the recognized letter.
6.4 Code
The front-end interface is built using React.js and Chakra UI for a clean and user-friendly design.
• The user interface is designed to provide an intuitive and seamless experience for users. The
interface includes:
• Webcam Feed: Displays the live video stream from the user’s webcam for gesture recognition.
• Camera Toggle Feature: Users can enable or disable the webcam for privacy.
• Interactive UI Elements: Built using Chakra UI to ensure a clean and responsive design.
6.6 Snapshots
7. Conclusion
The American Sign Language Detection System successfully classifies ASL hand gestures into 29
categories using deep learning models in TensorFlow.js. The system enables real-time ASL recognition
through a web-based interface, bridging the communication gap between ASL users and non-ASL
users. Future improvements could involve expanding the dataset, improving recognition accuracy,
and adding sentence-level ASL translation.
8. Future Scope
• Improved Accuracy: Fine-tune the model using more data and transfer learning techniques.
• Sentence Recognition: Extend recognition beyond single letters to detect words and phrases.
9. References