SignSpeak – Sign language translation system for hearing impaired

Kalaiselvi, G. and Badri, N. and Karnan, C. (2025) SignSpeak – Sign language translation system for hearing impaired. International Journal of Science and Research Archive, 15 (2). pp. 921-930. ISSN 2582-8185

[thumbnail of IJSRA-2025-1523.pdf] Article PDF
IJSRA-2025-1523.pdf - Published Version
Available under License Creative Commons Attribution Non-commercial Share Alike.

Download ( 640kB)

Abstract

The deaf and hard-of-hearing community face significant challenges navigating through daily-life. Difficulties such as hearing impairments or speech disabilities often limit communication for the community which is a crucial aspect of human life. To bridge this gap, Sign Language, is a methodology that involves hand movements with face interaction that acts a medium to converse is utilized to convey the ideas that a hard-of-hearing wishes to share. However, Sign Language exists in different accents and does not follow a universal common language and is quite uncommon for an average group of people to have knowledge base regarding it. There are translators that have Sign Language understanding that interpret information to the hearing-impaired via hand signs. But they exist few in numbers along with their availability varying at times. In such situations, hearing-impaired community cannot actively participate in interactions. SignSpeak, a real-time translation system is developed that records audio input from the user and provides the transcribed sign gloss clips that interprets the user's vocabulary in sign language. Bidirectional Encoder Representation from Transformers (BERT) - a deep learning algorithm combines PyAudio to obtain audio input with OpenAI's Whisper model to generate text transcription and reorder them. It generates tokens which are then read by the FFmpeg module for sign gloss video retrieval and stitching and present a complete video output.

Item Type: Article
Official URL: https://doi.org/10.30574/ijsra.2025.15.2.1523
Uncontrolled Keywords: Real-time Speech Translation; American Sign Language (ASL); Whisper ASR; BERT Transformer; Gloss Video Synthesis; Accessibility Technology; Multimodal Communication
Depositing User: Editor IJSRA
Date Deposited: 25 Jul 2025 15:18
Related URLs:
URI: https://eprint.scholarsrepository.com/id/eprint/1923