Sign Language to Text and Voice Conversion for Speaking Impaired People Using Convolutional Neural Network

Authors

  • K.K. Sreedeve Department of Computer Science and Engineering, Dhaanish Ahmed College of Engineering, Padappai, Chennai, Tamil Nadu, India.
  • K. Jayashree Department of Computer Science and Engineering, Dhaanish Ahmed College of Engineering, Padappai, Chennai, Tamil Nadu, India.
  • Rincya Mabujan Department of Computer Science and Engineering, Dhaanish Ahmed College of Engineering, Padappai, Chennai, Tamil Nadu, India.
  • B. Sneha Department of Computer Science and Engineering, Dhaanish Ahmed College of Engineering, Padappai, Chennai, Tamil Nadu, India.
  • N. Selvam Department of Computer Science and Engineering, Dhaanish Ahmed College of Engineering, Padappai, Chennai, Tamil Nadu, India.
  • M. Mohamed Sameer Ali Department of Computer Science and Engineering, Dhaanish Ahmed College of Engineering, Chennai, Tamil Nadu, India.

Keywords:

Text-To-Speech (TTS), Convolutional Neural Networks (CNN), Sign Language Detection, Translation System (SLDTS).

Abstract

Sign language, a visual-gestural form of communication predominantly utilised by individuals with hearing impairments, has distinct problems in routine encounters with non-signers. To solve this problem with communication, we suggest a complete Sign Language Detection and Translation System (SLDTS) that uses Convolutional Neural Networks (CNN) to recognise sign language and turn it into audio. It also uses keyword recognition to make communication more efficient. The SLDTS is built on deep learning methods, notably CNN-based models, that can accurately find and recognise sign language motions. These models learn quickly how to recognise different hand shapes, movements, and gestures by being trained on massive, annotated datasets of sign language gestures. By analysing input video streams or images in real time, the CNN-based sign language detection component of the SLDTS can efficiently recognise and understand sign language motions, creating the foundation for effective communication between sign language users and non-signers. When the SLDTS sees sign language gestures, it uses text-to-speech (TTS) synthesis to turn them into spoken words. This translation procedure takes the recognised signs and turns them into text, then combines them into audio output. The SLDTS uses powerful TTS technology to make sure that the translated audio is clear and sounds natural, so that those who don't know sign language may understand what the signer is trying to say. This feature lets people with and without hearing problems talk to each other in real time, which helps them understand each other better and interact more.

Downloads

Published

2025-08-17

Most read articles by the same author(s)