Indian Journal of Science and Technology
DOI: 10.17485/IJST/v16i45.2583
Year: 2023, Volume: 16, Issue: 45, Pages: 4177-4185
Original Article
Premkumar Duraisamy1*, A Abinayasrijanani2, M Amrit Candida2, P Dinesh Babu2
1Assistant Professor (Sr. G.), Department of Computer Science and Engineering, KPR Institute of Engineering and Technology, Coimbatore, Tamil Nadu, India
2B.E Student, Department of Computer Science and Engineering, KPR Institute of Engineering and Technology, Coimbatore, Tamil Nadu, India
*Corresponding Author
Email: [email protected]
Received Date:11 October 2023, Accepted Date:30 October 2023, Published Date:05 December 2023
Objective: The goal of the proposed work is to leverage deep learning technologies to create an efficient and accurate system for transforming sign language into text and speech. People deliver their ideas, feelings, and experiences to others around them via their interactions with each other. The hand gesture plays a significant role since it reflects the user's thoughts more rapidly than other motions (head, face, eye, and body). For deaf-mute people with disabilities, this is still not the case. Sign language facilitates communication among deaf-mute individuals. An individual who is deaf-mute can communicate without the use of acoustic noises by using sign language. Methods: Convolutional neural networks (CNNs) are generally used to recognize and extract characteristics from sign language motions. These neural networks are employed to recognize and extract critical features from sign language gestures. These features are processed by natural language processing models for textual translation. Finally, neural text-to-speech (TTS) technology is used to translate the textual translations into synthesized speech, thereby bridging the communication gap for the Deaf community. To establish an inclusive and accessible communication system, this technique combines computer vision, natural language processing, and speech synthesis. Findings: The datasets used in this technique include hand gesture images, which contain different hand poses and expressions. It is used to train and assess the model. The experiment findings show an accuracy of 97.6% with a precision of 94.1%, a recall of 96.8%, and an F1-score of 95.9%. Novelty: This approach displays a cogent translation from text to speech and achieves an outstanding translation accuracy of 97.6% from sign language to text, producing a natural and understandable output.
Keywords: Sign Language Translation, Deep Learning, Convolutional Neural Networks, SequenceToSequence Models, Attention Mechanisms, Neural TextToSpeech
© 2023 Duraisamy et al. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. Published By Indian Society for Education and Environment (iSee)
Subscribe now for latest articles and news.