IJSMT Journal

International Journal of Science, Strategic Management and Technology

An International, Peer-Reviewed, Open Access Scholarly Journal Indexed in recognized academic databases · DOI via Crossref The journal adheres to established scholarly publishing, peer-review, and research ethics guidelines set by the UGC

ISSN: 3108-1762 (Online)
webp (1)

Plagiarism Passed
Peer reviewed
Open Access

QUANTUM-ENHANCED SPATIO-TEMPORAL DEEP LEARNING FRAMEWORK FOR REAL-TIME SIGN LANGUAGE TRANSLATION

AUTHORS:
Suriya Durai Murugan T
Mentor
Dr M Ramnath, Dr M Kaliappan
Affiliation
Department of Artificial Intelligence and Data Science, Ramco institute of Technology, Rajapalayam, Tamil Nadu, India
CC BY 4.0 License:
This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Abstract

Real-time sign language translation remains a significant challenge in assistive technology due to the complexity of capturing dynamic hand gestures and interpreting them accurately. This paper presents a Hybrid Quantum CNN-LSTM based Real-Time Sign Language Translation System, designed to bridge the communication gap between hearing-impaired individuals and the general population. The system integrates computer vision, deep learning, and quantum machine learning, where live video input is captured using OpenCV and hand landmarks are extracted using MediaPipe to obtain 42 three-dimensional joint coordinates. These landmarks are processed as temporal sequences and passed through a hybrid architecture combining Convolutional Neural Networks (CNN) for spatial feature extraction, Long Short-Term Memory (LSTM) for temporal modeling, and a Quantum Convolutional Neural Network (QCNN) layer implemented using PennyLane to capture complex non-linear relationships through quantum embedding and entanglement. Unlike traditional deep learning models, the proposed system employs a parallel quantum-classical framework that enhances feature representation while maintaining computational efficiency. The model is trained using augmented datasets with techniques such as temporal expansion, geometric transformations, and noise injection to improve robustness and generalization. Experimental results demonstrate an F1-score of approximately 0.93, achieving competitive performance with reduced parameter complexity and enabling real-time translation on standard computing devices. This work highlights the potential of hybrid quantum-classical models in real-time computer vision applications and provides an efficient, scalable solution for assistive communication technologies

Keywords
Article Metrics
Article Views
13
PDF Downloads
0
HOW TO CITE
APA

MLA

Chicago

Copy

T, S. D. M. (2026). Quantum-Enhanced Spatio-Temporal Deep Learning Framework for Real-Time Sign Language Translation. International Journal of Science, Strategic Management and Technology, 02(04). https://doi.org/10.55041/ijsmt.v2i4.285

T, Suriya. "Quantum-Enhanced Spatio-Temporal Deep Learning Framework for Real-Time Sign Language Translation." International Journal of Science, Strategic Management and Technology, vol. 02, no. 04, 2026, pp. . doi:https://doi.org/10.55041/ijsmt.v2i4.285.

T, Suriya. "Quantum-Enhanced Spatio-Temporal Deep Learning Framework for Real-Time Sign Language Translation." International Journal of Science, Strategic Management and Technology 02, no. 04 (2026). https://doi.org/https://doi.org/10.55041/ijsmt.v2i4.285.

References
1.Simonyan and A. Zisserman, “Two-Stream Convolutional Networks for Action Recognition in Videos,” Advances in Neural Information Processing Systems (NeurIPS), 2014.

2.Donahue et al., “Long-term Recurrent Convolutional Networks for Visual Recognition and Description,” IEEE CVPR, 2015.

3.Hochreiter and J. Schmidhuber, “Long Short-Term Memory,” Neural Computation, vol. 9, no. 8, pp. 1735–1780, 1997.

4.Koller, H. Ney, and R. Bowden, “Deep Hand: How to Train a CNN on 1 Million Hand Images When Your Data Is Continuous and Weakly Labelled,” IEEE CVPR, 2016.

5.C. Camgoz et al., “Neural Sign Language Translation,” IEEE CVPR, 2018.

6.Vaswani et al., “Attention Is All You Need,”NeurIPS, 2017.

7.Dosovitskiy et al., “An Image Is Worth 16x16 Words: Transformers for Image Recognition at Scale,” ICLR, 2021.

8.Pu et al., “Iterative Alignment Network for Continuous Sign Language Recognition,” IEEE CVPR, 2019.

9.Chen et al., “A Simple Framework for Contrastive Learning of Visual Representations,” ICML, 2020.

10.M. Schuld and F. Petruccione, Supervised Learning with Quantum Computers, Springer, 201
Ethics and Compliance
✓ All ethical standards met
This article has undergone plagiarism screening and double-blind peer review. Editorial policies have been followed. Authors retain copyright under CC BY-NC 4.0 license. The research complies with ethical standards and institutional guidelines.
Indexed In
Similar Articles
Knowledge, Attitude, and Practice Study on Pharmacovigilance Among Healthcare Professionals
string(16) "Saurabh N. Verma" Verma, S. N.
(2026)
DOI: 10.55041/ijsmt.v1i2.003
Ultrasonic Radar System
string(13) "Mashalkar Y.S" Y.S, M.et al.
(2026)
DOI: 10.55041/ijsmt.v2i3.413
Invisible Geographies: Theoretical Perspectives on Hidden Human–Environment Interactions
string(22) "Dr. Madhuchhanda Dhole" Dhole, D. M.
(2026)
DOI: 10.55041/ijsmt.v2i4.170
Scroll to Top