ALGORITHM FOR SIGN LANGUAGE RECOGNITION USING CONVOLUTIONAL NEURAL NETWORK
DOI:
https://doi.org/10.18522/2311-3103-2026-1-%25pKeywords:
Sign language, convolutional neural network, machine learningAbstract
This paper addresses the problem of automatic recognition of Russian Sign Language (RSL) using computer vision and deep learning methods. The relevance of the study is driven by a steady increase in the number of people with hearing impairments: according to the World Health Organization, there are currently about 70 million deaf and hard-of-hearing individuals worldwide, and this number is projected to reach 630 million by 2035. The development of effective gesture recognition algorithms is an important direction for creating contactless human–machine interaction systems aimed at improving accessibility of information technologies and enhancing the quality of life for people with hearing disabilities. The aim of the study is to develop and experimentally validate an algorithm for real-time recognition of Russian Sign Language alphabet gestures in a video stream using a convolutional neural network. A specialized dataset was created, consisting of 430 images of hand gestures corresponding to the letters of the RSL alphabet, captured from different angles and under varying lighting conditions. The model was implemented using TensorFlow and Keras libraries, while integration with the video stream was performed using OpenCV and a marker-based hand tracking system. As a result of training and testing, the proposed model achieved a recognition accuracy of 99% on the test dataset. A comparative analysis with classical machine learning methods demonstrated the superiority of the convolutional neural network in terms of classification accuracy and robustness to external noise. The obtained results confirm the effectiveness of the proposed approach and its applicability for real-time systems intended for communication, educational, and rehabilitation applications, as well as for the development of advanced human–machine interaction interfaces.
References
1. World Health Organization. Deafness and hearing loss, 2023. Available at: https://www.who.int/news-room/fact-sheets/detail/deafness-and-hearing-loss (accessed 10 January 2026).
2. Dialekty russkogo zhestovogo yazyka [Dialects of russian sign language]. Available at: https://voginfo.ru/vog/2015/05/30/rslidioms/ (accessed 10 February 2024).
3. Cruz M., et al. Sign language recognition: A review, International Journal of Computer Vision, 2020.
4. Kim H., et al. Marker-based hand gesture recognition systems: A review, Sensors, 2022.
5. Foteinos K., et al. Visual Hand Gesture Recognition with Deep Learning: A Comprehensive Review of Methods, Datasets, Challenges and Future Research Directions, arXiv preprint arXiv:2507.04465, 2025.
6. Zhou Y., et al. CNN-LSTM architecture for sign language recognition, IEEE Access, 2021.
7. Wang L., et al. Spatio-temporal networks for dynamic hand gesture recognition, Neurocomputing, 2024.
8. Garcia F., et al. Robust CNN models with data augmentation for hand gesture recognition, Pattern Recognition, 2025.
9. Baihan A., Alutaibi A.I., Alshehri M., et al. Sign language recognition using modified deep learning net-work and hybrid optimization: a hybrid optimizer (HO) based optimized CNNSa-LSTM approach, Sci Rep., 2024, Vol. 14, Article 26111. DOI: 10.1038/s41598-024-76174-7.
10. Kumar S., et al. Vision-based markerless hand gesture recognition using deep learning, IEEE Transac-tions on Multimedia, 2022.
11. Safa Ameur, Mohamed Ali Mahjoub, Anouar Ben Khalifa. Spatial-temporal generative network based on deep long short-term memory autoencoder for hand skeleton data sequences reconstruction and recogni-tion, Engineering Applications of Artificial Intelligence, 2025, Vol. 161, Part C, Article 112289.
12. Al Mudawi N., Ansar H., Alazeb A., Aljuaid H., AlQahtani Y., Algarni A., Jalal A., Liu H. Innovative healthcare solutions: robust hand gesture recognition of daily life routines using 1D CNN, Front. Bio-eng. Biotechnol, 2024, Vol. 12, Article 1401803. DOI: 10.3389/fbioe.2024.1401803.
13. Patel R., et al. Efficient CNN models for embedded hand gesture recognition, Journal of Real-Time Image Processing, 2023.
14. Buynachev S.K., Boklag N.Yu. Osnovy programmirovaniya na yazyke Python [Fundamentals of pro-gramming in Python]. Ekaterinburg: Izd-vo Ural. un-ta, 2014, 91 p.
15. Zheron O. Prikladnoe mashinnoe obuchenie s pomoshch'yu Scikit-Learn i TensorFlow [Applied ma-chine learning with Scikit-Learn and TensorFlow]. Moscow: Dialektika, 2018, 688 p.
16. Makkini U. Python i analiz dannykh [Python and data analysis]: transl. from Engl. by A.A. Slinkina. Moscow: DMK Press, 2020, 540 p.
17. Set V. Glubokoe obuchenie: legkaya razrabotka proektov na Python [Deep learning: easy Python project development]. Saint Petersburg: Piter, 2021, 352 p.
18. Collette A. H5Py Documentation. Release Sep 06, 2019, 75 p.
19. Uchebnik «Mashinnoe zrenie» [.Textbook "Machine Vision"]. Vol. 2. Available at: https://copter-space.gitbook.io/uchebnik-mashinnoe-zrenie-tom-2 (accessed 10 April 2025).
20. Esan A., Okomba N., Adebiyi T., et al. Comparing CNN and Logistic Regression Results for Automated Detection Tasks. Available at: https://www.researchgate.net/publication/382880766 (accessed 10 No-vember 2025).








