of multimodal data, including video feeds, keypoints,
and optical flow, allows for the training of a unified
visual backbone that significantly boosts recognition
performance.
Deep learning model optimization is another key
strategy. Advanced models such as the BLSTM
(Bidirectional Long Short-Term Memory) model
decompose consecutive sentences into word vectors,
thereby enhancing the recognition of continuous sign
language sentences. The fusion of attention
mechanisms with connective temporal classification
methods enables the extraction and combination of
short-term spatio-temporal features and hand
movement trajectory features. This addresses
challenges related to redundant information and
alignment issues within the spatio-temporal
dimension.
To tackle the challenge of recognizing sign
language from non-specific individuals, data
enhancement and diversification are essential.
Expanding the training dataset to include a broader
range of signers improves the system's ability to
generalize. Techniques like image generation for data
augmentation can further strengthen the model's
robustness, ensuring high accuracy in real-time
recognition scenarios.
Lastly, the introduction of prior knowledge,
including motor and linguistic a priori, into the causal
temporal recognition framework is beneficial. This
incorporation refines the robustness of feature
extraction by providing a deeper understanding of the
contextual semantics and the nuances of sign
language gestures. By integrating these strategies,
sign language recognition systems can be made more
resilient and effective in complex and varied
environments.
4 CONCLUSIONS
A significant application of computer vision and
machine learning technologies in the field of
accessible communication is the recognition of sign
language using an image-based system. The
performance of the sign language recognition system
will continue to improve with ongoing technological
advancements, providing the hearing impaired with a
more convenient and effective means of
communication. In the future, it is required to
continue in-depth research to solve the current
problems and promote the further development of
sign language recognition technology. The system
has a broad application prospect. It can not only
provide more communication opportunities for the
hearing impaired to help them communicate with
others without barriers, but also can be applied in the
field of education to help teachers and students
understand sign language better and improve the
teaching effect. In addition, the sign language image
recognition system also has potential application
value in the fields of intelligent transportation, remote
control, and virtual reality.
REFERENCES
Amin, M. S., Rizvi, S. T. H., & Hossain, M. M. 2022. A
comparative review on applications of different sensors
for sign language recognition. Journal of Imaging, 8(4),
98.
Cheok, M. J., Omar, Z., & Jaward, M. H. 2019. A review
of hand gesture and sign language recognition
techniques. International Journal of Machine Learning
and Cybernetics, 10, 131-153.
Jones, G. A., Ni, D., & Wang, W. 2021. Nothing about us
without us: Deaf education and sign language access in
China. Deafness & Education International, 23(3),
179-200.
Subburaj, S., & Murugavalli, S. 2022. Survey on sign
language recognition in context of vision-based and
deep learning. Measurement: Sensors, 23, 100385.
Wadhawan, A., & Kumar, P. 2021. Sign language
recognition systems: A decade systematic literature
review. Archives of computational methods in
engineering, 28, 785-813.
Wiley, V., & Lucas, T. 2018. Computer vision and image
processing: a paper review. International Journal of
Artificial Intelligence Research, 2(1), 29-36.
Yang, G., Ding, X., Gao, Y., et al. 2023. Continuous Sign
Language Recognition with Complex Backgrounds
Based on Attention Mechanism. Journal of Wuhan
University (Natural Science Edition), 69(1), 97-105
Yin, Y., 2017. Research on Behavior Perception
Recognition Technology and System Based on Mobile
Devices. Nanjing University. Doctoral Dissertation.
Zhang, Y., Yuan, K., & Yang, X., 2001, A Novel Data
Glove for Sign Language Recognition. Journal of
University of Science and Technology Beijing, 23(4),
379-381
Zhang, X., Cao, H., & Li, H. 2023. Neurophysiological
effects of frequency, length, phonological
neighborhood density, and iconicity on sign
recognition. NeuroReport, 34(17), 817-824.