Sign Language Recognition With Self-Learning Fusion Model

Sign language recognition (SLR) is the task of recognizing human actions that represent the language, which is not only helpful for deaf-mute people but also a means for human-computer interaction. Although data from wearable sensors have been proven useful for this task, it is still difficult to co...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE sensors journal 2023-11, Vol.23 (22), p.27828-27840
Hauptverfasser: Vu, Hoai-Nam, Hoang, Trung, Tran, Cong, Pham, Cuong
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Sign language recognition (SLR) is the task of recognizing human actions that represent the language, which is not only helpful for deaf-mute people but also a means for human-computer interaction. Although data from wearable sensors have been proven useful for this task, it is still difficult to collect such data for training deep fusion models. In this study, our contributions are twofold: 1) we collect and release a dataset for SLR consisting of both video and sensor data obtained from wearable devices and 2) we propose the first self-learning fusion model for SLR, termed STSLR, that utilizes a portion of annotated data to simulate sensor embedding vectors. By virtue of the simulated sensor features, the video features from video-only data are enhanced to allow the fusion model to recognize the annotated actions more effectively. We empirically demonstrate the superiority of STSLR over competitive benchmarks on our newly released dataset and well-known publicly available ones.
ISSN:1530-437X
1558-1748
DOI:10.1109/JSEN.2023.3314728