On the role of multimodal learning in the recognition of sign language
Data
2018
Embargo
2019-12-31
Orientador
Coorientador
Título da revista
ISSN da revista
Título do volume
Editora
Springer
Idioma
Inglês
Título Alternativo
Resumo
Sign Language Recognition (SLR) has become one of the most important research areas in the field of human computer interaction. SLR systems are meant to automatically translate sign language into text or speech, in order to reduce the communicational gap between deaf and hearing people. The aim of this paper is to exploit multimodal learning techniques for an accurate SLR, making use of data provided by Kinect and Leap Motion. In this regard, single-modality approaches as well as different multimodal methods, mainly based on convolutional neural networks, are proposed. Our main contribution is a novel multimodal end-to-end neural network that explicitly models private feature representations that are specific to each modality and shared feature representations that are similar between modalities. By imposing such regularization in the learning process, the underlying idea is to increase the discriminative ability of the learned features and, hence, improve the generalization capability of the model. Experimental results demonstrate that multimodal learning yields an overall improvement in the sign recognition performance. In particular, the novel neural network architecture outperforms the current state-of-the-art methods for the SLR task.
Palavras-chave
Sign language recognition, Multimodal learning, Convolutional neural networks, Kinect Leap motion
Tipo de Documento
Artigo
Versão da Editora
Dataset
Citação
Ferreira, P. M., Cardoso, J. S., & Rebelo, A. (2018). On the role of multimodal learning in the recognition of sign language. Multimedia Tools and Applications, 1-22. https://doi.org/10.1007/s11042-018-6565-5. Disponível no Repositório UPT, http://hdl.handle.net/11328/2500
Identificadores
TID
Designação
Tipo de Acesso
Acesso Embargado