On the role of multimodal learning in the recognition of sign language

dc.contributor.authorFerreira, Pedro M.
dc.contributor.authorCardoso, Jaime S.
dc.contributor.authorRebelo, Ana
dc.date.accessioned2019-01-02T16:46:57Z
dc.date.available2019-01-02T16:46:57Z
dc.date.embargo2019-12-31
dc.date.issued2018
dc.description.abstractSign Language Recognition (SLR) has become one of the most important research areas in the field of human computer interaction. SLR systems are meant to automatically translate sign language into text or speech, in order to reduce the communicational gap between deaf and hearing people. The aim of this paper is to exploit multimodal learning techniques for an accurate SLR, making use of data provided by Kinect and Leap Motion. In this regard, single-modality approaches as well as different multimodal methods, mainly based on convolutional neural networks, are proposed. Our main contribution is a novel multimodal end-to-end neural network that explicitly models private feature representations that are specific to each modality and shared feature representations that are similar between modalities. By imposing such regularization in the learning process, the underlying idea is to increase the discriminative ability of the learned features and, hence, improve the generalization capability of the model. Experimental results demonstrate that multimodal learning yields an overall improvement in the sign recognition performance. In particular, the novel neural network architecture outperforms the current state-of-the-art methods for the SLR task.pt_PT
dc.identifier.citationFerreira, P. M., Cardoso, J. S., & Rebelo, A. (2018). On the role of multimodal learning in the recognition of sign language. Multimedia Tools and Applications, 1-22. https://doi.org/10.1007/s11042-018-6565-5. Disponível no Repositório UPT, http://hdl.handle.net/11328/2500pt_PT
dc.identifier.doihttps://doi.org/10.1007/s11042-018-6565-5pt_PT
dc.identifier.issn1573-7721
dc.identifier.urihttp://hdl.handle.net/11328/2500
dc.language.isoengpt_PT
dc.peerreviewedyespt_PT
dc.publisherSpringerpt_PT
dc.relation.publisherversionhttps://link.springer.com/article/10.1007%2Fs11042-018-6565-5pt_PT
dc.rightsembargoed accesspt_PT
dc.rights.urihttp://creativecommons.org/licenses/by/4.0/pt_PT
dc.subjectSign language recognitionpt_PT
dc.subjectMultimodal learningpt_PT
dc.subjectConvolutional neural networkspt_PT
dc.subjectKinect Leap motionpt_PT
dc.titleOn the role of multimodal learning in the recognition of sign languagept_PT
dc.typejournal articlept_PT
degois.publication.firstPage1pt_PT
degois.publication.lastPage12pt_PT
degois.publication.titleMultimedia Tools and Applicationspt_PT
dspace.entity.typePublicationen

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
2018PFerreiraMTA-converted.pdf
Size:
788.31 KB
Format:
Adobe Portable Document Format