Show simple item record

AuthorMuhammad Zaidi, Syed Aun
AuthorLatif, Siddique
AuthorQadir, Junaid
Available date2025-07-08T03:58:10Z
Publication Date2024
Publication NameIEEE Open Journal of the Computer Society
ResourceScopus
Identifierhttp://dx.doi.org/10.1109/OJCS.2024.3486904
ISSN26441268
URIhttp://hdl.handle.net/10576/66082
AbstractDespite the recent progress in emotion recognition, state-of-the-art systems are unable to achieve improved performance in cross-language settings. In this article we propose a Multimodal Dual Attention Transformer (MDAT) model to improve cross-language multimodal emotion recognition. Our model utilises pre-trained models for multimodal feature extraction and is equipped with dual attention mechanisms including graph attention and co-attention to capture complex dependencies across different modalities and languages to achieve improved cross-language multimodal emotion recognition. In addition, our model also exploits a transformer encoder layer for high-level feature representation to improve emotion classification accuracy. This novel construct preserves modality-specific emotional information while enhancing cross-modality and cross-language feature generalisation, resulting in improved performance with minimal target language data. We assess our model's performance on four publicly available emotion recognition datasets and establish its superior effectiveness compared to recent approaches and baseline models.
SponsorFunding text 1: This work was supported in part by Qatar University High Impact Internal under Grant QUHI-CENG23/24-127, and in part by Qatar National Library. The statements made herein are solely the responsibility of the authors. Open Access publication supported by Qatar National Library.; Funding text 2: The authors would like to acknowledge support from Qatar University High Impact Internal Grant (QUHI-CENG23/24-127). Open access funding is provided by Qatar National Library. The statements made herein are solely the responsibility of the authors.
Languageen
PublisherIEEE
SubjectCo-attention networks
graph attention networks
multi-modal learning
multimodal emotion recognition
TitleEnhancing Cross-Language Multimodal Emotion Recognition With Dual Attention Transformers
TypeArticle
Pagination684-693
Volume Number5
dc.accessType Open Access


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record