Show simple item record

dc.contributor.authorLuna Jiménez, Cristina
dc.contributor.authorGriol Barres, David 
dc.contributor.authorCallejas Carrión, Zoraida 
dc.date.accessioned2022-03-18T08:38:33Z
dc.date.available2022-03-18T08:38:33Z
dc.date.issued2021-12-30
dc.identifier.citationLuna-Jiménez, C... [et al.]. A Proposal for Multimodal Emotion Recognition Using Aural Transformers and Action Units on RAVDESS Dataset. Appl. Sci. 2022, 12, 327. [https://doi.org/10.3390/app12010327]es_ES
dc.identifier.urihttp://hdl.handle.net/10481/73537
dc.descriptionThe work leading to these results was supported by the Spanish Ministry of Science and Innovation through the projects GOMINOLA (PID2020-118112RB-C21 and PID2020-118112RB-C22, funded by MCIN/AEI/10.13039/501100011033), CAVIAR (TEC2017-84593-C2-1-R, funded by MCIN/AEI/10.13039/501100011033/FEDER "Una manera de hacer Europa"), and AMIC-PoC (PDC2021-120846-C42, funded by MCIN/AEI/10.13039/501100011033 and by "the European Union "NextGenerationEU/PRTR"). This research also received funding from the European Union's Horizon2020 research and innovation program under grant agreement No 823907 (http://menhir-project.eu, accessed on 17 November 2021). Furthermore, R.K.'s research was supported by the Spanish Ministry of Education (FPI grant PRE2018-083225).es_ES
dc.description.abstractEmotion recognition is attracting the attention of the research community due to its multiple applications in different fields, such as medicine or autonomous driving. In this paper, we proposed an automatic emotion recognizer system that consisted of a speech emotion recognizer (SER) and a facial emotion recognizer (FER). For the SER, we evaluated a pre-trained xlsr-Wav2Vec2.0 transformer using two transfer-learning techniques: embedding extraction and fine-tuning. The best accuracy results were achieved when we fine-tuned the whole model by appending a multilayer perceptron on top of it, confirming that the training was more robust when it did not start from scratch and the previous knowledge of the network was similar to the task to adapt. Regarding the facial emotion recognizer, we extracted the Action Units of the videos and compared the performance between employing static models against sequential models. Results showed that sequential models beat static models by a narrow difference. Error analysis reported that the visual systems could improve with a detector of high-emotional load frames, which opened a new line of research to discover new ways to learn from videos. Finally, combining these two modalities with a late fusion strategy, we achieved 86.70% accuracy on the RAVDESS dataset on a subject-wise 5-CV evaluation, classifying eight emotions. Results demonstrated that these modalities carried relevant information to detect users’ emotional state and their combination allowed to improve the final system performance.es_ES
dc.description.sponsorshipSpanish Government PID2020-118112RB-C21 PID2020-118112RB-C22 MCIN/AEI/10.13039/501100011033 TEC2017-84593-C2-1-R MCIN/AEI/10.13039/501100011033/FEDER PDC2021-120846-C42es_ES
dc.description.sponsorshipEuropean Union "NextGenerationEU/PRTR")es_ES
dc.description.sponsorshipEuropean Union's Horizon2020 research and innovation program 823907es_ES
dc.description.sponsorshipGerman Research Foundation (DFG) PRE2018-083225es_ES
dc.language.isoenges_ES
dc.publisherMDPIes_ES
dc.rightsAtribución 3.0 España*
dc.rights.urihttp://creativecommons.org/licenses/by/3.0/es/*
dc.subjectAudio–visual emotion recognitiones_ES
dc.subjectHuman-computer interaction es_ES
dc.subjectComputational paralinguisticses_ES
dc.subjectxlsr-Wav2Vec2.0 transformeres_ES
dc.subjectTransformeres_ES
dc.subjectTransfer learninges_ES
dc.subjectAction Unitses_ES
dc.subjectRAVDESSes_ES
dc.subjectSpeech emotion recognitiones_ES
dc.subjectFacial emotion recognitiones_ES
dc.titleA Proposal for Multimodal Emotion Recognition Using Aural Transformers and Action Units on RAVDESS Datasetes_ES
dc.typejournal articlees_ES
dc.relation.projectIDinfo:eu-repo/grantAgreement/EC/H2020/823907es_ES
dc.rights.accessRightsopen accesses_ES
dc.identifier.doi10.3390/app12010327
dc.type.hasVersionVoRes_ES


Files in this item

[PDF]

This item appears in the following Collection(s)

Show simple item record

Atribución 3.0 España
Except where otherwise noted, this item's license is described as Atribución 3.0 España