Combining speech- based and linguistic classifiers to recognize emotion in user spoken utterances
Identificadores
URI: https://hdl.handle.net/10481/88569Metadatos
Mostrar el registro completo del ítemFecha
2019Patrocinador
Work partially supported by Projects MINECO TEC2012-37832-C02-01, CI-CYT TEC2011-28626-C02-02, CAM CONTEXTS (S2009/TIC-1485).Resumen
In this paper we propose to combine speech-based and linguistic classification in order to obtain better emotion recognition results for user spoken utterances. Usually these approaches are considered in isolation and even developed by different
communities working on emotion recognition and sentiment analysis. We propose modeling the users emotional state by means of the fusion of the outputs generated with both approaches, taking into account information that is usually neglected in the individual approaches such as the interaction context and errors, and the peculiarities of transcribed spoken utterances. The fusion approach allows to employ different recognizers and can be integrated as an additional module in the architecture of a spoken conversational agent, using the information generated as an additional input for the dialog manager to decide the next system response. We have evaluated our proposal using three emotionally-colored databases and obtained very positive results.