Adaptive Data Boosting Technique for Robust Personalized Speech Emotion in Emotionally-Imbalanced Small-Sample Environments
Metadatos
Afficher la notice complèteAuteur
Bang, Jaehun; Hur, Taeho; Kim, Dohyeong; Huynh-The, Thien; Lee, Jongwon; Han, Yongkoo; Baños Legrán, Oresti; Kim, Jee-In; Lee, SungyoungEditorial
MDPI
Materia
Speech emotion recognition Personalization Machine learning Data selection Data augmentation
Date
2018-11-02Referencia bibliográfica
Bang, J.[et al.]. Adaptive Data Boosting Technique for Robust Personalized Speech Emotion in Emotionally-Imbalanced Small-Sample Environments. Sensors 2018, 18, 3744.
Patrocinador
This research was supported by an Institute for Information & Communications Technology Promotion (IITP) grant funded by the Korean government (MSIT) (No. 2017-0-00655). This research was supported by the MSIT (Ministry of Science and ICT), Korea, under the ITRC (Information Technology Research Center) support program (IITP-2017-0-01629) supervised by the IITP (Institute for Information & communications Technology Promotion). This research was supported by the MIST (Ministry of Science and ICT), Korea, under the National Program for Excellence in SW supervised by the IITP (Institute for Information & communications Technology Promotion) (2017-0-00093).Résumé
Personalized emotion recognition provides an individual training model for each target
user in order to mitigate the accuracy problem when using general training models collected from
multiple users. Existing personalized speech emotion recognition research has a cold-start problem
that requires a large amount of emotionally-balanced data samples from the target user when creating
the personalized training model. Such research is difficult to apply in real environments due to the
difficulty of collecting numerous target user speech data with emotionally-balanced label samples.
Therefore, we propose the Robust Personalized Emotion Recognition Framework with the Adaptive
Data Boosting Algorithm to solve the cold-start problem. The proposed framework incrementally
provides a customized training model for the target user by reinforcing the dataset by combining the
acquired target user speech with speech from other users, followed by applying SMOTE (Synthetic
Minority Over-sampling Technique)-based data augmentation. The proposed method proved
to be adaptive across a small number of target user datasets and emotionally-imbalanced data
environments through iterative experiments using the IEMOCAP (Interactive Emotional Dyadic
Motion Capture) database.