dc.contributor.author | Bang, Jaehun | |
dc.contributor.author | Hur, Taeho | |
dc.contributor.author | Kim, Dohyeong | |
dc.contributor.author | Huynh-The, Thien | |
dc.contributor.author | Lee, Jongwon | |
dc.contributor.author | Han, Yongkoo | |
dc.contributor.author | Baños Legrán, Oresti | |
dc.contributor.author | Kim, Jee-In | |
dc.contributor.author | Lee, Sungyoung | |
dc.date.accessioned | 2019-03-27T08:59:36Z | |
dc.date.available | 2019-03-27T08:59:36Z | |
dc.date.issued | 2018-11-02 | |
dc.identifier.citation | Bang, J.[et al.]. Adaptive Data Boosting Technique for Robust Personalized Speech Emotion in Emotionally-Imbalanced Small-Sample Environments. Sensors 2018, 18, 3744. | es_ES |
dc.identifier.issn | 1660-4601 | |
dc.identifier.uri | http://hdl.handle.net/10481/55222 | |
dc.description.abstract | Personalized emotion recognition provides an individual training model for each target
user in order to mitigate the accuracy problem when using general training models collected from
multiple users. Existing personalized speech emotion recognition research has a cold-start problem
that requires a large amount of emotionally-balanced data samples from the target user when creating
the personalized training model. Such research is difficult to apply in real environments due to the
difficulty of collecting numerous target user speech data with emotionally-balanced label samples.
Therefore, we propose the Robust Personalized Emotion Recognition Framework with the Adaptive
Data Boosting Algorithm to solve the cold-start problem. The proposed framework incrementally
provides a customized training model for the target user by reinforcing the dataset by combining the
acquired target user speech with speech from other users, followed by applying SMOTE (Synthetic
Minority Over-sampling Technique)-based data augmentation. The proposed method proved
to be adaptive across a small number of target user datasets and emotionally-imbalanced data
environments through iterative experiments using the IEMOCAP (Interactive Emotional Dyadic
Motion Capture) database. | es_ES |
dc.description.sponsorship | This research was supported by an Institute for Information & Communications Technology Promotion
(IITP) grant funded by the Korean government (MSIT) (No. 2017-0-00655). This research was supported by the
MSIT (Ministry of Science and ICT), Korea, under the ITRC (Information Technology Research Center) support
program (IITP-2017-0-01629) supervised by the IITP (Institute for Information & communications Technology
Promotion). This research was supported by the MIST (Ministry of Science and ICT), Korea, under the National
Program for Excellence in SW supervised by the IITP (Institute for Information & communications Technology
Promotion) (2017-0-00093). | es_ES |
dc.language.iso | eng | es_ES |
dc.publisher | MDPI | es_ES |
dc.rights | Atribución 3.0 España | * |
dc.rights.uri | http://creativecommons.org/licenses/by/3.0/es/ | * |
dc.subject | Speech emotion recognition | es_ES |
dc.subject | Personalization | es_ES |
dc.subject | Machine learning | es_ES |
dc.subject | Data selection | es_ES |
dc.subject | Data augmentation | es_ES |
dc.title | Adaptive Data Boosting Technique for Robust Personalized Speech Emotion in Emotionally-Imbalanced Small-Sample Environments | es_ES |
dc.type | journal article | es_ES |
dc.rights.accessRights | open access | es_ES |