Afficher la notice abrégée

dc.contributor.authorLobato Martín, Javier
dc.contributor.authorPérez Córdoba, José Luis 
dc.contributor.authorGonzález López, José Andrés 
dc.date.accessioned2024-11-11T09:17:36Z
dc.date.available2024-11-11T09:17:36Z
dc.date.issued2024-11-11
dc.identifier.citationLobato Martín, J., Pérez Córdoba, J.L., Gonzalez-Lopez, J.A. (2024) Direct Speech Synthesis from Non-audible Speech Biosignals: A Comparative Study. Proc. IberSPEECH 2024, 86-90, doi: 10.21437/IberSPEECH.2024-18es_ES
dc.identifier.urihttps://hdl.handle.net/10481/96807
dc.descriptionThis work was supported by grant PID2022-141378OBC22 funded by MICIU/AEI/10.13039/501100011033 and by ERDF/EU.es_ES
dc.description.abstractThis paper presents a speech restoration system that generates audible speech from articulatory movement data captured using Permanent Magnet Articulography (PMA). Several algorithms were explored for speech synthesis, including classical unit-selection and deep neural network (DNN) methods. A database containing simultaneous PMA and speech recordings from healthy subjects was used for training and validation. The system generates either direct waveforms or acoustic parameters, which are converted to audio via a vocoder. Results show intelligible speech synthesis is feasible, with Mel-Cepstral Distortion (MCD) values between 9.41 and 12.4 dB, and Short- Time Objective Intelligibility (STOI) scores ranging from 0.32 to 0.606, with a maximum near 0.9. Unit selection and recurrent neural network (RNN) methods performed best. Informal listening tests further confirmed the effectiveness of thesees_ES
dc.description.sponsorshipMICIU/AEI/10.13039/501100011033 PID2022-141378OBC22es_ES
dc.description.sponsorshipERDF/EUes_ES
dc.language.isoenges_ES
dc.publisherInternation Speech Communication Association (ISCA)es_ES
dc.rightsAtribución-NoComercial 4.0 Internacional*
dc.rights.urihttp://creativecommons.org/licenses/by-nc/4.0/*
dc.titleDirect Speech Synthesis from Non-audible Speech Biosignals: A Comparative Studyes_ES
dc.typeconference outputes_ES
dc.rights.accessRightsopen accesses_ES
dc.identifier.doi10.21437/IberSPEECH.2024-18
dc.type.hasVersionVoRes_ES


Fichier(s) constituant ce document

[PDF]

Ce document figure dans la(les) collection(s) suivante(s)

Afficher la notice abrégée

Atribución-NoComercial 4.0 Internacional
Excepté là où spécifié autrement, la license de ce document est décrite en tant que Atribución-NoComercial 4.0 Internacional