Mostrar el registro sencillo del ítem

dc.contributor.authorBenítez-Guijarro, Antonio
dc.contributor.authorCallejas Carrión, Zoraida 
dc.contributor.authorNoguera García, Manuel 
dc.contributor.authorBenghazi Akhlaki, Kawtar 
dc.date.accessioned2020-01-17T08:56:55Z
dc.date.available2020-01-17T08:56:55Z
dc.date.issued2019
dc.identifier.citationProceedings 2019, 31, 54; doi:10.3390/proceedings2019031054es_ES
dc.identifier.urihttp://hdl.handle.net/10481/58851
dc.description.abstractDevices with oral interfaces are enabling new interesting interaction scenarios and ways of interaction in ambient intelligence settings. The use of several of such devices in the same environment opens up the possibility to compare the inputs gathered from each one of them and perform a more accurate recognition and processing of user speech. However, the combination of multiple devices presents coordination challenges, as the processing of one voice signal by different speech processing units may result in conflicting outputs and it is necessary to decide which is the most reliable source. This paper presents an approach to rank several sources of spoken input in multi-device environments in order to give preference to the input with the highest estimated quality. The voice signals received by the multiple devices are assessed in terms of their calculated acoustic quality and the reliability of the speech recognition hypotheses produced. After this assessment, each input is assigned a unique score that allows the audio sources to be ranked so as to pick the best to be processed by the system. In order to validate this approach, we have performed an evaluation using a corpus of 4608 audios recorded in a two-room intelligent environment with 24 microphones. The experimental results show that our ranking approach makes it possible to successfully orchestrate an increasing number of acoustic inputs, obtaining better recognition rates than considering a single input, both in clear and noisy settings.es_ES
dc.description.sponsorshipThis research has received funding by the project DEP2015-70980-R of the Spanish Ministry of Economy and Competitiveness (MINECO) and European Regional Development Fund (ERDF), the European Union’s Horizon 2020 research and innovation programme under grant agreement No. 823907 (‘Mental health monitoring through interactive conversations’, MENHIR Project), as well as, received inputs from the COST Action IC1303 AAPELEes_ES
dc.language.isoenges_ES
dc.publisherMDPIes_ES
dc.relationH2020/823907es_ES
dc.rightsAtribución 3.0 España*
dc.rights.urihttp://creativecommons.org/licenses/by/3.0/es/*
dc.subjectHuman-computer interaction es_ES
dc.subjectSpoken interactiones_ES
dc.subjectSpeech recognitiones_ES
dc.subjectAmbient intelligencees_ES
dc.subjectCoordination of deviceses_ES
dc.titleCoordination of Speech Recognition Devices in Intelligent Environments with Multiple Responsive Deviceses_ES
dc.typeinfo:eu-repo/semantics/articlees_ES
dc.rights.accessRightsinfo:eu-repo/semantics/openAccesses_ES
dc.identifier.doihttps://doi.org/10.3390/proceedings2019031054


Ficheros en el ítem

[PDF]

Este ítem aparece en la(s) siguiente(s) colección(ones)

Mostrar el registro sencillo del ítem

Atribución 3.0 España
Excepto si se señala otra cosa, la licencia del ítem se describe como Atribución 3.0 España