Identificador persistente para citar o vincular este elemento:
http://hdl.handle.net/10553/110725
Título: | Improving user verification in human-robot interaction from audio or image inputs through sample quality assessment | Autores/as: | Freire Obregón, David Sebastián Rosales-Santana, Kevin Marín Reyes, Pedro Antonio Peñate Sánchez, Adrián Lorenzo Navarro, José Javier Castrillón Santana, Modesto Fernando |
Clasificación UNESCO: | 120304 Inteligencia artificial 2405 Biometría |
Palabras clave: | Biometric verification Audiovisual verification Human robot interaction |
Fecha de publicación: | 2021 | Proyectos: | ULPGC2018-08 Identificación automática de oradores en sesiones parlamentarias usando características audiovisuales. RTI2018-093337-B-I00 Re-identificación mUltimodal de participaNtes en competiciones dEpoRtivaS |
Publicación seriada: | Pattern Recognition Letters | Resumen: | In this paper, we tackle the task of improving biometric verification in the context of Human-Robot Interaction (HRI). A robot that wants to identify a specific person to provide a service can do so by either image verification or, if light conditions are not favourable, through voice verification. In our approach, we will take advantage of the possibility a robot has of recovering further data until it is sure of the identity of the person. The key contribution is that we select from both image and audio signals the parts that are of higher confidence. For images we use a system that looks at the face of each person and selects frames in which the confidence is high while keeping those frames separate in time to avoid using very similar facial appearance. For audio our approach tries to find the parts of the signal that contain a person talking, avoiding those in which noise is present by segmenting the signal. Once the parts of interest are found, each input is described with an independent deep learning architecture that obtains a descriptor for each kind of input (face/voice). We also present in this paper fusion methods that improve performance by combining the features from both face and voice, results to validate this are shown for each independent input and for the fusion methods. | URI: | http://hdl.handle.net/10553/110725 | ISSN: | 0167-8655 | DOI: | 10.1016/j.patrec.2021.06.014 | Fuente: | Pattern Recognition Letters, [ISSN 0167-8655] v. 149, p. 179-184, (September 2021) |
Colección: | Artículos |
Los elementos en ULPGC accedaCRIS están protegidos por derechos de autor con todos los derechos reservados, a menos que se indique lo contrario.