Please use this identifier to cite or link to this item:
http://hdl.handle.net/10553/73852
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Marras, Mirko | en_US |
dc.contributor.author | Marín-Reyes, Pedro A. | en_US |
dc.contributor.author | Lorenzo Navarro, José Javier | en_US |
dc.contributor.author | Castrillón Santana, Modesto Fernando | en_US |
dc.contributor.author | Fenu, Gianni | en_US |
dc.date.accessioned | 2020-07-28T12:26:12Z | - |
dc.date.available | 2020-07-28T12:26:12Z | - |
dc.date.issued | 2019 | en_US |
dc.identifier.isbn | 978-989-758-351-3 | en_US |
dc.identifier.issn | 2184-4313 | en_US |
dc.identifier.uri | http://hdl.handle.net/10553/73852 | - |
dc.description.abstract | Intelligent technologies have pervaded our daily life, making it easier for people to complete their activities. One emerging application is involving the use of robots for assisting people in various tasks (e.g., visiting a museum). In this context, it is crucial to enable robots to correctly identify people. Existing robots often use facial information to establish the identity of a person of interest. But, the face alone may not offer enough relevant information due to variations in pose, illumination, resolution and recording distance. Other biometric modalities like the voice can improve the recognition performance in these conditions. However, the existing datasets in robotic scenarios usually do not include the audio cue and tend to suffer from one or more limitations: most of them are acquired under controlled conditions, limited in number of identities or samples per user, collected by the same recording device, and/or not freely available. In this paper, we propose AveRobot, an audio-visual dataset of 111 participants vocalizing short sentences under robot assistance scenarios. The collection took place into a three-floor building through eight different cameras with built-in microphones. The performance for face and voice re-identification and verification was evaluated on this dataset with deep learning baselines, and compared against audio-visual datasets from diverse scenarios. The results showed that AveRobot is a challenging dataset for people re-identification and verification. | en_US |
dc.language | eng | en_US |
dc.relation | Identificación Automática de Oradores en Sesiones Parlamentarias Usando Características Audiovisuales. | en_US |
dc.relation.ispartof | ICPRAM (Setúbal) | en_US |
dc.source | Proceedings of the 8th International Conference on Pattern Recognition Applications and Methods - Volume 1: ICPRAM, p. 255-265 (2019) | en_US |
dc.subject | 120304 Inteligencia artificial | en_US |
dc.subject.other | Deep learning | en_US |
dc.subject.other | Face-voice dataset | en_US |
dc.subject.other | Human-robot interaction | en_US |
dc.subject.other | People re-identification | en_US |
dc.subject.other | People verification | en_US |
dc.title | AveroBot: an audio-visual dataset for people re-identification and verification in human-robot interaction | en_US |
dc.type | info:eu-repo/semantics/conferenceobject | en_US |
dc.type | ConferenceObject | en_US |
dc.relation.conference | 8th International Conference on Pattern Recognition Applications and Methods, ICPRAM 2019 | en_US |
dc.identifier.doi | 10.5220/0007690902550265 | en_US |
dc.identifier.scopus | 2-s2.0-85064638092 | - |
dc.identifier.scopus | 22333278500 | - |
dc.identifier.scopus | 15042453800 | - |
dc.contributor.orcid | 0000-0002-2834-2067 | - |
dc.contributor.orcid | 0000-0002-8673-2725 | - |
dc.contributor.orcid | #NODATA# | - |
dc.contributor.orcid | #NODATA# | - |
dc.contributor.orcid | #NODATA# | - |
dc.description.lastpage | 265 | en_US |
dc.description.firstpage | 255 | en_US |
dc.investigacion | Ingeniería y Arquitectura | en_US |
dc.type2 | Actas de congresos | en_US |
dc.description.numberofpages | 11 | en_US |
dc.utils.revision | Sí | en_US |
dc.identifier.ulpgc | Sí | en_US |
dc.contributor.buulpgc | BU-ING | en_US |
item.grantfulltext | open | - |
item.fulltext | Con texto completo | - |
crisitem.project.principalinvestigator | Castrillón Santana, Modesto Fernando | - |
crisitem.event.eventsstartdate | 19-02-2019 | - |
crisitem.event.eventsenddate | 21-02-2019 | - |
crisitem.author.dept | GIR SIANI: Inteligencia Artificial, Robótica y Oceanografía Computacional | - |
crisitem.author.dept | IU Sistemas Inteligentes y Aplicaciones Numéricas | - |
crisitem.author.dept | Departamento de Informática y Sistemas | - |
crisitem.author.dept | GIR SIANI: Inteligencia Artificial, Robótica y Oceanografía Computacional | - |
crisitem.author.dept | IU Sistemas Inteligentes y Aplicaciones Numéricas | - |
crisitem.author.dept | Departamento de Informática y Sistemas | - |
crisitem.author.orcid | 0000-0002-2834-2067 | - |
crisitem.author.orcid | 0000-0002-8673-2725 | - |
crisitem.author.parentorg | IU Sistemas Inteligentes y Aplicaciones Numéricas | - |
crisitem.author.parentorg | IU Sistemas Inteligentes y Aplicaciones Numéricas | - |
crisitem.author.fullName | Marín Reyes, Pedro Antonio | - |
crisitem.author.fullName | Lorenzo Navarro, José Javier | - |
crisitem.author.fullName | Castrillón Santana, Modesto Fernando | - |
Appears in Collections: | Actas de congresos |
Items in accedaCRIS are protected by copyright, with all rights reserved, unless otherwise indicated.