Please use this identifier to cite or link to this item: https://accedacris.ulpgc.es/jspui/handle/10553/159788
Title: An Evaluation of a Visual Question Answering Strategy for Zero-shot Facial Expression Recognition in Still Images
Authors: Salas Cáceres, José Ignacio 
Castrillón Santana, Modesto Fernando 
Freire Obregón, David Sebastián 
Santana Jaria, Oliverio Jesús 
Hernández Sosa, José Daniel 
Lorenzo Navarro, José Javier 
UNESCO Clasification: 120304 Inteligencia artificial
Keywords: Facial Expression Recognition
Visual Language Models
Large Language Models
AffectNet
Issue Date: 2025
Conference: International Conference on Visual Communications and Image Processing (VCIP 2025) 
Abstract: Facial expression recognition (FER) is a key research area in computer vision and human-computer interaction. Despite recent advances, challenges persist, especially in generalizing to new scenarios. In fact, zero-shot FER significantly reduces the performance of state-of-the-art FER models. The community has recently started to explore the integration of knowledge from Large Language Models for visual tasks. In this work, we evaluate a broad collection of Visual Language Models (VLMs), avoiding the lack of task-specific knowledge by adopting a Visual Question Answering strategy. We compare the proposed pipeline with state-of-the-art FER models, both integrating and excluding VLMs, evaluating well-known FER benchmarks: AffectNet, FERPlus, and RAF-DB. The results show state-of-the-art performance for some VLMs in zero-shot FER scenarios, suggesting a research line for further exploration to improve FER generalization.
URI: https://accedacris.ulpgc.es/jspui/handle/10553/159788
ISBN: 979-8-3315-6867-2
DOI: 10.1109/VCIP67698.2025.11396850
Source: International Conference on Visual Communications and Image Processing (VCIP 2025), 1-4 diciembre 2025, Klagenfurt, Austria
Appears in Collections:Actas de congresos
Show full item record

Google ScholarTM

Check

Altmetric


Share



Export metadata



Items in accedaCRIS are protected by copyright, with all rights reserved, unless otherwise indicated.