Improving Facial Expression Recognition Through Data Preparation and Merging

Please use this identifier to cite or link to this item: http://hdl.handle.net/10045/136107
Información del item - Informació de l'item - Item information
Title: Improving Facial Expression Recognition Through Data Preparation and Merging
Authors: Mejia-Escobar, Christian | Cazorla, Miguel | Martinez-Martin, Ester
Research Group/s: Robótica y Visión Tridimensional (RoViT)
Center, Department or Service: Universidad de Alicante. Departamento de Ciencia de la Computación e Inteligencia Artificial
Keywords: Artificial dataset | Deep Learning | Convolutional Neural Network | Emotion Recognition | Facial Expression Recognition | Stable Diffusion
Issue Date: 10-Jul-2023
Publisher: IEEE
Citation: IEEE Access. 2023, 11: 71339-71360. https://doi.org/10.1109/ACCESS.2023.3293728
Abstract: Human emotions present a major challenge for artificial intelligence. Automated emotion recognition based on facial expressions is important to robotics, medicine, psychology, education, security, arts, entertainment and more. Deep learning is promising for capturing complex emotional features. However, there is no training dataset that is large and representative of the full diversity of emotional expressions in all populations and contexts. Current facial datasets are incomplete, biased, unbalanced, error-prone and have different properties. Models learn these limitations and become dependent on specific datasets, hindering their ability to generalize to new data or real-world scenarios. Our work addresses these difficulties and provides the following contributions to improve emotion recognition: 1) a methodology for merging disparate in-the-wild datasets that increases the number of images and enriches the diversity of people, gestures, and attributes of resolution, color, background, lighting and image format; 2) a balanced, unbiased, and well-labeled evaluator dataset, built with a gender, age, and ethnicity predictor and the successful Stable Diffusion model. Single- and cross-dataset experimentation show that our method increases the generalization of the FER2013, NHFI and AffectNet datasets by 13.93%, 24.17% and 7.45%, respectively; and 3) we propose the first and largest artificial emotion dataset, which can complement real datasets in tasks related to facial expression.
Sponsor: This work has been funded by grant CIPROM/2021/017 awarded by the MEEBAI Project (Prometheus Programme for Research Groups on R&D Excellence) from Conselleria de Innovación, Universidades, Ciencia y Sociedad Digital of Generalitat Valenciana (Spain), and partially by the grant awarded by the Central University of Ecuador through budget certification No. 34 of March 25, 2022 for the development of the research project with code DOCT-DI-2020-37.
URI: http://hdl.handle.net/10045/136107
ISSN: 2169-3536
DOI: 10.1109/ACCESS.2023.3293728
Language: eng
Type: info:eu-repo/semantics/article
Rights: This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4.0/
Peer Review: si
Publisher version: https://doi.org/10.1109/ACCESS.2023.3293728
Appears in Collections:INV - RoViT - Artículos de Revistas

Files in This Item:
Files in This Item:
File Description SizeFormat 
ThumbnailMejia-Escobar_etal_2023_IEEE-Access.pdf3,3 MBAdobe PDFOpen Preview


This item is licensed under a Creative Commons License Creative Commons