A Methodology based on Rebalancing Techniques to measure and improve Fairness in Artificial Intelligence algorithms

Por favor, use este identificador para citar o enlazar este ítem: http://hdl.handle.net/10045/123225
Información del item - Informació de l'item - Item information
Título: A Methodology based on Rebalancing Techniques to measure and improve Fairness in Artificial Intelligence algorithms
Autor/es: Lavalle, Ana | Maté, Alejandro | Trujillo, Juan | García Carrasco, Jorge
Grupo/s de investigación o GITE: Lucentia
Centro, Departamento o Servicio: Universidad de Alicante. Departamento de Lenguajes y Sistemas Informáticos
Palabras clave: Artificial Intelligence | Fairness | Rebalancing techniques
Área/s de conocimiento: Lenguajes y Sistemas Informáticos
Fecha de publicación: 25-abr-2022
Editor: CEUR
Cita bibliográfica: Proceedings of the 24th International Workshop on Design, Optimization, Languages and Analytical Processing of Big Data (DOLAP), co-located with the 25th International Conference on Extending Database Technology and the 25th International Conference on Database Theory (EDBT/ICDT 2022). Edinburgh, UK, March 29, 2022. CEUR Workshop Proceedings, Vol-3130, 81-85
Resumen: Artificial Intelligence (AI) has become one of the key drivers for the next decade. As important decisions are increasingly supported or directly made by AI systems, concerns regarding the rationale and fairness in their outputs are becoming more and more prominent nowadays. Following the recent interest in fairer predictions, several metrics for measuring fairness have been proposed, leading to different objectives which may need to be addressed in different fashion. In this paper, we propose (i) a methodology for analyzing and improving fairness in AI predictions by selecting sensitive attributes that should be protected; (ii) We analyze how the most common rebalance approaches affect the fairness of AI predictions and how they compare to the alternatives of removing or creating separate classifiers for each group within a protected attribute. Finally, (iii) our methodology generates a set of tables that can be easily computed for choosing the best alternative in each particular case. The main advantage of our methodology is that it allows AI practitioners to measure and improve fairness in AI algorithms in a systematic way. In order to check our proposal, we have properly applied it to the COMPAS dataset, which has been widely demonstrated to be biased by several previous studies.
Patrocinador/es: This work has been co-funded by the AETHER-UA project (PID2020-112540RB-C43), funded by Spanish Ministry of Science and Innovation and the BALLADEER (PROMETEO/2021/088) projects, funded by the Conselleria de Innovación, Universidades, Ciencia y Sociedad Digital (Generalitat Valenciana).
URI: http://hdl.handle.net/10045/123225
ISSN: 1613-0073
Idioma: eng
Tipo: info:eu-repo/semantics/conferenceObject
Derechos: © Copyright 2022 for this paper by its author(s). Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
Revisión científica: si
Versión del editor: http://ceur-ws.org/Vol-3130/
Aparece en las colecciones:INV - LUCENTIA - Comunicaciones a Congresos, Conferencias, etc.

Archivos en este ítem:
Archivos en este ítem:
Archivo Descripción TamañoFormato 
ThumbnailLavalle_etal_2022_CEUR.pdf875,57 kBAdobe PDFAbrir Vista previa


Este ítem está licenciado bajo Licencia Creative Commons Creative Commons