Para depositar en Docta Complutense, identifícate con tu correo @ucm.es en el SSO institucional. Haz clic en el desplegable de INICIO DE SESIÓN situado en la parte superior derecha de la pantalla. Introduce tu correo electrónico y tu contraseña de la UCM y haz clic en el botón MI CUENTA UCM, no autenticación con contraseña.

Interpretability challenges in machine learning models

Citation

Marín Díaz, G., Carrasco González, R. A., & Gómez González, D. (2021). Interpretability challenges in machine learning models. En J. Pelegrín-Borondo, M. Arias-Oliva, K. Murata, & A. M. Lara Palma (Eds.), Moving technology ethics at the forefront of society, organisations and governments (pp. 205–217). Universidad de La Rioja. https://dialnet.unirioja.es/servlet/articulo?codigo=8036858

Abstract

Decisions based on Machine Learning (ML) algorithms are having an increasingly significant social impact; however, most of these systems are based on black box algorithms, models whose rules are not understandable to humans. On the other hand, different public and private organisations, as well as the scientific community, have recognised the problem of interpretability, focusing on the development of interpretable models (white box) or on methods that allow the explanation of black box models. The aim of this article is to propose a review of the historical evolution and current state of Machine Learning algorithms, analysing the need for interpretability. In this sense, the challenges of interpretability will be addressed from different points of view: in the field of research, legal, industry and regulatory bodies.

Research Projects

Organizational Units

Journal Issue

Description

Keywords