Explainable Artificial Intelligence, or XAI, is a vibrant research topic in the artificial intelligence community. It is raising growing interest across methods and domains, especially those involving high-stakes decision-making, such as the biomedical sector. Much has been written about the subject, yet XAI still lacks shared terminology and a framework capable of providing structural soundness to explanations, a crucial need for decisions that impact healthcare. In our work, we address these issues by proposing a novel definition of explanation that synthesizes insights from the existing literature. We recognize that explanations are not atomic, but rather the combination of evidence stemming from the model and its input-output mapping, along with the human interpretation of this evidence. Furthermore, we fit explanations into the properties of faithfulness (\textit{i.e.}, the explanation is an accurate description of the model's inner workings and decision-making process) and plausibility (\textit{i.e.}, how much the explanation seems convincing to the user). Our theoretical framework simplifies the operationalization of these properties and provides new insights into common explanation methods that we analyze through case studies. We explore the impact of our framework in the sensitive domain of biomedicine, where XAI can play a central role in generating trust by balancing faithfulness and plausibility.

Machine Learning Models Explanations As Interpretations of Evidence: A Theoretical Framework of Explainability and Its Implications on High-Stakes Biomedical Decision-Making

Matteo Rizzo;Alberto Veneri;Matteo Marcuzzo;Alessandro Zangari;Andrea Albarelli;Claudio Lucchese;Marco Salvatore Nobile;
2025-01-01

Abstract

Explainable Artificial Intelligence, or XAI, is a vibrant research topic in the artificial intelligence community. It is raising growing interest across methods and domains, especially those involving high-stakes decision-making, such as the biomedical sector. Much has been written about the subject, yet XAI still lacks shared terminology and a framework capable of providing structural soundness to explanations, a crucial need for decisions that impact healthcare. In our work, we address these issues by proposing a novel definition of explanation that synthesizes insights from the existing literature. We recognize that explanations are not atomic, but rather the combination of evidence stemming from the model and its input-output mapping, along with the human interpretation of this evidence. Furthermore, we fit explanations into the properties of faithfulness (\textit{i.e.}, the explanation is an accurate description of the model's inner workings and decision-making process) and plausibility (\textit{i.e.}, how much the explanation seems convincing to the user). Our theoretical framework simplifies the operationalization of these properties and provides new insights into common explanation methods that we analyze through case studies. We explore the impact of our framework in the sensitive domain of biomedicine, where XAI can play a central role in generating trust by balancing faithfulness and plausibility.
2025
24
File in questo prodotto:
File Dimensione Formato  
A_Theoretical_Framework_for_AI_Models_Explainability_with_Application_in_Biomedicine.pdf

accesso aperto

Tipologia: Documento in Pre-print
Licenza: Accesso libero (no vincoli)
Dimensione 890.02 kB
Formato Adobe PDF
890.02 kB Adobe PDF Visualizza/Apri

I documenti in ARCA sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/10278/5106607
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
social impact