Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Preprints and Working Papers
  4. The future of human-centric eXplainable Artificial Intelligence (XAI) is not post-hoc explanations
 
preprint

The future of human-centric eXplainable Artificial Intelligence (XAI) is not post-hoc explanations

Swamy, Vinitra  
•
Frej, Jibril Albachir  
•
Käser, Tanja  
2023

Explainable Artificial Intelligence (XAI) plays a crucial role in enabling human understanding and trust in deep learning systems, often defined as determining which features are most important to a model's prediction. As models get larger, more ubiquitous, and pervasive in aspects of daily life, explainability is necessary to avoid or minimize adverse effects of model mistakes. Unfortunately, current approaches in human-centric XAI (e.g. predictive tasks in healthcare, education, or personalized ads) tend to rely on a single explainer. This is a particularly concerning trend when considering that recent work has identified systematic disagreement in explainability methods when applied to the same points and underlying black-box models. In this paper, we therefore present a call for action to address the limitations of current state-of-the-art explainers. We propose to shift from post-hoc explainability to designing interpretable neural network architectures; moving away from approximation techniques in human-centric and high impact applications. We identify five needs of human-centric XAI (real-time, accurate, actionable, human-interpretable, and consistent) and propose two schemes for interpretable-by-design neural network workflows (adaptive routing for interpretable conditional computation and diagnostic benchmarks for iterative model learning). We postulate that the future of human-centric XAI is neither in explaining black-boxes nor in reverting to traditional, interpretable models, but in neural networks that are intrinsically interpretable.

  • Files
  • Details
  • Metrics
Type
preprint
DOI
10.48550/arxiv.2307.00364
Author(s)
Swamy, Vinitra  
Frej, Jibril Albachir  
Käser, Tanja  
Date Issued

2023

Note

Viewpoint paper, under review at JAIR

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
ML4ED  
AVP-E-LEARN  
Available on Infoscience
July 4, 2023
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/198837
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés