Critical empirical study on black-box explanations in AI - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Année : 2021

Critical empirical study on black-box explanations in AI

Résumé

This paper provides empirical concerns about post-hoc explanations of black-box ML models, one of the major trends in AI explainability (XAI), by showing its lack of interpretability and societal consequences. Using a representative consumer panel to test our assumptions, we report three main findings. First, we show that post-hoc explanations of black-box model tend to give partial and biased information on the underlying mechanism of the algorithm and can be subject to manipulation or information withholding by diverting users' attention. Secondly, we show the importance of tested behavioral indicators, in addition to self-reported perceived indicators, to provide a more comprehensive view of the dimensions of interpretability. This paper contributes to shedding new light on the actual theoretical debate between intrinsically transparent AI models and post-hoc explanations of black-box complex models-a debate which is likely to play a highly influential role in the future development and operationalization of AI systems.
Fichier principal
Vignette du fichier
final version 2.pdf (237.83 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-03357663 , version 1 (28-09-2021)

Identifiants

  • HAL Id : hal-03357663 , version 1

Citer

Jean-Marie John-Mathews. Critical empirical study on black-box explanations in AI. ICIS 2021 : 42nd International Conference on Information Systems, Association for Information Systems (AIS), Dec 2021, Austin, Texas, United States. ⟨hal-03357663⟩
199 Consultations
214 Téléchargements

Partager

Gmail Facebook X LinkedIn More