scispace - formally typeset
Open AccessPosted Content

Ground Truth Evaluation of Neural Network Explanations with CLEVR-XAI.

Reads0
Chats0
TLDR
In this article, a ground truth-based evaluation framework for explainable AI (XAI) methods based on the CLEVR visual question answering task is proposed, which provides a selective, controlled and realistic testbed for the evaluation of neural network explanations.
Abstract
The rise of deep learning in today's applications entailed an increasing need in explaining the model's decisions beyond prediction performances in order to foster trust and accountability Recently, the field of explainable AI (XAI) has developed methods that provide such explanations for already trained neural networks In computer vision tasks such explanations, termed heatmaps, visualize the contributions of individual pixels to the prediction So far XAI methods along with their heatmaps were mainly validated qualitatively via human-based assessment, or evaluated through auxiliary proxy tasks such as pixel perturbation, weak object localization or randomization tests Due to the lack of an objective and commonly accepted quality measure for heatmaps, it was debatable which XAI method performs best and whether explanations can be trusted at all In the present work, we tackle the problem by proposing a ground truth based evaluation framework for XAI methods based on the CLEVR visual question answering task Our framework provides a (1) selective, (2) controlled and (3) realistic testbed for the evaluation of neural network explanations We compare ten different explanation methods, resulting in new insights about the quality and properties of XAI methods, sometimes contradicting with conclusions from previous comparative studies The CLEVR-XAI dataset and the benchmarking code can be found at this https URL

read more

Citations
More filters
Journal ArticleDOI

Explaining Deep Neural Networks and Beyond: A Review of Methods and Applications

TL;DR: In this paper, the authors provide a timely overview of explainable AI, with a focus on 'post-hoc' explanations, explain its theoretical foundations, and put interpretability algorithms to a test both from a theory and comparative evaluation perspective using extensive simulations.
Journal ArticleDOI

Explaining Deep Neural Networks and Beyond: A Review of Methods and Applications

TL;DR: In this paper, the authors provide a timely overview of post hoc explanations and explain its theoretical foundations, and put interpretability algorithms to a test both from a theory and comparative evaluation perspective using extensive simulations, and demonstrate successful usage of XAI in a representative selection of application scenarios.
Posted Content

Toward Interpretable Machine Learning: Transparent Deep Neural Networks and Beyond

TL;DR: This work aims to provide a timely overview of this active emerging field of machine learning and explain its theoretical foundations, put interpretability algorithms to a test both from a theory and comparative evaluation perspective using extensive simulations, and outline best practice aspects.
Posted Content

Explaining Bayesian Neural Networks.

TL;DR: In this paper, the authors propose a holistic explanation framework for explaining BNNs, where the network weights follow a probability distribution, and thus the standard explanation extends to an explanation distribution.
Posted Content

Software for Dataset-wide XAI: From Local Explanations to Global Insights with Zennit, CoRelAy, and ViRelAy.

TL;DR: Zennit as discussed by the authors is a post-hoc attribution framework implemented in PyTorch and CoRelAy is a web-application to interactively explore data, attributions, and analysis results.
References
More filters
Journal ArticleDOI

ImageNet Large Scale Visual Recognition Challenge

TL;DR: The ImageNet Large Scale Visual Recognition Challenge (ILSVRC) as mentioned in this paper is a benchmark in object category classification and detection on hundreds of object categories and millions of images, which has been run annually from 2010 to present, attracting participation from more than fifty institutions.
Proceedings ArticleDOI

BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

TL;DR: BERT as mentioned in this paper pre-trains deep bidirectional representations from unlabeled text by jointly conditioning on both left and right context in all layers, which can be fine-tuned with just one additional output layer to create state-of-the-art models for a wide range of tasks.
Book ChapterDOI

Visualizing and Understanding Convolutional Networks

TL;DR: A novel visualization technique is introduced that gives insight into the function of intermediate feature layers and the operation of the classifier in large Convolutional Network models, used in a diagnostic role to find model architectures that outperform Krizhevsky et al on the ImageNet classification benchmark.
Proceedings ArticleDOI

"Why Should I Trust You?": Explaining the Predictions of Any Classifier

TL;DR: In this article, the authors propose LIME, a method to explain models by presenting representative individual predictions and their explanations in a non-redundant way, framing the task as a submodular optimization problem.
Journal ArticleDOI

Dermatologist-level classification of skin cancer with deep neural networks

TL;DR: This work demonstrates an artificial intelligence capable of classifying skin cancer with a level of competence comparable to dermatologists, trained end-to-end from images directly, using only pixels and disease labels as inputs.
Related Papers (5)
Trending Questions (1)
What ground truth evaluation methods are used to evaluate a vision system?

The paper does not mention any specific ground truth evaluation methods used to evaluate a vision system.