Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2022 Apr 8:5:825565.
doi: 10.3389/frai.2022.825565. eCollection 2022.

GANterfactual-Counterfactual Explanations for Medical Non-experts Using Generative Adversarial Learning

Affiliations

GANterfactual-Counterfactual Explanations for Medical Non-experts Using Generative Adversarial Learning

Silvan Mertes et al. Front Artif Intell. .

Abstract

With the ongoing rise of machine learning, the need for methods for explaining decisions made by artificial intelligence systems is becoming a more and more important topic. Especially for image classification tasks, many state-of-the-art tools to explain such classifiers rely on visual highlighting of important areas of the input data. Contrary, counterfactual explanation systems try to enable a counterfactual reasoning by modifying the input image in a way such that the classifier would have made a different prediction. By doing so, the users of counterfactual explanation systems are equipped with a completely different kind of explanatory information. However, methods for generating realistic counterfactual explanations for image classifiers are still rare. Especially in medical contexts, where relevant information often consists of textural and structural information, high-quality counterfactual images have the potential to give meaningful insights into decision processes. In this work, we present GANterfactual, an approach to generate such counterfactual image explanations based on adversarial image-to-image translation techniques. Additionally, we conduct a user study to evaluate our approach in an exemplary medical use case. Our results show that, in the chosen medical use-case, counterfactual explanations lead to significantly better results regarding mental models, explanation satisfaction, trust, emotions, and self-efficacy than two state-of-the art systems that work with saliency maps, namely LIME and LRP.

Keywords: counterfactual explanations; explainable AI; generative adversarial networks; image-to-image translation; interpretable machine learning; machine learning.

PubMed Disclaimer

Conflict of interest statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Figures

Figure 1
Figure 1
Schematic overview of our approach. A CycleGAN architecture is extended with the classifier that shall be explained. Both the generators of the CycleGAN include the classifier's decisions for the generated data into their loss function.
Figure 2
Figure 2
Example images of the used dataset. The top row shows images that are labeled as Normal, while the bottom row shows images labeled as Lung Opacity, indicating lungs that are suffering from pneumonia.
Figure 3
Figure 3
Examples of counterfactual images produced with our proposed approach. In each pair, the left image shows the original image, while the right image shows the corresponding counterfactual explanation. The red boxes were added manually to point the reader to the regions that were altered the most. The original images in the top row were classified as normal, while the original images in the bottom row were classified as pneumonia. The shown counterfactual images were all classified as the opposite as their respective counterpart.
Figure 4
Figure 4
Computational evaluation results of the counterfactual image generation performance. The confusion matrices show the number of samples out of each subset (Normal, Pneumonia, Total) of the rsna dataset that the classifier predicted to be the respective class before (y-axis) and after (x-axis) the samples had been translated by either the original CycleGAN or by our approach.
Figure 5
Figure 5
An example x-ray image classified as Pneumonia, as well as the different XAI visualizations used in our study when the slider is fully on the right side. Best viewed in color.
Figure 6
Figure 6
A simplified schematic of our prediction task.
Figure 7
Figure 7
Results of the explanation satisfaction and trust questionnaires. Error bars represent the 95% Confidence Interval (CI).
Figure 8
Figure 8
Results of the prediction task, and the task reflection questions. Error bars represent the 95% Confidence Interval (CI).
Figure 9
Figure 9
Results of the emotion questionnaires. Participants in the counterfactual condition felt significantly less angry and more relaxed compared to the LRP saliency map condition. For LIME, no significant differences were found. Error bars represent the 95% CI.
Figure 10
Figure 10
Significant differences regarding self-efficacy and general confidence of the participants in their predictions of the AI between the counterfactual condition and the saliency map conditions (LRP and LIME). Error bars represent the 95% CI.
Figure 11
Figure 11
Confidence of the participants in correct and false predictions. The significant difference between the counterfactual condition and the saliency map conditions is based on the confidence in correct predictions, not in the incorrect ones. Error bars represent the 95% CI.

Similar articles

Cited by

References

    1. Ahsan M. M., Gupta K. D., Islam M., Sen S., Rahman M. L., Hossain M. (2020). Study of different deep learning approach with explainable AI for screening patients with COVID-19 symptoms: using CT scan and chest x-ray image dataset. CoRR. Available online at: https://arxiv.org/abs/2007.12525
    1. Alqaraawi A., Schuessler M., Weiß P., Costanza E., Berthouze N. (2020). Evaluating saliency map explanations for convolutional neural networks: a user study, in IUI '20: 25th International Conference on Intelligent User Interfaces (Cagliari: ), 275–285. 10.1145/3377325.3377519 - DOI
    1. Anderson A., Dodge J., Sadarangani A., Juozapaitis Z., Newman E., Irvine J., et al. . (2019). Explaining reinforcement learning to mere mortals: an empirical study, in Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI-19 (Macao: ), 1328–1334. 10.24963/ijcai.2019/184 - DOI
    1. Arrieta A. B., Díaz-Rodríguez N., Del Ser J., Bennetot A., Tabik S., Barbado A., et al. . (2020). Explainable artificial intelligence (XAI): concepts, taxonomies, opportunities and challenges toward responsible AI. Inform. Fus. 58, 82–115. 10.1016/j.inffus.2019.12.012 - DOI
    1. Bach S., Binder A., Montavon G., Klauschen F., Müller K.-R., Samek W. (2015). On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation. PLoS ONE 10:e130140. 10.1371/journal.pone.0130140 - DOI - PMC - PubMed