Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
Review
. 2021 Jul 1:4:688969.
doi: 10.3389/fdata.2021.688969. eCollection 2021.

Principles and Practice of Explainable Machine Learning

Affiliations
Review

Principles and Practice of Explainable Machine Learning

Vaishak Belle et al. Front Big Data. .

Abstract

Artificial intelligence (AI) provides many opportunities to improve private and public life. Discovering patterns and structures in large troves of data in an automated manner is a core component of data science, and currently drives applications in diverse areas such as computational biology, law and finance. However, such a highly positive impact is coupled with a significant challenge: how do we understand the decisions suggested by these systems in order that we can trust them? In this report, we focus specifically on data-driven methods-machine learning (ML) and pattern recognition models in particular-so as to survey and distill the results and observations from the literature. The purpose of this report can be especially appreciated by noting that ML models are increasingly deployed in a wide range of businesses. However, with the increasing prevalence and complexity of methods, business stakeholders in the very least have a growing number of concerns about the drawbacks of models, data-specific biases, and so on. Analogously, data science practitioners are often not aware about approaches emerging from the academic literature or may struggle to appreciate the differences between different methods, so end up using industry standards such as SHAP. Here, we have undertaken a survey to help industry practitioners (but also data scientists more broadly) understand the field of explainable machine learning better and apply the right tools. Our latter sections build a narrative around a putative data scientist, and discuss how she might go about explaining her models by asking the right questions. From an organization viewpoint, after motivating the area broadly, we discuss the main developments, including the principles that allow us to study transparent models vs. opaque models, as well as model-specific or model-agnostic post-hoc explainability approaches. We also briefly reflect on deep learning models, and conclude with a discussion about future research directions.

Keywords: black-box models; explainable AI; machine learning; survey; transparent models.

PubMed Disclaimer

Conflict of interest statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Figures

FIGURE 1
FIGURE 1
Concerns faced by various stakeholders.
FIGURE 2
FIGURE 2
A taxonomic view on XAI.
FIGURE 3
FIGURE 3
Jane’s agenda and challenge: which model offers the best trade-off in terms of accuracy vs. explainability?
FIGURE 4
FIGURE 4
Jane’s choices: should she go for a transparent model or an opaque one?
FIGURE 5
FIGURE 5
As transparent models become increasingly complex they may lose their explainability features. The primary goal is to maintain a balance between explainability and accuracy. In cases where this is not possible, opaque models paired with post hoc XAI approaches provide an alternative solution.
FIGURE 6
FIGURE 6
Jane decides to use SHAP, but cannot resolve all of the stakeholder’s questions. Its also worth noting that although SHAP is an important method for explaining opaque models, users should be aware of its limitations, often arising from either the optimization objective or the underlying approximation.
FIGURE 7
FIGURE 7
Visualizations can facilitate understanding the model’s reasoning, both on an instance and a global level. Most of these approaches make a set of assumptions, so choosing the appropriate one depends on the application.
FIGURE 8
FIGURE 8
Counterfactuals produce a hypothetical instance, representing a minimal set of changes of the original one, so the model classifies it in a different category.
FIGURE 9
FIGURE 9
Local explanations as rules. High precision means that the rule is robust and that similar instances will get the same outcome. High coverage means that large number of the points satisfy the rule’s premises, so the rule “generalizes” better.
FIGURE 10
FIGURE 10
The quality of a ML model is vastly affected by the quality of the data it is trained on. Finding influential points that can, for example, alter the decision boundary or encourage the model to take a certain decision, contributes in having a more complete picture of the model’s reasoning.
FIGURE 11
FIGURE 11
Extracting rules from a random forest. Frequency of a rule is defined as the proportion of data instances satisfying the rule condition. The frequency measures the popularity of the rule. Error of a rule is defined as the number of incorrectly classified instances determined by the rule. So she is able to say that for 80% of the customers with 100% accuracy (ie. 0% error), when income >20 k and there are zero missed payments, the application is approved.
FIGURE 12
FIGURE 12
A short comparison of model agnostic vs. model specific approaches.
FIGURE 13
FIGURE 13
A list of possible questions of interest when explaining a model. This highlights the need for combining multiple techniques together and that there is no catch-all approach.
FIGURE 14
FIGURE 14
A sample pipeline, that is, a “cheat sheet” of sorts for approaching explainability.
FIGURE 15
FIGURE 15
Using SHAP, PDF and counterfactuals, visualized in terms of instances.
FIGURE 16
FIGURE 16
Using anchors, deletion diagnostics and intrees, visualized in terms of instances.
FIGURE 17
FIGURE 17
Possible avenues for XAI research.

References

    1. Adebayo J., Kagal L. (2016). Iterative Orthogonal Feature Projection for Diagnosing Bias in Black-Box Models, FATML Workshop 2016, New York, NY.
    1. Agrahari R., Foroushani A., Docking T. R., Chang L., Duns G., Hudoba M., Karsan A., Zare H. (2018). Applications of Bayesian Network Models in Predicting Types of Hematological Malignancies. Scientific Reports. - PMC - PubMed
    1. Arrieta A. B., Díaz-Rodríguez N., Del Ser J., Bennetot A., Tabik S., Barbado A., et al. (2019). Explainable Artificial Intelligence (Xai): Concepts, Taxonomies, Opportunities and Challenges toward Responsible Ai. arXiv preprint arXiv:1910.10045.
    1. Augasta M. G., Kathirvalavakumar T. (2012). Reverse Engineering the Neural Networks for Rule Extraction in Classification Problems. Heidelberg, Germany: Neural Processing Letters.
    1. Auret L., Aldrich C. (2012). Interpretation of Nonlinear Relationships between Process Variables by Use of Random Forests. Minerals Eng. 35, 27–42. 10.1016/j.mineng.2012.05.008 - DOI

LinkOut - more resources