Desiderata for Interpretability: Explaining Decision Tree Predictions with Counterfactuals

Authors

  • Kacper Sokol University of Bristol
  • Peter Flach University of Bristol

DOI:

https://doi.org/10.1609/aaai.v33i01.330110035

Abstract

Explanations in machine learning come in many forms, but a consensus regarding their desired properties is still emerging. In our work we collect and organise these explainability desiderata and discuss how they can be used to systematically evaluate properties and quality of an explainable system using the case of class-contrastive counterfactual statements. This leads us to propose a novel method for explaining predictions of a decision tree with counterfactuals. We show that our model-specific approach exploits all the theoretical advantages of counterfactual explanations, hence improves decision tree interpretability by decoupling the quality of the interpretation from the depth and width of the tree.

Downloads

Published

2019-07-17

How to Cite

Sokol, K., & Flach, P. (2019). Desiderata for Interpretability: Explaining Decision Tree Predictions with Counterfactuals. Proceedings of the AAAI Conference on Artificial Intelligence, 33(01), 10035-10036. https://doi.org/10.1609/aaai.v33i01.330110035

Issue

Section

Student Abstract Track