Concepedia

Publication | Open Access

Explaining Explanations in AI

696

Citations

89

References

2019

Year

TLDR

Interpretability research has largely produced simplified surrogate models that approximate the decision logic of complex AI systems, serving as pedagogical tools for predicting decisions and diagnosing failures, yet Box’s maxim reminds us that all models are wrong but some are useful. The authors aim to distinguish between surrogate models and philosophical/sociological explanations, contrasting schools of thought on what constitutes an explanation and proposing that machine learning could benefit from a broader perspective. They describe surrogate models as do‑it‑yourself kits that enable practitioners to answer counterfactual or contrastive questions without external assistance. They find that while surrogate models provide valuable explanatory power, delivering them as explanations is more difficult than necessary, and alternative explanation forms may lack similar trade‑offs, suggesting that machine learning could benefit from a wider explanatory framework.

Abstract

Recent work on interpretability in machine learning and AI has focused on the building of simplified models that approximate the true criteria used to make decisions. These models are a useful pedagogical device for teaching trained professionals how to predict what decisions will be made by the complex system, and most importantly how the system might break. However, when considering any such model it's important to remember Box's maxim that "All models are wrong but some are useful." We focus on the distinction between these models and explanations in philosophy and sociology. These models can be understood as a "do it yourself kit" for explanations, allowing a practitioner to directly answer "what if questions" or generate contrastive explanations without external assistance. Although a valuable ability, giving these models as explanations appears more difficult than necessary, and other forms of explanation may not have the same trade-offs. We contrast the different schools of thought on what makes an explanation, and suggest that machine learning might benefit from viewing the problem more broadly.

References

YearCitations

2017

20.1K

2016

14K

2024

13.1K

2014

8.1K

2017

7.6K

2013

4.9K

2017

2.6K

2017

2.4K

2016

2.2K

2016

2.1K

Page 1