Abstract
Explainable artificial intelligence provides tools to better understand predictive models and their decisions, but many such methods are limited to producing insights with respect to a single class. When generating explanations for several classes, reasoning over them to obtain a comprehensive view may be difficult since they can present competing or contradictory evidence. To address this challenge, we introduce the novel paradigm of multi-class explanations. We outline the theory behind such techniques and propose a local surrogate model based on multi-output regression trees—called LIMETREE—that offers faithful and consistent explanations of multiple classes for individual predictions while being post-hoc, model-agnostic and data-universal. On top of strong fidelity guarantees, our implementation delivers a range of diverse explanation types, including counterfactual statements favored in the literature. We evaluate our algorithm with respect to explainability desiderata, through quantitative experiments and via a pilot user study, on image and tabular data classification tasks, comparing it with LIME, which is a state-of-the-art surrogate explainer. Our contributions demonstrate the benefits of multi-class explanations and the wide-ranging advantages of our method across a diverse set of scenarios.
Original language | English |
---|---|
Article number | 929 |
Number of pages | 31 |
Journal | Electronics |
Volume | 14 |
Issue number | 5 |
Early online date | 26 Feb 2025 |
DOIs | |
Publication status | Published - 1 Mar 2025 |
Bibliographical note
Publisher Copyright:© 2025 by the authors.
Keywords
- model-agnostic
- artificial intelligence
- decision tree
- interpretability
- post-hoc
- surrogate
- explainability
- machine learning