Testing machine learning explanation methods

There are many methods for explaining why a machine learning model produces a given output in response to a given input. The relative merits of these methods are often debated using theoretical arguments and illustrative examples. This paper provides a large-scale empirical test of four widely used...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Neural computing & applications 2023-08, Vol.35 (24), p.18073-18084
1. Verfasser: Anderson, Andrew A.
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:There are many methods for explaining why a machine learning model produces a given output in response to a given input. The relative merits of these methods are often debated using theoretical arguments and illustrative examples. This paper provides a large-scale empirical test of four widely used explanation methods by comparing how well their algorithmically generated denial reasons align with lender-provided denial reasons using a dataset of home mortgage applications. On a held-out sample of 10,000 denied applications, Shapley additive explanations (SHAP) correspond most closely with lender-provided reasons. SHAP is also the most computationally efficient. As a second contribution, this paper presents a method for computing integrated gradient explanations that can be used for non-differentiable models such as XGBoost.
ISSN:0941-0643
1433-3058
DOI:10.1007/s00521-023-08597-8