Testing machine learning explanation methods
There are many methods for explaining why a machine learning model produces a given output in response to a given input. The relative merits of these methods are often debated using theoretical arguments and illustrative examples. This paper provides a large-scale empirical test of four widely used...
Gespeichert in:
Veröffentlicht in: | Neural computing & applications 2023-08, Vol.35 (24), p.18073-18084 |
---|---|
1. Verfasser: | |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | There are many methods for explaining why a machine learning model produces a given output in response to a given input. The relative merits of these methods are often debated using theoretical arguments and illustrative examples. This paper provides a large-scale empirical test of four widely used explanation methods by comparing how well their algorithmically generated denial reasons align with lender-provided denial reasons using a dataset of home mortgage applications. On a held-out sample of 10,000 denied applications, Shapley additive explanations (SHAP) correspond most closely with lender-provided reasons. SHAP is also the most computationally efficient. As a second contribution, this paper presents a method for computing integrated gradient explanations that can be used for non-differentiable models such as XGBoost. |
---|---|
ISSN: | 0941-0643 1433-3058 |
DOI: | 10.1007/s00521-023-08597-8 |