Augmentation by Counterfactual Explanation -- Fixing an Overconfident Classifier
A highly accurate but overconfident model is ill-suited for deployment in critical applications such as healthcare and autonomous driving. The classification outcome should reflect a high uncertainty on ambiguous in-distribution samples that lie close to the decision boundary. The model should also...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | A highly accurate but overconfident model is ill-suited for deployment in
critical applications such as healthcare and autonomous driving. The
classification outcome should reflect a high uncertainty on ambiguous
in-distribution samples that lie close to the decision boundary. The model
should also refrain from making overconfident decisions on samples that lie far
outside its training distribution, far-out-of-distribution (far-OOD), or on
unseen samples from novel classes that lie near its training distribution
(near-OOD). This paper proposes an application of counterfactual explanations
in fixing an over-confident classifier. Specifically, we propose to fine-tune a
given pre-trained classifier using augmentations from a counterfactual
explainer (ACE) to fix its uncertainty characteristics while retaining its
predictive performance. We perform extensive experiments with detecting
far-OOD, near-OOD, and ambiguous samples. Our empirical results show that the
revised model have improved uncertainty measures, and its performance is
competitive to the state-of-the-art methods. |
---|---|
DOI: | 10.48550/arxiv.2210.12196 |