Quantifying Local Model Validity using Active Learning
Real-world applications of machine learning models are often subject to legal or policy-based regulations. Some of these regulations require ensuring the validity of the model, i.e., the approximation error being smaller than a threshold. A global metric is generally too insensitive to determine the...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Real-world applications of machine learning models are often subject to legal
or policy-based regulations. Some of these regulations require ensuring the
validity of the model, i.e., the approximation error being smaller than a
threshold. A global metric is generally too insensitive to determine the
validity of a specific prediction, whereas evaluating local validity is costly
since it requires gathering additional data.We propose learning the model error
to acquire a local validity estimate while reducing the amount of required data
through active learning. Using model validation benchmarks, we provide
empirical evidence that the proposed method can lead to an error model with
sufficient discriminative properties using a relatively small amount of data.
Furthermore, an increased sensitivity to local changes of the validity bounds
compared to alternative approaches is demonstrated. |
---|---|
DOI: | 10.48550/arxiv.2406.07474 |