Leveraging generative models to characterize the failure conditions of image classifiers
The IJCAI-ECAI-22 Workshop on Artificial Intelligence Safety (AISafety 2022), Jul 2022, Vienna, Austria We address in this work the question of identifying the failure conditions of a given image classifier. To do so, we exploit the capacity of producing controllable distributions of high quality im...
Gespeichert in:
Hauptverfasser: | , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | The IJCAI-ECAI-22 Workshop on Artificial Intelligence Safety
(AISafety 2022), Jul 2022, Vienna, Austria We address in this work the question of identifying the failure conditions of
a given image classifier. To do so, we exploit the capacity of producing
controllable distributions of high quality image data made available by recent
Generative Adversarial Networks (StyleGAN2): the failure conditions are
expressed as directions of strong performance degradation in the generative
model latent space. This strategy of analysis is used to discover corner cases
that combine multiple sources of corruption, and to compare in more details the
behavior of different classifiers. The directions of degradation can also be
rendered visually by generating data for better interpretability. Some
degradations such as image quality can affect all classes, whereas other ones
such as shape are more class-specific. The approach is demonstrated on the
MNIST dataset that has been completed by two sources of corruption: noise and
blur, and shows a promising way to better understand and control the risks of
exploiting Artificial Intelligence components for safety-critical applications. |
---|---|
DOI: | 10.48550/arxiv.2410.12814 |