What Is Wrong with My Model? Identifying Systematic Problems with Semantic Data Slicing
Machine learning models make mistakes, yet sometimes it is difficult to identify the systematic problems behind the mistakes. Practitioners engage in various activities, including error analysis, testing, auditing, and red-teaming, to form hypotheses of what can go (or has gone) wrong with their mod...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Machine learning models make mistakes, yet sometimes it is difficult to
identify the systematic problems behind the mistakes. Practitioners engage in
various activities, including error analysis, testing, auditing, and
red-teaming, to form hypotheses of what can go (or has gone) wrong with their
models. To validate these hypotheses, practitioners employ data slicing to
identify relevant examples. However, traditional data slicing is limited by
available features and programmatic slicing functions. In this work, we propose
SemSlicer, a framework that supports semantic data slicing, which identifies a
semantically coherent slice, without the need for existing features. SemSlicer
uses Large Language Models to annotate datasets and generate slices from any
user-defined slicing criteria. We show that SemSlicer generates accurate slices
with low cost, allows flexible trade-offs between different design dimensions,
reliably identifies under-performing data slices, and helps practitioners
identify useful data slices that reflect systematic problems. |
---|---|
DOI: | 10.48550/arxiv.2409.09261 |