SoK: Unintended Interactions among Machine Learning Defenses and Risks
Machine learning (ML) models cannot neglect risks to security, privacy, and fairness. Several defenses have been proposed to mitigate such risks. When a defense is effective in mitigating one risk, it may correspond to increased or decreased susceptibility to other risks. Existing research lacks an...
Gespeichert in:
Hauptverfasser: | , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Machine learning (ML) models cannot neglect risks to security, privacy, and
fairness. Several defenses have been proposed to mitigate such risks. When a
defense is effective in mitigating one risk, it may correspond to increased or
decreased susceptibility to other risks. Existing research lacks an effective
framework to recognize and explain these unintended interactions. We present
such a framework, based on the conjecture that overfitting and memorization
underlie unintended interactions. We survey existing literature on unintended
interactions, accommodating them within our framework. We use our framework to
conjecture on two previously unexplored interactions, and empirically validate
our conjectures. |
---|---|
DOI: | 10.48550/arxiv.2312.04542 |