Multiple Descents in Unsupervised Learning: The Role of Noise, Domain Shift and Anomalies
The phenomenon of double descent has recently gained attention in supervised learning. It challenges the conventional wisdom of the bias-variance trade-off by showcasing a surprising behavior. As the complexity of the model increases, the test error initially decreases until reaching a certain point...
Gespeichert in:
Hauptverfasser: | , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | The phenomenon of double descent has recently gained attention in supervised
learning. It challenges the conventional wisdom of the bias-variance trade-off
by showcasing a surprising behavior. As the complexity of the model increases,
the test error initially decreases until reaching a certain point where the
model starts to overfit the train set, causing the test error to rise. However,
deviating from classical theory, the error exhibits another decline when
exceeding a certain degree of over-parameterization. We study the presence of
double descent in unsupervised learning, an area that has received little
attention and is not yet fully understood. We conduct extensive experiments
using under-complete auto-encoders (AEs) for various applications, such as
dealing with noisy data, domain shifts, and anomalies. We use synthetic and
real data and identify model-wise, epoch-wise, and sample-wise double descent
for all the aforementioned applications. Finally, we assessed the usability of
the AEs for detecting anomalies and mitigating the domain shift between
datasets. Our findings indicate that over-parameterized models can improve
performance not only in terms of reconstruction, but also in enhancing
capabilities for the downstream task. |
---|---|
DOI: | 10.48550/arxiv.2406.11703 |