Functional Risk Minimization
The field of Machine Learning has changed significantly since the 1970s. However, its most basic principle, Empirical Risk Minimization (ERM), remains unchanged. We propose Functional Risk Minimization~(FRM), a general framework where losses compare functions rather than outputs. This results in bet...
Gespeichert in:
Hauptverfasser: | , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | The field of Machine Learning has changed significantly since the 1970s.
However, its most basic principle, Empirical Risk Minimization (ERM), remains
unchanged. We propose Functional Risk Minimization~(FRM), a general framework
where losses compare functions rather than outputs. This results in better
performance in supervised, unsupervised, and RL experiments. In the FRM
paradigm, for each data point $(x_i,y_i)$ there is function $f_{\theta_i}$ that
fits it: $y_i = f_{\theta_i}(x_i)$. This allows FRM to subsume ERM for many
common loss functions and to capture more realistic noise processes. We also
show that FRM provides an avenue towards understanding generalization in the
modern over-parameterized regime, as its objective can be framed as finding the
simplest model that fits the training data. |
---|---|
DOI: | 10.48550/arxiv.2412.21149 |