Goal scoring, coherent loss and applications to machine learning
Motivated by the binary classification problem in machine learning, we study in this paper a class of decision problems where the decision maker has a list of goals, from which he aims to attain the maximal possible number of goals. In binary classification, this essentially means seeking a predicti...
Gespeichert in:
Veröffentlicht in: | Mathematical programming 2020-07, Vol.182 (1-2), p.103-140 |
---|---|
Hauptverfasser: | , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Motivated by the binary classification problem in machine learning, we study in this paper a class of decision problems where the decision maker has a list of goals, from which he aims to attain the maximal possible number of goals. In binary classification, this essentially means seeking a prediction rule to achieve the lowest probability of misclassification, and computationally it involves minimizing a (difficult) non-convex, 0–1 loss function. To address the intractability, previous methods consider minimizing the
cumulative loss
—the sum of convex surrogates of the 0–1 loss of each goal. We revisit this paradigm and develop instead an
axiomatic
framework by proposing a set of salient properties on functions for goal scoring and then propose the
coherent loss
approach, which is a tractable upper-bound of the loss over the
entire set
of goals. We show that the proposed approach yields a strictly tighter approximation to the total loss (i.e., the number of missed goals) than any convex cumulative loss approach while preserving the convexity of the underlying optimization problem. Moreover, this approach, applied to for binary classification, also has a robustness interpretation which builds a connection to robust SVMs. |
---|---|
ISSN: | 0025-5610 1436-4646 |
DOI: | 10.1007/s10107-019-01387-y |