Classify and Generate Reciprocally: Simultaneous Positive-Unlabelled Learning and Conditional Generation with Extra Data
The scarcity of class-labeled data is a ubiquitous bottleneck in many machine learning problems. While abundant unlabeled data typically exist and provide a potential solution, it is highly challenging to exploit them. In this paper, we address this problem by leveraging Positive-Unlabeled~(PU) clas...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | The scarcity of class-labeled data is a ubiquitous bottleneck in many machine
learning problems. While abundant unlabeled data typically exist and provide a
potential solution, it is highly challenging to exploit them. In this paper, we
address this problem by leveraging Positive-Unlabeled~(PU) classification and
the conditional generation with extra unlabeled data \emph{simultaneously}. In
particular, we present a novel training framework to jointly target both PU
classification and conditional generation when exposed to extra data,
especially out-of-distribution unlabeled data, by exploring the interplay
between them: 1) enhancing the performance of PU classifiers with the
assistance of a novel Classifier-Noise-Invariant Conditional GAN~(CNI-CGAN)
that is robust to noisy labels, 2) leveraging extra data with predicted labels
from a PU classifier to help the generation. Theoretically, we prove the
optimal condition of CNI-CGAN, and experimentally, we conducted extensive
evaluations on diverse datasets, verifying the simultaneous improvements in
both classification and generation. |
---|---|
DOI: | 10.48550/arxiv.2006.07841 |