Improved Synthetic Training for Reading Comprehension
Automatically generated synthetic training examples have been shown to improve performance in machine reading comprehension (MRC). Compared to human annotated gold standard data, synthetic training data has unique properties, such as high availability at the possible expense of quality. In view of s...
Gespeichert in:
Hauptverfasser: | , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Automatically generated synthetic training examples have been shown to
improve performance in machine reading comprehension (MRC). Compared to human
annotated gold standard data, synthetic training data has unique properties,
such as high availability at the possible expense of quality. In view of such
differences, in this paper, we explore novel applications of synthetic examples
to MRC. Our proposed pre-training and knowledge distillation strategies show
significant improvements over existing methods. In a particularly surprising
discovery, we observe that synthetic distillation often yields students that
can outperform the teacher model. |
---|---|
DOI: | 10.48550/arxiv.2010.12776 |