VOLTA: Improving Generative Diversity by Variational Mutual Information Maximizing Autoencoder
The natural language generation domain has witnessed great success thanks to Transformer models. Although they have achieved state-of-the-art generative quality, they often neglect generative diversity. Prior attempts to tackle this issue suffer from either low model capacity or over-complicated arc...
Gespeichert in:
Hauptverfasser: | , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | The natural language generation domain has witnessed great success thanks to
Transformer models. Although they have achieved state-of-the-art generative
quality, they often neglect generative diversity. Prior attempts to tackle this
issue suffer from either low model capacity or over-complicated architectures.
Some recent methods employ the VAE framework to enhance diversity, but their
latent variables fully depend on the input context, restricting exploration of
the latent space. In this paper, we introduce VOLTA, a framework that elevates
generative diversity by bridging Transformer with VAE via a more effective
cross-attention-based connection, departing from conventional embedding
concatenation or summation. Additionally, we propose integrating InfoGAN-style
latent codes to enable input-independent variability, further diversifying the
generation. Moreover, our framework accommodates discrete inputs alongside its
existing support for continuous inputs. We perform comprehensive experiments
with two types of Transformers on six datasets from three different NLG tasks
to show that our approach can significantly improve generative diversity while
maintaining generative quality. |
---|---|
DOI: | 10.48550/arxiv.2307.00852 |