SparseGAN: Sparse Generative Adversarial Network for Text Generation
It is still a challenging task to learn a neural text generation model under the framework of generative adversarial networks (GANs) since the entire training process is not differentiable. The existing training strategies either suffer from unreliable gradient estimations or imprecise sentence repr...
Gespeichert in:
Hauptverfasser: | , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Yuan, Liping Zeng, Jiehang Zheng, Xiaoqing |
description | It is still a challenging task to learn a neural text generation model under
the framework of generative adversarial networks (GANs) since the entire
training process is not differentiable. The existing training strategies either
suffer from unreliable gradient estimations or imprecise sentence
representations. Inspired by the principle of sparse coding, we propose a
SparseGAN that generates semantic-interpretable, but sparse sentence
representations as inputs to the discriminator. The key idea is that we treat
an embedding matrix as an over-complete dictionary, and use a linear
combination of very few selected word embeddings to approximate the output
feature representation of the generator at each time step. With such
semantic-rich representations, we not only reduce unnecessary noises for
efficient adversarial training, but also make the entire training process fully
differentiable. Experiments on multiple text generation datasets yield
performance improvements, especially in sequence-level metrics, such as BLEU. |
doi_str_mv | 10.48550/arxiv.2103.11578 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2103_11578</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2103_11578</sourcerecordid><originalsourceid>FETCH-LOGICAL-a678-8649460151cde0d49011a0b20ac4de00975edc18a0526b49e0acb452276304e83</originalsourceid><addsrcrecordid>eNo9j0FPg0AUhPfiwVR_gCf3D4DvLbvL4o1UxSZNeyh38oDXhFhL8yBY_721bXqaycxkkk-pJ4TYBufgheTYTbFBSGJEl4Z79bY5kAxc5KtXfbG64D0Ljd3EOm8nloGko51e8fjTy5fe9qJLPo63Xb9_UHdb2g38eNWZKj_ey_lntFwXi3m-jMinIQreZtYDOmxahtZmgEhQG6DGngLIUsdtg4HAGV_bjE9FbZ0xqU_Ackhm6vlye8aoDtJ9k_xW_zjVGSf5A9zQRAw</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>SparseGAN: Sparse Generative Adversarial Network for Text Generation</title><source>arXiv.org</source><creator>Yuan, Liping ; Zeng, Jiehang ; Zheng, Xiaoqing</creator><creatorcontrib>Yuan, Liping ; Zeng, Jiehang ; Zheng, Xiaoqing</creatorcontrib><description>It is still a challenging task to learn a neural text generation model under
the framework of generative adversarial networks (GANs) since the entire
training process is not differentiable. The existing training strategies either
suffer from unreliable gradient estimations or imprecise sentence
representations. Inspired by the principle of sparse coding, we propose a
SparseGAN that generates semantic-interpretable, but sparse sentence
representations as inputs to the discriminator. The key idea is that we treat
an embedding matrix as an over-complete dictionary, and use a linear
combination of very few selected word embeddings to approximate the output
feature representation of the generator at each time step. With such
semantic-rich representations, we not only reduce unnecessary noises for
efficient adversarial training, but also make the entire training process fully
differentiable. Experiments on multiple text generation datasets yield
performance improvements, especially in sequence-level metrics, such as BLEU.</description><identifier>DOI: 10.48550/arxiv.2103.11578</identifier><language>eng</language><subject>Computer Science - Computation and Language</subject><creationdate>2021-03</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2103.11578$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2103.11578$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Yuan, Liping</creatorcontrib><creatorcontrib>Zeng, Jiehang</creatorcontrib><creatorcontrib>Zheng, Xiaoqing</creatorcontrib><title>SparseGAN: Sparse Generative Adversarial Network for Text Generation</title><description>It is still a challenging task to learn a neural text generation model under
the framework of generative adversarial networks (GANs) since the entire
training process is not differentiable. The existing training strategies either
suffer from unreliable gradient estimations or imprecise sentence
representations. Inspired by the principle of sparse coding, we propose a
SparseGAN that generates semantic-interpretable, but sparse sentence
representations as inputs to the discriminator. The key idea is that we treat
an embedding matrix as an over-complete dictionary, and use a linear
combination of very few selected word embeddings to approximate the output
feature representation of the generator at each time step. With such
semantic-rich representations, we not only reduce unnecessary noises for
efficient adversarial training, but also make the entire training process fully
differentiable. Experiments on multiple text generation datasets yield
performance improvements, especially in sequence-level metrics, such as BLEU.</description><subject>Computer Science - Computation and Language</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNo9j0FPg0AUhPfiwVR_gCf3D4DvLbvL4o1UxSZNeyh38oDXhFhL8yBY_721bXqaycxkkk-pJ4TYBufgheTYTbFBSGJEl4Z79bY5kAxc5KtXfbG64D0Ljd3EOm8nloGko51e8fjTy5fe9qJLPo63Xb9_UHdb2g38eNWZKj_ey_lntFwXi3m-jMinIQreZtYDOmxahtZmgEhQG6DGngLIUsdtg4HAGV_bjE9FbZ0xqU_Ackhm6vlye8aoDtJ9k_xW_zjVGSf5A9zQRAw</recordid><startdate>20210322</startdate><enddate>20210322</enddate><creator>Yuan, Liping</creator><creator>Zeng, Jiehang</creator><creator>Zheng, Xiaoqing</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20210322</creationdate><title>SparseGAN: Sparse Generative Adversarial Network for Text Generation</title><author>Yuan, Liping ; Zeng, Jiehang ; Zheng, Xiaoqing</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a678-8649460151cde0d49011a0b20ac4de00975edc18a0526b49e0acb452276304e83</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Computer Science - Computation and Language</topic><toplevel>online_resources</toplevel><creatorcontrib>Yuan, Liping</creatorcontrib><creatorcontrib>Zeng, Jiehang</creatorcontrib><creatorcontrib>Zheng, Xiaoqing</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Yuan, Liping</au><au>Zeng, Jiehang</au><au>Zheng, Xiaoqing</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>SparseGAN: Sparse Generative Adversarial Network for Text Generation</atitle><date>2021-03-22</date><risdate>2021</risdate><abstract>It is still a challenging task to learn a neural text generation model under
the framework of generative adversarial networks (GANs) since the entire
training process is not differentiable. The existing training strategies either
suffer from unreliable gradient estimations or imprecise sentence
representations. Inspired by the principle of sparse coding, we propose a
SparseGAN that generates semantic-interpretable, but sparse sentence
representations as inputs to the discriminator. The key idea is that we treat
an embedding matrix as an over-complete dictionary, and use a linear
combination of very few selected word embeddings to approximate the output
feature representation of the generator at each time step. With such
semantic-rich representations, we not only reduce unnecessary noises for
efficient adversarial training, but also make the entire training process fully
differentiable. Experiments on multiple text generation datasets yield
performance improvements, especially in sequence-level metrics, such as BLEU.</abstract><doi>10.48550/arxiv.2103.11578</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2103.11578 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2103_11578 |
source | arXiv.org |
subjects | Computer Science - Computation and Language |
title | SparseGAN: Sparse Generative Adversarial Network for Text Generation |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-03T09%3A03%3A22IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=SparseGAN:%20Sparse%20Generative%20Adversarial%20Network%20for%20Text%20Generation&rft.au=Yuan,%20Liping&rft.date=2021-03-22&rft_id=info:doi/10.48550/arxiv.2103.11578&rft_dat=%3Carxiv_GOX%3E2103_11578%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |