Syntax-Aware Language Modeling with Recurrent Neural Networks

Neural language models (LMs) are typically trained using only lexical features, such as surface forms of words. In this paper, we argue this deprives the LM of crucial syntactic signals that can be detected at high confidence using existing parsers. We present a simple but highly effective approach...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Blythe, Duncan, Akbik, Alan, Vollgraf, Roland
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Blythe, Duncan
Akbik, Alan
Vollgraf, Roland
description Neural language models (LMs) are typically trained using only lexical features, such as surface forms of words. In this paper, we argue this deprives the LM of crucial syntactic signals that can be detected at high confidence using existing parsers. We present a simple but highly effective approach for training neural LMs using both lexical and syntactic information, and a novel approach for applying such LMs to unparsed text using sequential Monte Carlo sampling. In experiments on a range of corpora and corpus sizes, we show our approach consistently outperforms standard lexical LMs in character-level language modeling; on the other hand, for word-level models the models are on a par with standard language models. These results indicate potential for expanding LMs beyond lexical surface features to higher-level NLP features for character-level models.
doi_str_mv 10.48550/arxiv.1803.03665
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_1803_03665</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1803_03665</sourcerecordid><originalsourceid>FETCH-LOGICAL-a675-27affe7cbf7f44193edaab292d3da113fe277e2af8b626fb6d310745275dfa193</originalsourceid><addsrcrecordid>eNotj71uwjAURr0wVNAH6FS_QIJ_YjsMHRCCtlIAibJHN_jeNCIEZJIG3r6UMp3l06dzGHuRIk5SY8QYwqX6iWUqdCy0teaJvX1dmxYu0bSHgDyDpuygRL48eqyrpuR91X7zDe66ELBp-Qq7APUNbX8M-_OIDQjqMz4_OGTbxXw7-4iy9fvnbJpFYJ2JlAMidLuCHCWJnGj0AIWaKK89SKkJlXOogNLCKkuF9VoKlxjljCe47Yfs9f_27p-fQnWAcM3_OvJ7h_4FcMJDeA</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Syntax-Aware Language Modeling with Recurrent Neural Networks</title><source>arXiv.org</source><creator>Blythe, Duncan ; Akbik, Alan ; Vollgraf, Roland</creator><creatorcontrib>Blythe, Duncan ; Akbik, Alan ; Vollgraf, Roland</creatorcontrib><description>Neural language models (LMs) are typically trained using only lexical features, such as surface forms of words. In this paper, we argue this deprives the LM of crucial syntactic signals that can be detected at high confidence using existing parsers. We present a simple but highly effective approach for training neural LMs using both lexical and syntactic information, and a novel approach for applying such LMs to unparsed text using sequential Monte Carlo sampling. In experiments on a range of corpora and corpus sizes, we show our approach consistently outperforms standard lexical LMs in character-level language modeling; on the other hand, for word-level models the models are on a par with standard language models. These results indicate potential for expanding LMs beyond lexical surface features to higher-level NLP features for character-level models.</description><identifier>DOI: 10.48550/arxiv.1803.03665</identifier><language>eng</language><subject>Computer Science - Computation and Language ; Computer Science - Learning</subject><creationdate>2018-03</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/1803.03665$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.1803.03665$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Blythe, Duncan</creatorcontrib><creatorcontrib>Akbik, Alan</creatorcontrib><creatorcontrib>Vollgraf, Roland</creatorcontrib><title>Syntax-Aware Language Modeling with Recurrent Neural Networks</title><description>Neural language models (LMs) are typically trained using only lexical features, such as surface forms of words. In this paper, we argue this deprives the LM of crucial syntactic signals that can be detected at high confidence using existing parsers. We present a simple but highly effective approach for training neural LMs using both lexical and syntactic information, and a novel approach for applying such LMs to unparsed text using sequential Monte Carlo sampling. In experiments on a range of corpora and corpus sizes, we show our approach consistently outperforms standard lexical LMs in character-level language modeling; on the other hand, for word-level models the models are on a par with standard language models. These results indicate potential for expanding LMs beyond lexical surface features to higher-level NLP features for character-level models.</description><subject>Computer Science - Computation and Language</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2018</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj71uwjAURr0wVNAH6FS_QIJ_YjsMHRCCtlIAibJHN_jeNCIEZJIG3r6UMp3l06dzGHuRIk5SY8QYwqX6iWUqdCy0teaJvX1dmxYu0bSHgDyDpuygRL48eqyrpuR91X7zDe66ELBp-Qq7APUNbX8M-_OIDQjqMz4_OGTbxXw7-4iy9fvnbJpFYJ2JlAMidLuCHCWJnGj0AIWaKK89SKkJlXOogNLCKkuF9VoKlxjljCe47Yfs9f_27p-fQnWAcM3_OvJ7h_4FcMJDeA</recordid><startdate>20180302</startdate><enddate>20180302</enddate><creator>Blythe, Duncan</creator><creator>Akbik, Alan</creator><creator>Vollgraf, Roland</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20180302</creationdate><title>Syntax-Aware Language Modeling with Recurrent Neural Networks</title><author>Blythe, Duncan ; Akbik, Alan ; Vollgraf, Roland</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a675-27affe7cbf7f44193edaab292d3da113fe277e2af8b626fb6d310745275dfa193</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2018</creationdate><topic>Computer Science - Computation and Language</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Blythe, Duncan</creatorcontrib><creatorcontrib>Akbik, Alan</creatorcontrib><creatorcontrib>Vollgraf, Roland</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Blythe, Duncan</au><au>Akbik, Alan</au><au>Vollgraf, Roland</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Syntax-Aware Language Modeling with Recurrent Neural Networks</atitle><date>2018-03-02</date><risdate>2018</risdate><abstract>Neural language models (LMs) are typically trained using only lexical features, such as surface forms of words. In this paper, we argue this deprives the LM of crucial syntactic signals that can be detected at high confidence using existing parsers. We present a simple but highly effective approach for training neural LMs using both lexical and syntactic information, and a novel approach for applying such LMs to unparsed text using sequential Monte Carlo sampling. In experiments on a range of corpora and corpus sizes, we show our approach consistently outperforms standard lexical LMs in character-level language modeling; on the other hand, for word-level models the models are on a par with standard language models. These results indicate potential for expanding LMs beyond lexical surface features to higher-level NLP features for character-level models.</abstract><doi>10.48550/arxiv.1803.03665</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.1803.03665
ispartof
issn
language eng
recordid cdi_arxiv_primary_1803_03665
source arXiv.org
subjects Computer Science - Computation and Language
Computer Science - Learning
title Syntax-Aware Language Modeling with Recurrent Neural Networks
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-08T05%3A48%3A03IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Syntax-Aware%20Language%20Modeling%20with%20Recurrent%20Neural%20Networks&rft.au=Blythe,%20Duncan&rft.date=2018-03-02&rft_id=info:doi/10.48550/arxiv.1803.03665&rft_dat=%3Carxiv_GOX%3E1803_03665%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true