Distributional Inclusion Vector Embedding for Unsupervised Hypernymy Detection

Modeling hypernymy, such as poodle is-a dog, is an important generalization aid to many NLP tasks, such as entailment, coreference, relation extraction, and question answering. Supervised learning from labeled hypernym sources, such as WordNet, limits the coverage of these models, which can be addre...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Chang, Haw-Shiuan, Wang, ZiYun, Vilnis, Luke, McCallum, Andrew
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Chang, Haw-Shiuan
Wang, ZiYun
Vilnis, Luke
McCallum, Andrew
description Modeling hypernymy, such as poodle is-a dog, is an important generalization aid to many NLP tasks, such as entailment, coreference, relation extraction, and question answering. Supervised learning from labeled hypernym sources, such as WordNet, limits the coverage of these models, which can be addressed by learning hypernyms from unlabeled text. Existing unsupervised methods either do not scale to large vocabularies or yield unacceptably poor accuracy. This paper introduces distributional inclusion vector embedding (DIVE), a simple-to-implement unsupervised method of hypernym discovery via per-word non-negative vector embeddings which preserve the inclusion property of word contexts in a low-dimensional and interpretable space. In experimental evaluations more comprehensive than any previous literature of which we are aware-evaluating on 11 datasets using multiple existing as well as newly proposed scoring functions-we find that our method provides up to double the precision of previous unsupervised embeddings, and the highest average performance, using a much more compact word representation, and yielding many new state-of-the-art results.
doi_str_mv 10.48550/arxiv.1710.00880
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_1710_00880</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1710_00880</sourcerecordid><originalsourceid>FETCH-LOGICAL-a670-a547364184ee9de7e214335c40e60c35384853d42b06b2fe09a1d935a3ff12a03</originalsourceid><addsrcrecordid>eNotj81OwzAMx3PhgAYPwIm8QIdTJ216RNtgkya4jF2rtHGmSG02pe1E354wOP0_ZFv-MfYkYCm1UvBi4re_LkWZCgCt4Z59rP0wRt9Moz8H0_FdaLtpSJ4fqR3PkW_6hqz14cRdSl9hmC4Ur34gy7dzsmHuZ76mMU2nrQd250w30OO_LtjhbXNYbbP95_tu9brPTFFCZpQssZBCS6LKUkm5kIiqlUAFtKhQp3fRyryBoskdQWWErVAZdE7kBnDBnv_O3oDqS_S9iXP9C1bfwPAHFuRJEg</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Distributional Inclusion Vector Embedding for Unsupervised Hypernymy Detection</title><source>arXiv.org</source><creator>Chang, Haw-Shiuan ; Wang, ZiYun ; Vilnis, Luke ; McCallum, Andrew</creator><creatorcontrib>Chang, Haw-Shiuan ; Wang, ZiYun ; Vilnis, Luke ; McCallum, Andrew</creatorcontrib><description>Modeling hypernymy, such as poodle is-a dog, is an important generalization aid to many NLP tasks, such as entailment, coreference, relation extraction, and question answering. Supervised learning from labeled hypernym sources, such as WordNet, limits the coverage of these models, which can be addressed by learning hypernyms from unlabeled text. Existing unsupervised methods either do not scale to large vocabularies or yield unacceptably poor accuracy. This paper introduces distributional inclusion vector embedding (DIVE), a simple-to-implement unsupervised method of hypernym discovery via per-word non-negative vector embeddings which preserve the inclusion property of word contexts in a low-dimensional and interpretable space. In experimental evaluations more comprehensive than any previous literature of which we are aware-evaluating on 11 datasets using multiple existing as well as newly proposed scoring functions-we find that our method provides up to double the precision of previous unsupervised embeddings, and the highest average performance, using a much more compact word representation, and yielding many new state-of-the-art results.</description><identifier>DOI: 10.48550/arxiv.1710.00880</identifier><language>eng</language><subject>Computer Science - Computation and Language</subject><creationdate>2017-10</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/1710.00880$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.1710.00880$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Chang, Haw-Shiuan</creatorcontrib><creatorcontrib>Wang, ZiYun</creatorcontrib><creatorcontrib>Vilnis, Luke</creatorcontrib><creatorcontrib>McCallum, Andrew</creatorcontrib><title>Distributional Inclusion Vector Embedding for Unsupervised Hypernymy Detection</title><description>Modeling hypernymy, such as poodle is-a dog, is an important generalization aid to many NLP tasks, such as entailment, coreference, relation extraction, and question answering. Supervised learning from labeled hypernym sources, such as WordNet, limits the coverage of these models, which can be addressed by learning hypernyms from unlabeled text. Existing unsupervised methods either do not scale to large vocabularies or yield unacceptably poor accuracy. This paper introduces distributional inclusion vector embedding (DIVE), a simple-to-implement unsupervised method of hypernym discovery via per-word non-negative vector embeddings which preserve the inclusion property of word contexts in a low-dimensional and interpretable space. In experimental evaluations more comprehensive than any previous literature of which we are aware-evaluating on 11 datasets using multiple existing as well as newly proposed scoring functions-we find that our method provides up to double the precision of previous unsupervised embeddings, and the highest average performance, using a much more compact word representation, and yielding many new state-of-the-art results.</description><subject>Computer Science - Computation and Language</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2017</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj81OwzAMx3PhgAYPwIm8QIdTJ216RNtgkya4jF2rtHGmSG02pe1E354wOP0_ZFv-MfYkYCm1UvBi4re_LkWZCgCt4Z59rP0wRt9Moz8H0_FdaLtpSJ4fqR3PkW_6hqz14cRdSl9hmC4Ur34gy7dzsmHuZ76mMU2nrQd250w30OO_LtjhbXNYbbP95_tu9brPTFFCZpQssZBCS6LKUkm5kIiqlUAFtKhQp3fRyryBoskdQWWErVAZdE7kBnDBnv_O3oDqS_S9iXP9C1bfwPAHFuRJEg</recordid><startdate>20171002</startdate><enddate>20171002</enddate><creator>Chang, Haw-Shiuan</creator><creator>Wang, ZiYun</creator><creator>Vilnis, Luke</creator><creator>McCallum, Andrew</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20171002</creationdate><title>Distributional Inclusion Vector Embedding for Unsupervised Hypernymy Detection</title><author>Chang, Haw-Shiuan ; Wang, ZiYun ; Vilnis, Luke ; McCallum, Andrew</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a670-a547364184ee9de7e214335c40e60c35384853d42b06b2fe09a1d935a3ff12a03</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2017</creationdate><topic>Computer Science - Computation and Language</topic><toplevel>online_resources</toplevel><creatorcontrib>Chang, Haw-Shiuan</creatorcontrib><creatorcontrib>Wang, ZiYun</creatorcontrib><creatorcontrib>Vilnis, Luke</creatorcontrib><creatorcontrib>McCallum, Andrew</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Chang, Haw-Shiuan</au><au>Wang, ZiYun</au><au>Vilnis, Luke</au><au>McCallum, Andrew</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Distributional Inclusion Vector Embedding for Unsupervised Hypernymy Detection</atitle><date>2017-10-02</date><risdate>2017</risdate><abstract>Modeling hypernymy, such as poodle is-a dog, is an important generalization aid to many NLP tasks, such as entailment, coreference, relation extraction, and question answering. Supervised learning from labeled hypernym sources, such as WordNet, limits the coverage of these models, which can be addressed by learning hypernyms from unlabeled text. Existing unsupervised methods either do not scale to large vocabularies or yield unacceptably poor accuracy. This paper introduces distributional inclusion vector embedding (DIVE), a simple-to-implement unsupervised method of hypernym discovery via per-word non-negative vector embeddings which preserve the inclusion property of word contexts in a low-dimensional and interpretable space. In experimental evaluations more comprehensive than any previous literature of which we are aware-evaluating on 11 datasets using multiple existing as well as newly proposed scoring functions-we find that our method provides up to double the precision of previous unsupervised embeddings, and the highest average performance, using a much more compact word representation, and yielding many new state-of-the-art results.</abstract><doi>10.48550/arxiv.1710.00880</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.1710.00880
ispartof
issn
language eng
recordid cdi_arxiv_primary_1710_00880
source arXiv.org
subjects Computer Science - Computation and Language
title Distributional Inclusion Vector Embedding for Unsupervised Hypernymy Detection
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-25T10%3A29%3A57IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Distributional%20Inclusion%20Vector%20Embedding%20for%20Unsupervised%20Hypernymy%20Detection&rft.au=Chang,%20Haw-Shiuan&rft.date=2017-10-02&rft_id=info:doi/10.48550/arxiv.1710.00880&rft_dat=%3Carxiv_GOX%3E1710_00880%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true