JAKET: Joint Pre-training of Knowledge Graph and Language Understanding

Knowledge graphs (KGs) contain rich information about world knowledge, entities and relations. Thus, they can be great supplements to existing pre-trained language models. However, it remains a challenge to efficiently integrate information from KG into language modeling. And the understanding of a...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Yu, Donghan, Zhu, Chenguang, Yang, Yiming, Zeng, Michael
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Yu, Donghan
Zhu, Chenguang
Yang, Yiming
Zeng, Michael
description Knowledge graphs (KGs) contain rich information about world knowledge, entities and relations. Thus, they can be great supplements to existing pre-trained language models. However, it remains a challenge to efficiently integrate information from KG into language modeling. And the understanding of a knowledge graph requires related context. We propose a novel joint pre-training framework, JAKET, to model both the knowledge graph and language. The knowledge module and language module provide essential information to mutually assist each other: the knowledge module produces embeddings for entities in text while the language module generates context-aware initial embeddings for entities and relations in the graph. Our design enables the pre-trained model to easily adapt to unseen knowledge graphs in new domains. Experimental results on several knowledge-aware NLP tasks show that our proposed framework achieves superior performance by effectively leveraging knowledge in language understanding.
doi_str_mv 10.48550/arxiv.2010.00796
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2010_00796</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2010_00796</sourcerecordid><originalsourceid>FETCH-LOGICAL-a676-a6514a97b07555b8f25f359d37cac5945346c078775b55e82d43535a87ff63d03</originalsourceid><addsrcrecordid>eNotj0FuwjAURL1hUdEeoKv6AqEm9red7hCioRCpXYR19BPbwRI4yKQtvX1d6GZGehqN9Ah5nLOZ0ADsGePFf81ylgBjqpB3pNwstqv6hW4GH0b6EW02RvTBh54Ojm7D8H2wpre0jHjaUwyGVhj6T0xoF4yN5zGxtL4nE4eHs3347ympX1f1cp1V7-XbclFlKJVMAXOBhWqZAoBWuxwch8Jw1WEHhQAuZMeUVgpaAKtzIzhwQK2ck9wwPiVPt9urSXOK_ojxp_kzaq5G_BctfUQz</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>JAKET: Joint Pre-training of Knowledge Graph and Language Understanding</title><source>arXiv.org</source><creator>Yu, Donghan ; Zhu, Chenguang ; Yang, Yiming ; Zeng, Michael</creator><creatorcontrib>Yu, Donghan ; Zhu, Chenguang ; Yang, Yiming ; Zeng, Michael</creatorcontrib><description>Knowledge graphs (KGs) contain rich information about world knowledge, entities and relations. Thus, they can be great supplements to existing pre-trained language models. However, it remains a challenge to efficiently integrate information from KG into language modeling. And the understanding of a knowledge graph requires related context. We propose a novel joint pre-training framework, JAKET, to model both the knowledge graph and language. The knowledge module and language module provide essential information to mutually assist each other: the knowledge module produces embeddings for entities in text while the language module generates context-aware initial embeddings for entities and relations in the graph. Our design enables the pre-trained model to easily adapt to unseen knowledge graphs in new domains. Experimental results on several knowledge-aware NLP tasks show that our proposed framework achieves superior performance by effectively leveraging knowledge in language understanding.</description><identifier>DOI: 10.48550/arxiv.2010.00796</identifier><language>eng</language><subject>Computer Science - Computation and Language</subject><creationdate>2020-10</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2010.00796$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2010.00796$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Yu, Donghan</creatorcontrib><creatorcontrib>Zhu, Chenguang</creatorcontrib><creatorcontrib>Yang, Yiming</creatorcontrib><creatorcontrib>Zeng, Michael</creatorcontrib><title>JAKET: Joint Pre-training of Knowledge Graph and Language Understanding</title><description>Knowledge graphs (KGs) contain rich information about world knowledge, entities and relations. Thus, they can be great supplements to existing pre-trained language models. However, it remains a challenge to efficiently integrate information from KG into language modeling. And the understanding of a knowledge graph requires related context. We propose a novel joint pre-training framework, JAKET, to model both the knowledge graph and language. The knowledge module and language module provide essential information to mutually assist each other: the knowledge module produces embeddings for entities in text while the language module generates context-aware initial embeddings for entities and relations in the graph. Our design enables the pre-trained model to easily adapt to unseen knowledge graphs in new domains. Experimental results on several knowledge-aware NLP tasks show that our proposed framework achieves superior performance by effectively leveraging knowledge in language understanding.</description><subject>Computer Science - Computation and Language</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj0FuwjAURL1hUdEeoKv6AqEm9red7hCioRCpXYR19BPbwRI4yKQtvX1d6GZGehqN9Ah5nLOZ0ADsGePFf81ylgBjqpB3pNwstqv6hW4GH0b6EW02RvTBh54Ojm7D8H2wpre0jHjaUwyGVhj6T0xoF4yN5zGxtL4nE4eHs3347ympX1f1cp1V7-XbclFlKJVMAXOBhWqZAoBWuxwch8Jw1WEHhQAuZMeUVgpaAKtzIzhwQK2ck9wwPiVPt9urSXOK_ojxp_kzaq5G_BctfUQz</recordid><startdate>20201002</startdate><enddate>20201002</enddate><creator>Yu, Donghan</creator><creator>Zhu, Chenguang</creator><creator>Yang, Yiming</creator><creator>Zeng, Michael</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20201002</creationdate><title>JAKET: Joint Pre-training of Knowledge Graph and Language Understanding</title><author>Yu, Donghan ; Zhu, Chenguang ; Yang, Yiming ; Zeng, Michael</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a676-a6514a97b07555b8f25f359d37cac5945346c078775b55e82d43535a87ff63d03</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Computer Science - Computation and Language</topic><toplevel>online_resources</toplevel><creatorcontrib>Yu, Donghan</creatorcontrib><creatorcontrib>Zhu, Chenguang</creatorcontrib><creatorcontrib>Yang, Yiming</creatorcontrib><creatorcontrib>Zeng, Michael</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Yu, Donghan</au><au>Zhu, Chenguang</au><au>Yang, Yiming</au><au>Zeng, Michael</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>JAKET: Joint Pre-training of Knowledge Graph and Language Understanding</atitle><date>2020-10-02</date><risdate>2020</risdate><abstract>Knowledge graphs (KGs) contain rich information about world knowledge, entities and relations. Thus, they can be great supplements to existing pre-trained language models. However, it remains a challenge to efficiently integrate information from KG into language modeling. And the understanding of a knowledge graph requires related context. We propose a novel joint pre-training framework, JAKET, to model both the knowledge graph and language. The knowledge module and language module provide essential information to mutually assist each other: the knowledge module produces embeddings for entities in text while the language module generates context-aware initial embeddings for entities and relations in the graph. Our design enables the pre-trained model to easily adapt to unseen knowledge graphs in new domains. Experimental results on several knowledge-aware NLP tasks show that our proposed framework achieves superior performance by effectively leveraging knowledge in language understanding.</abstract><doi>10.48550/arxiv.2010.00796</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2010.00796
ispartof
issn
language eng
recordid cdi_arxiv_primary_2010_00796
source arXiv.org
subjects Computer Science - Computation and Language
title JAKET: Joint Pre-training of Knowledge Graph and Language Understanding
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-06T06%3A31%3A59IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=JAKET:%20Joint%20Pre-training%20of%20Knowledge%20Graph%20and%20Language%20Understanding&rft.au=Yu,%20Donghan&rft.date=2020-10-02&rft_id=info:doi/10.48550/arxiv.2010.00796&rft_dat=%3Carxiv_GOX%3E2010_00796%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true