Knowledge Base Completion: Baseline strikes back (Again)

Knowledge Base Completion (KBC) has been a very active area lately. Several recent KBCpapers propose architectural changes, new training methods, or even new formulations. KBC systems are usually evaluated on standard benchmark datasets: FB15k, FB15k-237, WN18, WN18RR, and Yago3-10. Most existing me...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Jain, Prachi, Rathi, Sushant, Mausam, Chakrabarti, Soumen
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Jain, Prachi
Rathi, Sushant
Mausam
Chakrabarti, Soumen
description Knowledge Base Completion (KBC) has been a very active area lately. Several recent KBCpapers propose architectural changes, new training methods, or even new formulations. KBC systems are usually evaluated on standard benchmark datasets: FB15k, FB15k-237, WN18, WN18RR, and Yago3-10. Most existing methods train with a small number of negative samples for each positive instance in these datasets to save computational costs. This paper discusses how recent developments allow us to use all available negative samples for training. We show that Complex, when trained using all available negative samples, gives near state-of-the-art performance on all the datasets. We call this approach COMPLEX-V2. We also highlight how various multiplicative KBC methods, recently proposed in the literature, benefit from this train-ing regime and become indistinguishable in terms of performance on most datasets. Our work calls for a reassessment of their individual value, in light of these findings.
doi_str_mv 10.48550/arxiv.2005.00804
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2005_00804</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2005_00804</sourcerecordid><originalsourceid>FETCH-LOGICAL-a674-42955f359bd39dc989df2f782eb36b514e1c3161155a0fd3b3edebf8936f25d73</originalsourceid><addsrcrecordid>eNotj71uwjAURr0wVJQH6ITHdkhq-_omNhtE_RNIXdgjO75GFiFBCYL27aumnY70DUffYexBilwbRPHshq90zZUQmAthhL5jZtv1t5bCgfjGjcSr_nRu6ZL6bjUNbeqIj5chHWnk3jVH_rg-uNQ93bNZdO1Ii3_O2f71ZV-9Z7vPt49qvctcUepMK4sYAa0PYENjjQ1RxdIo8lB4lJpkA7KQEtGJGMADBfLRWCiiwlDCnC3_tNP1-jykkxu-69-EekqAH_rIP9g</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Knowledge Base Completion: Baseline strikes back (Again)</title><source>arXiv.org</source><creator>Jain, Prachi ; Rathi, Sushant ; Mausam ; Chakrabarti, Soumen</creator><creatorcontrib>Jain, Prachi ; Rathi, Sushant ; Mausam ; Chakrabarti, Soumen</creatorcontrib><description>Knowledge Base Completion (KBC) has been a very active area lately. Several recent KBCpapers propose architectural changes, new training methods, or even new formulations. KBC systems are usually evaluated on standard benchmark datasets: FB15k, FB15k-237, WN18, WN18RR, and Yago3-10. Most existing methods train with a small number of negative samples for each positive instance in these datasets to save computational costs. This paper discusses how recent developments allow us to use all available negative samples for training. We show that Complex, when trained using all available negative samples, gives near state-of-the-art performance on all the datasets. We call this approach COMPLEX-V2. We also highlight how various multiplicative KBC methods, recently proposed in the literature, benefit from this train-ing regime and become indistinguishable in terms of performance on most datasets. Our work calls for a reassessment of their individual value, in light of these findings.</description><identifier>DOI: 10.48550/arxiv.2005.00804</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Learning</subject><creationdate>2020-05</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2005.00804$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2005.00804$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Jain, Prachi</creatorcontrib><creatorcontrib>Rathi, Sushant</creatorcontrib><creatorcontrib>Mausam</creatorcontrib><creatorcontrib>Chakrabarti, Soumen</creatorcontrib><title>Knowledge Base Completion: Baseline strikes back (Again)</title><description>Knowledge Base Completion (KBC) has been a very active area lately. Several recent KBCpapers propose architectural changes, new training methods, or even new formulations. KBC systems are usually evaluated on standard benchmark datasets: FB15k, FB15k-237, WN18, WN18RR, and Yago3-10. Most existing methods train with a small number of negative samples for each positive instance in these datasets to save computational costs. This paper discusses how recent developments allow us to use all available negative samples for training. We show that Complex, when trained using all available negative samples, gives near state-of-the-art performance on all the datasets. We call this approach COMPLEX-V2. We also highlight how various multiplicative KBC methods, recently proposed in the literature, benefit from this train-ing regime and become indistinguishable in terms of performance on most datasets. Our work calls for a reassessment of their individual value, in light of these findings.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj71uwjAURr0wVJQH6ITHdkhq-_omNhtE_RNIXdgjO75GFiFBCYL27aumnY70DUffYexBilwbRPHshq90zZUQmAthhL5jZtv1t5bCgfjGjcSr_nRu6ZL6bjUNbeqIj5chHWnk3jVH_rg-uNQ93bNZdO1Ii3_O2f71ZV-9Z7vPt49qvctcUepMK4sYAa0PYENjjQ1RxdIo8lB4lJpkA7KQEtGJGMADBfLRWCiiwlDCnC3_tNP1-jykkxu-69-EekqAH_rIP9g</recordid><startdate>20200502</startdate><enddate>20200502</enddate><creator>Jain, Prachi</creator><creator>Rathi, Sushant</creator><creator>Mausam</creator><creator>Chakrabarti, Soumen</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20200502</creationdate><title>Knowledge Base Completion: Baseline strikes back (Again)</title><author>Jain, Prachi ; Rathi, Sushant ; Mausam ; Chakrabarti, Soumen</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a674-42955f359bd39dc989df2f782eb36b514e1c3161155a0fd3b3edebf8936f25d73</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Jain, Prachi</creatorcontrib><creatorcontrib>Rathi, Sushant</creatorcontrib><creatorcontrib>Mausam</creatorcontrib><creatorcontrib>Chakrabarti, Soumen</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Jain, Prachi</au><au>Rathi, Sushant</au><au>Mausam</au><au>Chakrabarti, Soumen</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Knowledge Base Completion: Baseline strikes back (Again)</atitle><date>2020-05-02</date><risdate>2020</risdate><abstract>Knowledge Base Completion (KBC) has been a very active area lately. Several recent KBCpapers propose architectural changes, new training methods, or even new formulations. KBC systems are usually evaluated on standard benchmark datasets: FB15k, FB15k-237, WN18, WN18RR, and Yago3-10. Most existing methods train with a small number of negative samples for each positive instance in these datasets to save computational costs. This paper discusses how recent developments allow us to use all available negative samples for training. We show that Complex, when trained using all available negative samples, gives near state-of-the-art performance on all the datasets. We call this approach COMPLEX-V2. We also highlight how various multiplicative KBC methods, recently proposed in the literature, benefit from this train-ing regime and become indistinguishable in terms of performance on most datasets. Our work calls for a reassessment of their individual value, in light of these findings.</abstract><doi>10.48550/arxiv.2005.00804</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2005.00804
ispartof
issn
language eng
recordid cdi_arxiv_primary_2005_00804
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Learning
title Knowledge Base Completion: Baseline strikes back (Again)
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-28T20%3A35%3A47IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Knowledge%20Base%20Completion:%20Baseline%20strikes%20back%20(Again)&rft.au=Jain,%20Prachi&rft.date=2020-05-02&rft_id=info:doi/10.48550/arxiv.2005.00804&rft_dat=%3Carxiv_GOX%3E2005_00804%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true