Porous Lattice-based Transformer Encoder for Chinese NER

COLING 2020 Incorporating lattices into character-level Chinese named entity recognition is an effective method to exploit explicit word information. Recent works extend recurrent and convolutional neural networks to model lattice inputs. However, due to the DAG structure or the variable-sized poten...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Mengge, Xue, Bowen, Yu, Tingwen, Liu, Yue, Zhang, Erli, Meng, Bin, Wang
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Mengge, Xue
Bowen, Yu
Tingwen, Liu
Yue, Zhang
Erli, Meng
Bin, Wang
description COLING 2020 Incorporating lattices into character-level Chinese named entity recognition is an effective method to exploit explicit word information. Recent works extend recurrent and convolutional neural networks to model lattice inputs. However, due to the DAG structure or the variable-sized potential word set for lattice inputs, these models prevent the convenient use of batched computation, resulting in serious inefficient. In this paper, we propose a porous lattice-based transformer encoder for Chinese named entity recognition, which is capable to better exploit the GPU parallelism and batch the computation owing to the mask mechanism in transformer. We first investigate the lattice-aware self-attention coupled with relative position representations to explore effective word information in the lattice structure. Besides, to strengthen the local dependencies among neighboring tokens, we propose a novel porous structure during self-attentional computation processing, in which every two non-neighboring tokens are connected through a shared pivot node. Experimental results on four datasets show that our model performs up to 9.47 times faster than state-of-the-art models, while is roughly on a par with its performance. The source code of this paper can be obtained from https://github.com/xxx/xxx.
doi_str_mv 10.48550/arxiv.1911.02733
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_1911_02733</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1911_02733</sourcerecordid><originalsourceid>FETCH-LOGICAL-a673-fad7f24af9e486acf091ec3ba06c20aaa8e26b43f5bda72853c29d2cee2a4f1b3</originalsourceid><addsrcrecordid>eNotj82KwjAURrNxMagPMCvzAq35a5supVRHKKNI9-UmucGCtpKoOG8_js7q8G0-ziHkk7NU6SxjSwiP_p7ykvOUiULKD6L3YxhvkTZwvfYWEwMRHW0DDNGP4YyB1oMd3ZPPSatjP2BE-l0fZmTi4RRx_s8padd1W30lzW6zrVZNAnkhEw-u8EKBL1HpHKxnJUcrDbDcCgYAGkVulPSZcVAInUkrSicsogDluZFTsnjfvtS7S-jPEH66v4TulSB_AapTQZk</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Porous Lattice-based Transformer Encoder for Chinese NER</title><source>arXiv.org</source><creator>Mengge, Xue ; Bowen, Yu ; Tingwen, Liu ; Yue, Zhang ; Erli, Meng ; Bin, Wang</creator><creatorcontrib>Mengge, Xue ; Bowen, Yu ; Tingwen, Liu ; Yue, Zhang ; Erli, Meng ; Bin, Wang</creatorcontrib><description>COLING 2020 Incorporating lattices into character-level Chinese named entity recognition is an effective method to exploit explicit word information. Recent works extend recurrent and convolutional neural networks to model lattice inputs. However, due to the DAG structure or the variable-sized potential word set for lattice inputs, these models prevent the convenient use of batched computation, resulting in serious inefficient. In this paper, we propose a porous lattice-based transformer encoder for Chinese named entity recognition, which is capable to better exploit the GPU parallelism and batch the computation owing to the mask mechanism in transformer. We first investigate the lattice-aware self-attention coupled with relative position representations to explore effective word information in the lattice structure. Besides, to strengthen the local dependencies among neighboring tokens, we propose a novel porous structure during self-attentional computation processing, in which every two non-neighboring tokens are connected through a shared pivot node. Experimental results on four datasets show that our model performs up to 9.47 times faster than state-of-the-art models, while is roughly on a par with its performance. The source code of this paper can be obtained from https://github.com/xxx/xxx.</description><identifier>DOI: 10.48550/arxiv.1911.02733</identifier><language>eng</language><subject>Computer Science - Computation and Language ; Computer Science - Information Retrieval</subject><creationdate>2019-11</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,781,886</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/1911.02733$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.1911.02733$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Mengge, Xue</creatorcontrib><creatorcontrib>Bowen, Yu</creatorcontrib><creatorcontrib>Tingwen, Liu</creatorcontrib><creatorcontrib>Yue, Zhang</creatorcontrib><creatorcontrib>Erli, Meng</creatorcontrib><creatorcontrib>Bin, Wang</creatorcontrib><title>Porous Lattice-based Transformer Encoder for Chinese NER</title><description>COLING 2020 Incorporating lattices into character-level Chinese named entity recognition is an effective method to exploit explicit word information. Recent works extend recurrent and convolutional neural networks to model lattice inputs. However, due to the DAG structure or the variable-sized potential word set for lattice inputs, these models prevent the convenient use of batched computation, resulting in serious inefficient. In this paper, we propose a porous lattice-based transformer encoder for Chinese named entity recognition, which is capable to better exploit the GPU parallelism and batch the computation owing to the mask mechanism in transformer. We first investigate the lattice-aware self-attention coupled with relative position representations to explore effective word information in the lattice structure. Besides, to strengthen the local dependencies among neighboring tokens, we propose a novel porous structure during self-attentional computation processing, in which every two non-neighboring tokens are connected through a shared pivot node. Experimental results on four datasets show that our model performs up to 9.47 times faster than state-of-the-art models, while is roughly on a par with its performance. The source code of this paper can be obtained from https://github.com/xxx/xxx.</description><subject>Computer Science - Computation and Language</subject><subject>Computer Science - Information Retrieval</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2019</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj82KwjAURrNxMagPMCvzAq35a5supVRHKKNI9-UmucGCtpKoOG8_js7q8G0-ziHkk7NU6SxjSwiP_p7ykvOUiULKD6L3YxhvkTZwvfYWEwMRHW0DDNGP4YyB1oMd3ZPPSatjP2BE-l0fZmTi4RRx_s8padd1W30lzW6zrVZNAnkhEw-u8EKBL1HpHKxnJUcrDbDcCgYAGkVulPSZcVAInUkrSicsogDluZFTsnjfvtS7S-jPEH66v4TulSB_AapTQZk</recordid><startdate>20191106</startdate><enddate>20191106</enddate><creator>Mengge, Xue</creator><creator>Bowen, Yu</creator><creator>Tingwen, Liu</creator><creator>Yue, Zhang</creator><creator>Erli, Meng</creator><creator>Bin, Wang</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20191106</creationdate><title>Porous Lattice-based Transformer Encoder for Chinese NER</title><author>Mengge, Xue ; Bowen, Yu ; Tingwen, Liu ; Yue, Zhang ; Erli, Meng ; Bin, Wang</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a673-fad7f24af9e486acf091ec3ba06c20aaa8e26b43f5bda72853c29d2cee2a4f1b3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2019</creationdate><topic>Computer Science - Computation and Language</topic><topic>Computer Science - Information Retrieval</topic><toplevel>online_resources</toplevel><creatorcontrib>Mengge, Xue</creatorcontrib><creatorcontrib>Bowen, Yu</creatorcontrib><creatorcontrib>Tingwen, Liu</creatorcontrib><creatorcontrib>Yue, Zhang</creatorcontrib><creatorcontrib>Erli, Meng</creatorcontrib><creatorcontrib>Bin, Wang</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Mengge, Xue</au><au>Bowen, Yu</au><au>Tingwen, Liu</au><au>Yue, Zhang</au><au>Erli, Meng</au><au>Bin, Wang</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Porous Lattice-based Transformer Encoder for Chinese NER</atitle><date>2019-11-06</date><risdate>2019</risdate><abstract>COLING 2020 Incorporating lattices into character-level Chinese named entity recognition is an effective method to exploit explicit word information. Recent works extend recurrent and convolutional neural networks to model lattice inputs. However, due to the DAG structure or the variable-sized potential word set for lattice inputs, these models prevent the convenient use of batched computation, resulting in serious inefficient. In this paper, we propose a porous lattice-based transformer encoder for Chinese named entity recognition, which is capable to better exploit the GPU parallelism and batch the computation owing to the mask mechanism in transformer. We first investigate the lattice-aware self-attention coupled with relative position representations to explore effective word information in the lattice structure. Besides, to strengthen the local dependencies among neighboring tokens, we propose a novel porous structure during self-attentional computation processing, in which every two non-neighboring tokens are connected through a shared pivot node. Experimental results on four datasets show that our model performs up to 9.47 times faster than state-of-the-art models, while is roughly on a par with its performance. The source code of this paper can be obtained from https://github.com/xxx/xxx.</abstract><doi>10.48550/arxiv.1911.02733</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.1911.02733
ispartof
issn
language eng
recordid cdi_arxiv_primary_1911_02733
source arXiv.org
subjects Computer Science - Computation and Language
Computer Science - Information Retrieval
title Porous Lattice-based Transformer Encoder for Chinese NER
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-15T07%3A03%3A24IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Porous%20Lattice-based%20Transformer%20Encoder%20for%20Chinese%20NER&rft.au=Mengge,%20Xue&rft.date=2019-11-06&rft_id=info:doi/10.48550/arxiv.1911.02733&rft_dat=%3Carxiv_GOX%3E1911_02733%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true