Enhancing Structure Preservation in Coreference Resolution by Constrained Graph Encoding
Coreference resolution is a challenging yet practical problem. Most previous methods are designed to better utilize sequential features of language but can hardly capture the structural associations between mentions. In addition, it is often observed that during long-term training, the embeddings pr...
Gespeichert in:
Veröffentlicht in: | IEEE/ACM transactions on audio, speech, and language processing speech, and language processing, 2022, Vol.30, p.2557-2567 |
---|---|
Hauptverfasser: | , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Coreference resolution is a challenging yet practical problem. Most previous methods are designed to better utilize sequential features of language but can hardly capture the structural associations between mentions. In addition, it is often observed that during long-term training, the embeddings projected from unrelated mentions tend to move closer or even mix together, which increases the difficulty of learning decision boundaries. To tackle these issues: i) We propose a general graph schema derived from diverse knowledge sources (e.g., lemma, type, and semantic roles) to directly link mentions, so that rich information can be exchanged via the relevant connections; ii) We impose two adaptive constraints during graph encoding to regularize the embedding space. One is used to force different sub-modules to generate consistent predictions for the same mention pairs, and the other aims to make the learned embeddings corresponding to unrelated mentions more distinguishable while those of coreferential mentions more similar. Results on two public datasets (ECB+ and ACE05) show that our model consistently outperforms state-of-the-art baselines under different settings with p-value less than 0.01 in t-test, especially learning effectively from the limited labeled data. |
---|---|
ISSN: | 2329-9290 2329-9304 |
DOI: | 10.1109/TASLP.2022.3193222 |