Neural machine translation for in‐text citation classification

The quality of scientific publications can be measured by quantitative indices such as the h‐index, Source Normalized Impact per Paper, or g‐index. However, these measures lack to explain the function or reasons for citations and the context of citations from citing publication to cited publication....

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Journal of the American Society for Information Science and Technology 2023-10, Vol.74 (10), p.1229-1240
Hauptverfasser: Safder, Iqra, Ali, Momin, Aljohani, Naif Radi, Nawaz, Raheel, Hassan, Saeed‐Ul
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:The quality of scientific publications can be measured by quantitative indices such as the h‐index, Source Normalized Impact per Paper, or g‐index. However, these measures lack to explain the function or reasons for citations and the context of citations from citing publication to cited publication. We argue that citation context may be considered while calculating the impact of research work. However, mining citation context from unstructured full‐text publications is a challenging task. In this paper, we compiled a data set comprising 9,518 citations context. We developed a deep learning‐based architecture for citation context classification. Unlike feature‐based state‐of‐the‐art models, our proposed focal‐loss and class‐weight‐aware BiLSTM model with pretrained GloVe embedding vectors use citation context as input to outperform them in multiclass citation context classification tasks. Our model improves on the baseline state‐of‐the‐art by achieving an F1 score of 0.80 with an accuracy of 0.81 for citation context classification. Moreover, we delve into the effects of using different word embeddings on the performance of the classification model and draw a comparison between fastText, GloVe, and spaCy pretrained word embeddings.
ISSN:2330-1635
2330-1643
DOI:10.1002/asi.24817