Towards employing native information in citation function classification
Yang Zhang (),
Rongying Zhao (),
Yufei Wang,
Haihua Chen,
Adnan Mahmood,
Munazza Zaib,
Wei Emma Zhang and
Quan Z. Sheng
Additional contact information
Yang Zhang: Wuhan University
Rongying Zhao: Wuhan University
Yufei Wang: Macquarie University
Haihua Chen: University of North Texas
Adnan Mahmood: Macquarie University
Munazza Zaib: Macquarie University
Wei Emma Zhang: The University of Adelaide
Quan Z. Sheng: Macquarie University
Scientometrics, 2022, vol. 127, issue 11, No 26, 6557-6577
Abstract:
Abstract Citations play a fundamental role in supporting authors’ contribution claims throughout a scientific paper. Labelling citation instances with different function labels is indispensable for understanding a scientific text. A single citation is the linkage between two scientific papers in the citation network. These citations encompass rich native information, including context of the citation, citation location, citing and cited paper titles, DOI, and the website’s URL. Nevertheless, previous studies have ignored such rich native information during the process of datasets’ accumulation, thereby resulting in a lack of comprehensive yet significantly valuable features for the citation function classification task. In this paper, we argue that such important information should not be ignored, and accordingly, we extract and integrate all of the native information features into different neural text representation models via trainable embeddings and free text. We first construct a new dataset entitled, NI-Cite, comprising a large number of labelled citations with five key native features (Citation Context, Section Name, Title, DOI, Web URL) against each dataset instance. In addition, we propose to exploit the recently developed text representation models integrated with such information to evaluate the performance of citation function classification task. The experimental results demonstrate that the native information features suggested in this paper enhance the overall classification performance.
Keywords: Citation function classification; Pretrained language model; Natural language processing; Native information (search for similar items in EconPapers)
Date: 2022
References: View references in EconPapers View complete reference list from CitEc
Citations: View citations in EconPapers (5)
Downloads: (external link)
http://link.springer.com/10.1007/s11192-021-04242-0 Abstract (text/html)
Access to the full text of the articles in this series is restricted.
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:spr:scient:v:127:y:2022:i:11:d:10.1007_s11192-021-04242-0
Ordering information: This journal article can be ordered from
http://www.springer.com/economics/journal/11192
DOI: 10.1007/s11192-021-04242-0
Access Statistics for this article
Scientometrics is currently edited by Wolfgang Glänzel
More articles in Scientometrics from Springer, Akadémiai Kiadó
Bibliographic data for series maintained by Sonal Shukla () and Springer Nature Abstracting and Indexing ().