Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2021
DOI: 10.18653/v1/2021.naacl-main.359
|View full text |Cite
|
Sign up to set email alerts
|

On the Use of Context for Predicting Citation Worthiness of Sentences in Scholarly Articles

Abstract: In this paper, we study the importance of context in predicting the citation worthiness of sentences in scholarly articles. We formulate this problem as a sequence labeling task solved using a hierarchical BiLSTM model. We contribute a new benchmark dataset containing over two million sentences and their corresponding labels. We preserve the sentence order in this dataset and perform document-level train/test splits, which importantly allows incorporating contextual information in the modeling process. We eval… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 20 publications
(11 reference statements)
0
3
0
Order By: Relevance
“…Farber et al [3] and Bonab et al [4] utilized convolutional recurrent neural networks on diverse datasets. Context-aware citation detection was introduced by Gosangi et al [5] with the ACL-cite dataset, integrating BiLSTMs and transformer-based embeddings. Wright et al [6] delved into citation worthiness extensively, incorporating domain adaptation and transfer learning techniques.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Farber et al [3] and Bonab et al [4] utilized convolutional recurrent neural networks on diverse datasets. Context-aware citation detection was introduced by Gosangi et al [5] with the ACL-cite dataset, integrating BiLSTMs and transformer-based embeddings. Wright et al [6] delved into citation worthiness extensively, incorporating domain adaptation and transfer learning techniques.…”
Section: Related Workmentioning
confidence: 99%
“…We experimented with different models trained on our dataset to establish the baselines for the task of citation-worthiness detection (RQ2). For this assessment, we used our subset with 1M entries 5 . The split contains sentences sampled over all jurisdictions.…”
Section: Experimentation and Dicussionmentioning
confidence: 99%
See 1 more Smart Citation