2021
DOI: 10.1007/978-3-030-72113-8_20
|View full text |Cite
|
Sign up to set email alerts
|

Answer Sentence Selection Using Local and Global Context in Transformer Models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
14
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
2

Relationship

4
4

Authors

Journals

citations
Cited by 12 publications
(14 citation statements)
references
References 18 publications
0
14
0
Order By: Relevance
“…For the task of AS2, initial efforts embedded the question and candidates using CNNs (Severyn and Moschitti, 2015), weight aligned networks (Shen et al, 2017;Tran et al, 2018;Tay et al, 2018) and compare-aggregate architectures (Wang and Jiang, 2016;Bian et al, 2017;Yoon et al, 2019). Recent progress has stemmed from the application of transformer models for performing AS2 (Garg et al, 2020;Han et al, 2021;Lauriola and Moschitti, 2021). On the data front, small datasets like TrecQA (Wang et al, 2007) and WikiQA (Yang et al, 2015) have been supplemented with datasets such as ASNQ (Garg et al, 2020) having several million QA pairs.…”
Section: Related Workmentioning
confidence: 99%
“…For the task of AS2, initial efforts embedded the question and candidates using CNNs (Severyn and Moschitti, 2015), weight aligned networks (Shen et al, 2017;Tran et al, 2018;Tay et al, 2018) and compare-aggregate architectures (Wang and Jiang, 2016;Bian et al, 2017;Yoon et al, 2019). Recent progress has stemmed from the application of transformer models for performing AS2 (Garg et al, 2020;Han et al, 2021;Lauriola and Moschitti, 2021). On the data front, small datasets like TrecQA (Wang et al, 2007) and WikiQA (Yang et al, 2015) have been supplemented with datasets such as ASNQ (Garg et al, 2020) having several million QA pairs.…”
Section: Related Workmentioning
confidence: 99%
“…Triplet loss (Hoffer and Ailon, 2015) has been used in few-shot classification methods. Although introduced for images, it has been successfully adapted in natural language processing (Wei et al, 2021;Lauriola and Moschitti, 2021). Triplet loss enables the network to distinguish been positive and negative examples of a class.…”
Section: Triplet Lossmentioning
confidence: 99%
“…Similarly to previous work (Tan et al, 2017;Lauriola and Moschitti, 2021), we define local context Loc k (C i,j ) for candidate C i,j as the sentences immediately preceding and succeeding each answer candidate within a window of 2k + 1 sentences, i.e., Loc k (C i,j ) = C i,j−k , . .…”
Section: Local Contextmentioning
confidence: 99%
“…Their approach, while interesting, is limited to entitiesbased context, and specific to Wikipedia and MR domain. For AS2, Lauriola and Moschitti (2021) proposed a model that uses local context as defined by the preceding and following sentences of the target answer. They also introduced a simple bagof-words representation of documents as global context, which did not show significant improvement over non-contextual AS2 models.…”
Section: Introductionmentioning
confidence: 99%