Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2022
DOI: 10.18653/v1/2022.acl-long.76
|View full text |Cite
|
Sign up to set email alerts
|

Sentence-aware Contrastive Learning for Open-Domain Passage Retrieval

Abstract: Training dense passage representations via contrastive learning has been shown effective for Open-Domain Passage Retrieval (ODPR). Existing studies focus on further optimizing by improving negative sampling strategy or extra pretraining. However, these studies keep unknown in capturing passage with internal representation conflicts from improper modeling granularity. Specifically, under our observation that a passage can be organized by multiple semantically different sentences, modeling such a passage as a un… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
0
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(3 citation statements)
references
References 15 publications
0
0
0
Order By: Relevance
“…Distant supervision has been used to train MRC models in low-resource settings, and two main kinds of approaches have been proposed to address the mislabeling problem: (1) filtering noisy labels, and (2) modeling answer spans as latent variables. The noise filtering approaches learn to score and rank DS instances based on answer span positions Tay et al, 2018;Swayamdipta et al, 2018;Clark and Gardner, 2018;Lin et al, 2018;Joshi et al, 2017;Chen et al, 2017), question-passage similarities (Hong et al, 2022;Qin et al, 2021;Shao et al, 2021;Deng et al, 2021) and model confidences Zhu et al, 2022). The latent variable-based approaches jointly train MRC models and identify correct answer spans using hard-EM algorithms (Zhao et al, 2021;Min et al, 2019;Cheng et al, 2020).…”
Section: Relate Workmentioning
confidence: 99%
“…Distant supervision has been used to train MRC models in low-resource settings, and two main kinds of approaches have been proposed to address the mislabeling problem: (1) filtering noisy labels, and (2) modeling answer spans as latent variables. The noise filtering approaches learn to score and rank DS instances based on answer span positions Tay et al, 2018;Swayamdipta et al, 2018;Clark and Gardner, 2018;Lin et al, 2018;Joshi et al, 2017;Chen et al, 2017), question-passage similarities (Hong et al, 2022;Qin et al, 2021;Shao et al, 2021;Deng et al, 2021) and model confidences Zhu et al, 2022). The latent variable-based approaches jointly train MRC models and identify correct answer spans using hard-EM algorithms (Zhao et al, 2021;Min et al, 2019;Cheng et al, 2020).…”
Section: Relate Workmentioning
confidence: 99%
“…Nous verrons à la section suivante que les tâches de pré-entraînement proposées par et Ram et al (2022) génèrent plusieurs pseudo-questions à partir du même passage, en accord avec ces résultats. Ces travaux sont également liés à ceux de Zhang et al (2022) et Hong et al (2022, qui produisent plusieurs représentations par passage. Hong et al (2022), en particulier, démontrent une meilleure robustesse de leur modèle au changement de domaine.…”
Section: Recherche D'information 421 Recherche Neuronale : Dense Ou P...unclassified
“…Ces travaux sont également liés à ceux de Zhang et al (2022) et Hong et al (2022, qui produisent plusieurs représentations par passage. Hong et al (2022), en particulier, démontrent une meilleure robustesse de leur modèle au changement de domaine.…”
Section: Recherche D'information 421 Recherche Neuronale : Dense Ou P...unclassified