2022
DOI: 10.48550/arxiv.2212.08037
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Attributed Question Answering: Evaluation and Modeling for Attributed Large Language Models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 10 publications
(19 citation statements)
references
References 0 publications
0
3
0
Order By: Relevance
“…Madaan et al present an iterative self-refinement algorithm that alternates between feedback and refinement. Additionally, LLMs are also used to evaluate attribution between generated answers and references [2,51].…”
Section: Related Work 21 Open-domain Question Answeringmentioning
confidence: 99%
See 2 more Smart Citations
“…Madaan et al present an iterative self-refinement algorithm that alternates between feedback and refinement. Additionally, LLMs are also used to evaluate attribution between generated answers and references [2,51].…”
Section: Related Work 21 Open-domain Question Answeringmentioning
confidence: 99%
“…Current ODQA methods follow two main paradigms in preparation for answering questions: (1) The retrieve-then-read paradigm retrieves pertinent evidence documents from an external corpus and generates an answer based on them [16,18]. Since retrieval models often rely on well-curated corpora like Wikipedia, they can provide highly factual and accurate information about the question; (2) The generate-then-read paradigm directly employs language models to generate virtual documents [49], diversifying the evidence sources and enhancing answer coverage for the question.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Attribution and Fact Checking Our goal is closely related to works that check if LM-generated texts are faithful to a given source text (Bohnet et al, 2022;Honovich et al, 2022). This problem has been addressed via several approaches,…”
Section: Related Workmentioning
confidence: 99%
“…Modern language models (LMs) often generate inconsistent (Elazar et al, 2021), non-attributable (Rashkin et al, 2021;Bohnet et al, 2022;Liu et al, 2023a), or factually incorrect text (Tam et al, 2022;Devaraj et al, 2022;Maynez et al, 2020), thus negatively impacting the reliability of these models (Amodei et al, 2016;Hendrycks et al, 2021). This has prompted the community to develop methods that calibrate the confidence of model predictions to better align with their quality (Brundage et al, 2020).…”
Section: Introductionmentioning
confidence: 99%