2022
DOI: 10.1609/aaai.v36i10.21382
|View full text |Cite
|
Sign up to set email alerts
|

Generation-Focused Table-Based Intermediate Pre-training for Free-Form Question Answering

Abstract: Question answering over semi-structured tables has attracted significant attention in the NLP community. However, most of the existing work focus on questions that can be answered with short-form answer, i.e. the answer is often a table cell or aggregation of multiple cells. This can mismatch with the intents of users who want to ask more complex questions that require free-form answers such as explanations. To bridge the gap, most recently, pre-trained sequence-to-sequence language models such as T5 are … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
0
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 37 publications
0
0
0
Order By: Relevance
“…Our pretraining objectives include low-level tasks that are more focused on retrieving the underlying data from the chart images and high-level tasks that align closely with the downstream tasks. (Shi et al, 2022) to generate synthetic open-ended QA pairs. Specifically, a T5 model (Raffel et al, 2020) pretrained on SQuAD (Rajpurkar et al, 2016) is employed to generate an open-ended question for each summary.…”
Section: Pretraining Objectivesmentioning
confidence: 99%
“…Our pretraining objectives include low-level tasks that are more focused on retrieving the underlying data from the chart images and high-level tasks that align closely with the downstream tasks. (Shi et al, 2022) to generate synthetic open-ended QA pairs. Specifically, a T5 model (Raffel et al, 2020) pretrained on SQuAD (Rajpurkar et al, 2016) is employed to generate an open-ended question for each summary.…”
Section: Pretraining Objectivesmentioning
confidence: 99%
“…pretrains the PLMs with structure-augmented objectives (Herzig et al 2020;Deng et al 2021). Specifically, these works usually design unsupervised or weakly-supervised objectives for implicitly modeling the database structures with external or synthetic data corpus (Yu et al 2021b;Shi et al 2022). Although effective, further pretraining a large PLM can incur substantial costs and extra overheads (Yu et al 2021a).…”
Section: Introductionmentioning
confidence: 99%