2021
DOI: 10.48550/arxiv.2107.07261
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Turning Tables: Generating Examples from Semi-structured Tables for Endowing Language Models with Reasoning Skills

Abstract: Models pre-trained with a language modeling objective possess ample world knowledge and language skills, but are known to struggle in tasks that require reasoning. In this work, we propose to leverage semi-structured tables, and automatically generate at scale questionparagraph pairs, where answering the question requires reasoning over multiple facts in the paragraph. We add a pre-training step over this synthetic data, which includes examples that require 16 different reasoning skills such as number comparis… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
8
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(8 citation statements)
references
References 24 publications
(43 reference statements)
0
8
0
Order By: Relevance
“…Thus, more effort needs to be put in using semi-structured data for model evaluation. Such approaches can be applied on other datasets such as WikiTableQA (Pasupat and Liang, 2015), TabFact (Chen et al, 2019), Hy-bridQA (Chen et al, 2020b;Zayats et al, 2021;Oguz et al, 2020), OpenTableQA (Chen et al, 2021), ToTTo (Parikh et al, 2020, Turing Tables (Yoran et al, 2021) i.e. table to text generation tasks, LogicTable and (Chen et al, 2020a) and recently proposed tabular reasoning models proposed in TAPAS (Müller et al, 2021;Herzig et al, 2020), TaBERT (Yin et al, 2020), TABBIE (Iida et al, 2021), TabGCN (Pramanick and Bhattacharya, 2021) and RCI (Glass et al, 2021).…”
Section: Discussion and Related Workmentioning
confidence: 99%
“…Thus, more effort needs to be put in using semi-structured data for model evaluation. Such approaches can be applied on other datasets such as WikiTableQA (Pasupat and Liang, 2015), TabFact (Chen et al, 2019), Hy-bridQA (Chen et al, 2020b;Zayats et al, 2021;Oguz et al, 2020), OpenTableQA (Chen et al, 2021), ToTTo (Parikh et al, 2020, Turing Tables (Yoran et al, 2021) i.e. table to text generation tasks, LogicTable and (Chen et al, 2020a) and recently proposed tabular reasoning models proposed in TAPAS (Müller et al, 2021;Herzig et al, 2020), TaBERT (Yin et al, 2020), TABBIE (Iida et al, 2021), TabGCN (Pramanick and Bhattacharya, 2021) and RCI (Glass et al, 2021).…”
Section: Discussion and Related Workmentioning
confidence: 99%
“…Given a table and an executable SQL query, TaPEx uses the query's execution result (obtained through an off-the-shelf SQL executor, e.g., MySQL) to supervise the TaLM as a neural executor. Yoran et al [103] generate at scale question-paragraph pairs that require different reasoning skills to enhance the numerical reasoning abilities in table QA. [69], described above, also has benefits for this task.…”
Section: Objectives By Downstream Tasksmentioning
confidence: 99%
“…Tabular Reasoning. Recent studies investigate various NLP tasks on semi-structured tabular data, including tabular NLI and fact verification Gupta et al, 2020;Zhang and Balog, 2019), tabular probing , various question answering and semantic parsing tasks (Pasupat and Liang, 2015;Krishnamurthy et al, 2017;Abbas et al, 2016;Sun et al, 2016;Chen et al, 2020b;Lin et al, 2020;Zayats et al, 2021;Oguz et al, 2020;Chen et al, 2021, inter alia), and table-to-text generation (e.g., Nan et al, 2021;Yoran et al, 2021;Chen et al, 2020a). Several strategies for representing Wikipedia relational tables were recently proposed, such as TAPAS (Herzig et al, 2020), TaBERT (Yin et al, 2020), TabStruc , TABBIE (Iida et al, 2021), TabGCN (Pramanick andBhattacharya, 2021) and RCI (Glass et al, 2021).…”
Section: Related Workmentioning
confidence: 99%