2021
DOI: 10.1609/aaai.v35i15.17627
|View full text |Cite
|
Sign up to set email alerts
|

Learning Contextual Representations for Semantic Parsing with Generation-Augmented Pre-Training

Abstract: Most recently, there has been significant interest in learning contextual representations for various NLP tasks, by leveraging large scale text corpora to train powerful language models with self-supervised learning objectives, such as Masked Language Model (MLM). Based on a pilot study, we observe three issues of existing general-purpose language models when they are applied in the text-to-SQL semantic parsers: fail to detect the column mentions in the utterances, to infer the column mentions from the cell va… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
18
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 47 publications
(18 citation statements)
references
References 50 publications
0
18
0
Order By: Relevance
“…( 9) Global-GNN [3] proposes a semantic parser that globally reasons about the structure of the output query to make a more contextually informed selection of database constants. Many previous works design adaptive PLM models for specific Text-to-SQL models to achieve better result, such as GAP [35], GRAPPA [45], STRUG [9]. For fair comparison, except for comparing the result on a unified pre-training model BERT-large, we also report the result with model adaptive PLM.…”
Section: Methodsmentioning
confidence: 99%
“…( 9) Global-GNN [3] proposes a semantic parser that globally reasons about the structure of the output query to make a more contextually informed selection of database constants. Many previous works design adaptive PLM models for specific Text-to-SQL models to achieve better result, such as GAP [35], GRAPPA [45], STRUG [9]. For fair comparison, except for comparing the result on a unified pre-training model BERT-large, we also report the result with model adaptive PLM.…”
Section: Methodsmentioning
confidence: 99%
“…To make the task feasible, we keep the condition values (e.g., prescription name) the same in this work, except for genders and vital signs, as this is another major challenge in semantic parsing [30,29]. When splitting the dataset into train, validation, and test sets, we ensure that all the question templates are present in each split.…”
Section: Taskmentioning
confidence: 99%
“…We choose MIMICSQL, which contains simple SQL queries that can all be parsed with the Spider grammar, and EHRSQL for the healthcare datasets. Among many SOTA models in the Spider leaderboard, we use Generation-Augmented Pre-training (GAP) [29] to test its zero-shot domain transfer performance.…”
Section: Model Developmentmentioning
confidence: 99%
“…Among them, TaBERT (Yin et al 2020) and TAPAS (Herzig et al 2020) design structure-related unsupervised objectives for further pretraining the BERT model over millions of web tables and the surrounding text. GRAPPA (Yu et al 2021) and GAP (Shi et al 2021) utilize data augmentation techniques to synthesize high-quality pretraining corpus and respectively pretrain a RoBERTa and a BART model.…”
Section: Semantic Parsingmentioning
confidence: 99%
“…Despite the promising performance, current PLM-based approaches most regard both input and output as plain text sequences and neglect the structural information contained in the sentences (Yin et al 2020;Shi et al 2021), such as the database (DB) or knowledge base (KB) schema that essentially constitutes the key semantics of the target SQL or SPARQL logical forms. As a result, these PLM-based models often suffer from the hallucination issue (Ji et al 2022) and may generate incorrect logical form structures that are unfaithful to the input utterance (Nicosia, Qu, and Altun 2021;Gupta et al 2022).…”
Section: Introductionmentioning
confidence: 99%