Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1278
|View full text |Cite
|
Sign up to set email alerts
|

Semantic graph parsing with recurrent neural network DAG grammars

Abstract: Semantic parses are directed acyclic graphs (DAGs), so semantic parsing should be modeled as graph prediction. But predicting graphs presents difficult technical challenges, so it is simpler and more common to predict the linearized graphs found in semantic parsing datasets using well-understood sequence models. The cost of this simplicity is that the predicted strings may not be wellformed graphs. We present recurrent neural network DAG grammars, a graph-aware sequence model that ensures only well-formed grap… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
26
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
2
1

Relationship

3
5

Authors

Journals

citations
Cited by 19 publications
(28 citation statements)
references
References 26 publications
0
26
0
Order By: Relevance
“…Interestingly, the two-encoder setup seems to be preferable for these smaller, non-English data sets. For 2.2.0, we outperform the system of Fancellu et al (2019) for German and Italian and obtain competitive scores for Dutch.…”
Section: Robustness Across Languagesmentioning
confidence: 90%
“…Interestingly, the two-encoder setup seems to be preferable for these smaller, non-English data sets. For 2.2.0, we outperform the system of Fancellu et al (2019) for German and Italian and obtain competitive scores for Dutch.…”
Section: Robustness Across Languagesmentioning
confidence: 90%
“…Most approaches that produce semantic graphs (see for an overview) model distributions over graphs directly (Dozat and Manning, 2018;Zhang et al, 2019;He and Choi, 2020;Cai and Lam, 2020), while others make use of derivation trees that compositionally evaluate to graphs (Groschwitz et al, 2018;Chen et al, 2018;Fancellu et al, 2019;Lindemann et al, 2019). AM dependency parsing belongs to the latter category.…”
Section: Related Workmentioning
confidence: 99%
“…Because AMR are naturally structured objects (e.g. tree structures), semantic AMR parsing methods based on deep graph generative models are deemed as promising [19,35,88,130,145]. These methods represent the semantics of a sentence as a semantic graph (i.e., a sub-graph of a knowledge base) and treat semantic parsing as a semantic graph matching/generation process.…”
Section: Semantic Parsingmentioning
confidence: 99%