Proceedings of the 10th International Conference on Natural Language Generation 2017
DOI: 10.18653/v1/w17-3518
|View full text |Cite
|
Sign up to set email alerts
|

The WebNLG Challenge: Generating Text from RDF Data

Abstract: The WebNLG challenge consists in mapping sets of RDF triples to text. It provides a common benchmark on which to train, evaluate and compare "microplanners", i.e. generation systems that verbalise a given content by making a range of complex interacting choices including referring expression generation, aggregation, lexicalisation, surface realisation and sentence segmentation. In this paper, we introduce the microplanning task, describe data preparation, introduce our evaluation methodology, analyse participa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
285
0

Year Published

2018
2018
2020
2020

Publication Types

Select...
5
4
1

Relationship

1
9

Authors

Journals

citations
Cited by 259 publications
(327 citation statements)
references
References 15 publications
0
285
0
Order By: Relevance
“…Marcheggiani and Perez Beltrachini (2018) show that explicitly encoding the structure of the graph is beneficial with respect to sequential encoding. They evaluate their model on two tasks, WebNLG (Gardent et al, 2017) and SR11Deep (Belz et al, 2011), but do not apply it to AMR benchmarks. Song et al (2018) and Beck et al (2018) apply recurrent neural networks to directly encode AMR graphs.…”
Section: Related Workmentioning
confidence: 99%
“…Marcheggiani and Perez Beltrachini (2018) show that explicitly encoding the structure of the graph is beneficial with respect to sequential encoding. They evaluate their model on two tasks, WebNLG (Gardent et al, 2017) and SR11Deep (Belz et al, 2011), but do not apply it to AMR benchmarks. Song et al (2018) and Beck et al (2018) apply recurrent neural networks to directly encode AMR graphs.…”
Section: Related Workmentioning
confidence: 99%
“…Our impression is that recurring entity names in the training data coming from multiple reference texts for the same input lead to overfitting on the training vocabulary and to poor generalization to novel inputs. This is also reflected by the rather unsatisfying performance of subword-based approaches in the E2E 1 and WebNLG challenge (ADAPT system (Gardent et al, 2017b)).…”
Section: Input and Output Representationsmentioning
confidence: 99%
“…We evaluate the submitted systems by comparing them to a challenging baseline using automatic evaluation metrics (including novel text-based measures) as well as human evaluation (see Section 7). Note that, while there are other concurrent studies comparing a limited number of end-to-end NLG approaches (Novikova et al, 2017a Gardent et al, 2017a) which emerged during the E2E NLG Challenge, this is the first research to evaluate novel end-to-end generation at scale using human assessment.…”
Section: Introductionmentioning
confidence: 99%