Proceedings of the 11th International Conference on Natural Language Generation 2018
DOI: 10.18653/v1/w18-6535
|View full text |Cite
|
Sign up to set email alerts
|

Can Neural Generators for Dialogue Learn Sentence Planning and Discourse Structuring?

Abstract: Responses in task-oriented dialogue systems often realize multiple propositions whose ultimate form depends on the use of sentence planning and discourse structuring operations. For example a recommendation may consist of an explicitly evaluative utterance e.g. Chanpen Thai is the best option, along with content related by the justification discourse relation, e.g. It has great food and service, that combines multiple propositions into a single phrase. While neural generation methods integrate sentence plannin… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
34
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 30 publications
(34 citation statements)
references
References 50 publications
0
34
0
Order By: Relevance
“…Second, semantic errors were computed following Reed et al (2018), where we implemented a script to estimate the coverage automatically based on regular expression matching. 28 This allowed us to produce an independent estimate of the proportion of outputs with missing or added information (see Table 12).…”
Section: Error Analysis: Input Mr Coveragementioning
confidence: 99%
See 2 more Smart Citations
“…Second, semantic errors were computed following Reed et al (2018), where we implemented a script to estimate the coverage automatically based on regular expression matching. 28 This allowed us to produce an independent estimate of the proportion of outputs with missing or added information (see Table 12).…”
Section: Error Analysis: Input Mr Coveragementioning
confidence: 99%
“…28 This allowed us to produce an independent estimate of the proportion of outputs with missing or added information (see Table 12). Following Reed et al (2018), we also computed the slot error rate (SER) using this pattern-matching approach and the following formula: 29 SER = # missed + # added + # value errors + # repetitions # slots (5) Here, missed stands for slot values missing from the realisations, added denotes additional information not present in the MR (hallucinations), value errors denote correctly realised slots with incorrect values (e.g., specifying low price range instead of high),…”
Section: Error Analysis: Input Mr Coveragementioning
confidence: 99%
See 1 more Smart Citation
“…This poses a dual challenge: First, since the MR does not specify these discourse relations, crowdworkers creating the dataset in turn have no instructions on when to use them, and must thus use their own judgment in creating a natural-sounding response. While the E2E organizers tout the resulting response variations as a plus, Reed et al (2018) find that current neural systems are unable to learn to express discourse relations effectively with this dataset, and explore ways of enriching input MRs to do so. Indeed, now that the E2E system outputs have been released, a search through outputs from all participating systems reveals only 43 outputs (0.4% out of 10080) containing contrastive tokens, on a test set containing about 300 contrastive samples.…”
Section: Limitations Of Flat Mrsmentioning
confidence: 99%
“…To produce a cleaned version of the E2E data, we used the original human textual references, but paired them with correctly matching MRs. 4 To this end, we reimplemented the slot matching script of Reed et al (2018), which tags MR slots and values using regular expressions. We tuned our expressions based on the first 500 instances from the E2E development set and ran the script on the full dataset, producing corrected MRs for all human references (see Figure 1).…”
Section: Cleaning the Meaning Representationsmentioning
confidence: 99%