Proceedings of the Third Workshop on Narrative Understanding 2021
DOI: 10.18653/v1/2021.nuse-1.9
|View full text |Cite
|
Sign up to set email alerts
|

Fabula Entropy Indexing: Objective Measures of Story Coherence

Abstract: Automated story generation remains a difficult area of research because it lacks strong objective measures. Generated stories may be linguistically sound, but in many cases suffer poor narrative coherence required for a compelling, logically-sound story. To address this, we present Fabula Entropy Indexing (FEI), an evaluation method to assess story coherence by measuring the degree to which human participants agree with each other when answering true/false questions about stories. We devise two theoretically g… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(8 citation statements)
references
References 26 publications
0
8
0
Order By: Relevance
“…The task of generating compelling story ideas is difficult to formulate and evaluate because it is inherently subjective. Our work focuses on ensuring coherence, a key characteristic that has been identified and extensively used in previous literature [6,7,19,67]. The suggested story ideas should fit seamlessly within the user's narrative, i.e.…”
Section: Trope Suggestionmentioning
confidence: 99%
“…The task of generating compelling story ideas is difficult to formulate and evaluate because it is inherently subjective. Our work focuses on ensuring coherence, a key characteristic that has been identified and extensively used in previous literature [6,7,19,67]. The suggested story ideas should fit seamlessly within the user's narrative, i.e.…”
Section: Trope Suggestionmentioning
confidence: 99%
“…In Bai et al (2022b) human input defines the constitution but AI feedback is used to implement it during training. A seed of human generated examples guiding synthetic data generation also applies elsewhere (Bang et al, 2022;Castricato et al, 2022;Honovich et al, 2022;. Other articles adopt human labels on pre-existing datasets (Böhm et al, 2019;Liu et al, 2021a;Arora et al, 2022;, or leverage implicit feedback data from stories (Nahian et al, 2020) and social media such as Reddit or StackOverflow (Gao et al, 2020;Askell et al, 2021;Bai et al, 2022a).…”
Section: Collecting Feedbackmentioning
confidence: 99%
“…It varies how far removed the human input is, for example in designing the constitution (Bai et al, 2022a), in determining the automated metric (Nguyen et al, 2022;Korbak et al, 2023) or in compiling the word lists to measure political bias (Liu et al, 2021b). Often another model is treated as the 'oracle' to simulate human rewards- Gao et al (2018), for example, simulate preferences on two summaries with perfect, noisy and logistic noisy "oracles" based on ROGUE scores; Wang et al (2021) take the reward as human revisions from parallel machine translation corpora; while others deploy the rewards from a value, moral or toxicity classifier trained on crowdworker labels to reinforce a generator (Qiu et al, 2021;Castricato et al, 2022;Pyatkin et al, 2022).…”
Section: Integrating Feedbackmentioning
confidence: 99%
“…Assessing narratives is a complex and non-trivial task. The goal is to create a narrative that is both syntactically correct (e.g., coherent and consistent) and semantically rich (e.g., novel and interesting) [43]- [45]. Perez y Perez and Ortiz [46] proposed a model to evaluate interestingness based on novelty and correct story recount, with emphasis on the story's opening, closure, and dramatic tensions.…”
Section: Related Workmentioning
confidence: 99%