2022
DOI: 10.1007/978-3-031-11644-5_13
|View full text |Cite
|
Sign up to set email alerts
|

Towards Human-Like Educational Question Generation with Large Language Models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
13
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 30 publications
(16 citation statements)
references
References 19 publications
0
13
0
Order By: Relevance
“…Recent work [39,40] has shown that GPT-3 is capable of generating natural questions that require long-form and informative answers, in comparison to existing approaches [41,42] that mainly generate questions designed for short and specific answers. In this work, we prompt GPT-3 to explicitly generate strategic questions about the event.…”
Section: Strategic Questions As Section Headingsmentioning
confidence: 99%
“…Recent work [39,40] has shown that GPT-3 is capable of generating natural questions that require long-form and informative answers, in comparison to existing approaches [41,42] that mainly generate questions designed for short and specific answers. In this work, we prompt GPT-3 to explicitly generate strategic questions about the event.…”
Section: Strategic Questions As Section Headingsmentioning
confidence: 99%
“…For few-shot prompting, we use the text-davinci-003 model (GPT-3.5) from OpenAI 5 with 128 as the maximum token output, 0.7 for temperature and 1.0 for nucleus sampling. Following previous recommendations (Wang et al, 2022;Elkins et al, 2023), we choose the 5-shot setting: beyond the query, 5 examples (each composed of text, question and answer) are incorporated into the prompt. The 5 examples have been randomly extracted from the train set based on the following criterion: the selected examples are consistent with the target attribute (either narrative element or explicitness) one aims to control in the generation process.…”
Section: Implementation Detailsmentioning
confidence: 99%
“…Linguistic Quality: To evaluate the linguistic quality of the generated questions and answers, we report perplexity, grammatical error, and diversity metrics. For perplexity, our motivation is that previous studies (Wang et al, 2022) claim there is a relation between perplexity and coherence, in a way that perplexity is inversely related to the coherence of the generated text: the lower the perplexity score, the higher the coherence. For the sake of computational efficiency, we use GPT-2 (Radford et al, 2019) to compute perplexity.…”
Section: Implementation Detailsmentioning
confidence: 99%
See 2 more Smart Citations