Proceedings of the 2022 ACM Conference on International Computing Education Research - Volume 1 2022
DOI: 10.1145/3501385.3543957
|View full text |Cite
|
Sign up to set email alerts
|

Automatic Generation of Programming Exercises and Code Explanations Using Large Language Models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
27
0
2

Year Published

2022
2022
2023
2023

Publication Types

Select...
3
3
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 180 publications
(57 citation statements)
references
References 62 publications
0
27
0
2
Order By: Relevance
“…Large language models can not only support the assessment of student's solutions but also assist in the automatic generation of exercises. Using few-shot learning, [40] showed that the OpenAI Codex model is able to provide a variety of programming tasks together with the correct solution, automated tests to verify the student's solutions, and additional code explanations. With regard to testing factual knowledge in general, [41] proposed a framework to automatically generate question-answer pairs.…”
Section: Review Of Research Applying Large Languagementioning
confidence: 99%
“…Large language models can not only support the assessment of student's solutions but also assist in the automatic generation of exercises. Using few-shot learning, [40] showed that the OpenAI Codex model is able to provide a variety of programming tasks together with the correct solution, automated tests to verify the student's solutions, and additional code explanations. With regard to testing factual knowledge in general, [41] proposed a framework to automatically generate question-answer pairs.…”
Section: Review Of Research Applying Large Languagementioning
confidence: 99%
“…Figure 3 shows an example where the overall feedback is bad quality and successfully rejected, though parts of the generated explanation are correct; this could potentially be useful for tutors in a human-in-the-loop approach. 4 When comparing PyFiXV P≥70 with any other technique in Figure 6a, the results are significantly different w.r.t. χ 2 tests [41] (p ≤ 0.0001); here, we use contingency tables with two rows (techniques) and four columns (240 data points mapped to four possible precision/coverage outcomes).…”
Section: Resultsmentioning
confidence: 91%
“…students in a large introductory programming course [3]. Subsequently, recent works have shown promising results in using Codex on various programming education scenarios, including generating new programming assignments [4], providing code explanations [5], and enhancing programming-error-messages [6].…”
Section: Introductionmentioning
confidence: 99%
“…Similarly, BERT-generated doctor-patient dialogues were also found to be indistinguishable from actual doctor-patient dialogues, which can be used to create virtual standard patients for medical students' diagnosis practice training [57]. Additionally, for introductory programming courses, the state-ofthe-art LLMs, Codex, could generate sensible and novel exercises for students along with an appropriate sample solution (around three out of four times) and accurate code explanation (67% accuracy) [45].…”
Section: Practical Challenges -Rq2mentioning
confidence: 96%