2024
DOI: 10.1088/1361-6404/ad1420
|View full text |Cite
|
Sign up to set email alerts
|

How understanding large language models can inform the use of ChatGPT in physics education

Giulia Polverini,
Bor Gregorcic

Abstract: The paper aims to fulfil three main functions: (1) to serve as an introduction for the physics education community to the functioning of Large Language Models (LLMs), (2) to present a series of illustrative examples demonstrating how prompt-engineering techniques can impact LLMs performance on conceptual physics tasks and (3) to discuss potential implications of the understanding of LLMs and prompt engineering for physics teaching and learning. We first summarise existing research on the performance of a popul… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
6
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 15 publications
(8 citation statements)
references
References 73 publications
2
6
0
Order By: Relevance
“…The first relevant finding of this case study is that we were able to find an introductory conceptual physics question that ChatGPT-4 did not answer correctly. This further supports previous findings, suggesting that its performance on even basic physics tasks significantly limits the usefulness of the chatbot as a physics tutor [15].…”
Section: Chatgptsupporting
confidence: 91%
See 1 more Smart Citation
“…The first relevant finding of this case study is that we were able to find an introductory conceptual physics question that ChatGPT-4 did not answer correctly. This further supports previous findings, suggesting that its performance on even basic physics tasks significantly limits the usefulness of the chatbot as a physics tutor [15].…”
Section: Chatgptsupporting
confidence: 91%
“…This does not seem to be the case in physics, where even the state-of-the-art chatbot ChatGPT-4 3 [14] still has some way to go before its performance can be considered expert-like. Surprisingly, this is true not only for advanced physics topics but even for introductory conceptual physics, where it can still clumsily fail [15]. Therefore, the effectiveness of the technology is contingent upon the users' expertise in the subject matter, their ability to craft useful prompts, and critically evaluate the outputs.…”
Section: Introductionmentioning
confidence: 99%
“…"Set the Scene" improves the answer due to the context sensitivity of large language models. 44 "Be specific" is similar to what Zheng et al propose as "Detailed Instruction". 46 "Simplify your language" is sometimes contradicting specificity, but a consequence of how large language models work with natural language.…”
Section: The Five S Prompting Frameworksupporting
confidence: 53%
“…46 "Simplify your language" is sometimes contradicting specificity, but a consequence of how large language models work with natural language. 44 "Structure the output" again is a principle found with Zheng et al 46 "Share Feedback" is a further strategy developed due to the high dependency of large language models on the data provided to it. While this prompting framework is sometimes contradictory and, in some ways, fuzzy, it might serve as a facilitator for reflecting on students' own prompts.…”
Section: The Five S Prompting Frameworkmentioning
confidence: 99%
See 1 more Smart Citation