Opportunities and challenges of large language models in physics education
ORAL · Invited
Abstract
Specifically, we report on a comparison of pre-service teachers creating physics tasks using either ChatGPT or a textbook. The correctness of tasks was similar across both groups and both groups adjusted the difficulty level well for their target audience. Moreover, both groups struggled with task specificity and often omitted key solution information. However, those using textbooks created clearer tasks and embedded them more effectively in contexts. While ChatGPT users praised its user-friendliness, they faced issues with output quality.
Additionally, we show how LLMs are sensitive to students’ misconceptions and can aid in developing and validating Concept Inventories. Concept inventories are crucial for checking learners' understanding of physics concepts, but the creation and validation of such inventories are often resource intensive. In this context, we show how ChatGPT enables the creation synthetically generated empirical data, significantly accelerating the creation and validation of concept tests.
Overall, LLMs offer novel perspectives on the intersection of AI and physics education. However, their integration into teaching and learning should be carefully managed, considering their limitations and ensuring they complement traditional teaching methods.
–
Publication: Kasneci, E., Seßler, K., Küchemann, S., Bannert, M., Dementieva, D., Fischer, F., ... & Kasneci, G. (2023). ChatGPT for good? On opportunities and challenges of large language models for education. Learning and individual differences, 103, 102274.<br><br>Kieser, F., Wulff, P., Kuhn, J., & Küchemann, S. (2023). Educational data augmentation in physics education research using ChatGPT. Physical Review Physics Education Research, 19(2), 020150.<br><br>Krupp, L., Steinert, S., Kiefer-Emmanouilidis, M., Avila, K. E., Lukowicz, P., Kuhn, J., Küchemann, S., & Karolus, J. (2023). Challenges and Opportunities of Moderating Usage of Large Language Models in Education. arXiv preprint arXiv:2312.14969.<br><br>Krupp, L., Steinert, S., Kiefer-Emmanouilidis, M., Avila, K. E., Lukowicz, P., Kuhn, J., Küchemann, S., & Karolus, J. (2023). Unreflected Acceptance--Investigating the Negative Consequences of ChatGPT-Assisted Problem Solving in Physics Education. arXiv preprint arXiv:2309.03087.<br><br>Küchemann, S., Steinert, S., Revenga, N., Schweinberger, M., Dinc, Y., Avila, K. E., & Kuhn, J. (2023). Can ChatGPT support prospective teachers in physics task development?. Physical Review Physics Education Research, 19(2), 020128.<br><br>Küchemann, S., Avila, K. E., Dinc, Y., Hortmann, C., Revenga, N., Ruf, V., … Kuhn, J. (2024). Are Large Multimodal Foundation Models all we need? On Opportunities and Challenges of these Models in Education. Retrieved from osf.io/preprints/edarxiv/n7dvf <br><br>Steinert, S., Avila, K. E., Ruzika, S., Kuhn, J., & Küchemann, S. (2023). Harnessing Large Language Models to Enhance Self-Regulated Learning via Formative Feedback. arXiv preprint arXiv:2311.13984.
Presenters
-
Stefan Küchemann
Ludwig-Maximilians-Universität München
Authors
-
Stefan Küchemann
Ludwig-Maximilians-Universität München