How do physics students evaluate artificial intelligence responses on comprehension questions? A study on the perceived scientific accuracy and linguistic quality of ChatGPT

This study aimed at evaluating how students perceive the linguistic quality and scientific accuracy of ChatGPT responses to physics comprehension questions. A total of 102 first- and second-year physics students were confronted with three questions of progressing difficulty from introductory mechanics (rolling motion, waves, and fluid dynamics). Each question was presented with four different responses. All responses were attributed to ChatGPT, but in reality one sample solution was created by the researchers. All ChatGPT responses obtained in this study were wrong, imprecise, incomplete, or misleading. We found little differences in the perceived linguistic quality between ChatGPT responses and the sample solution. However, the students rated the overall scientific accuracy of the responses significantly differently, with the sample solution being rated best for the questions of low and medium difficulty. The discrepancy between the sample solution and the ChatGPT responses increased with the level of self-assessed knowledge of the question content. For the question of highest difficulty (fluid dynamics) that was unknown to most students, a ChatGPT response was rated just as good as the sample solution. Thus, this study provides data on the students' perception of ChatGPT responses and the factors influencing their perception. The results highlight the need for careful evaluation of ChatGPT responses both by instructors and students, particularly regarding scientific accuracy. Therefore, future research could explore the potential of similar"spot the bot"-activities in physics education to foster students' critical thinking skills.

[1]  Philipp Bitzenbauer ChatGPT in physics education: A pilot study on easy-to-implement activities , 2023, Contemporary Educational Technology.

[2]  S. Fergus,et al.  Evaluating Academic Answers Generated Using ChatGPT , 2023, Journal of Chemical Education.

[3]  Colin G. West Advances in apparent conceptual physics reasoning in GPT-4 , 2023, ArXiv.

[4]  B. Gregorcic,et al.  ChatGPT and the frustrated Socrates , 2023, Physics Education.

[5]  Colin G. West AI and the FCI: Can ChatGPT Project an Understanding of Introductory Physics? , 2023, ArXiv.

[6]  G. Kortemeyer Could an artificial-intelligence agent pass an introductory physics course? , 2023, Physical Review Physics Education Research.

[7]  Jörg Wittwer,et al.  Misconceptions in Physics Explainer Videos and the Illusion of Understanding: an Experimental Study , 2022, International Journal of Science and Mathematics Education.

[8]  Mónica Cortiñas,et al.  Teachers’ attitudes towards chatbots in education: a technology acceptance model approach considering the effect of social language, bot proactiveness, and users’ characteristics , 2021, Educational Studies.

[9]  Abejide Ade-Ibijola,et al.  Chatbots applications in education: A systematic review , 2021, Comput. Educ. Artif. Intell..

[10]  K. Bhattacharjee,et al.  Adoption of artificial intelligence in higher education: a quantitative analysis using structural equation modelling , 2020, Education and Information Technologies.

[11]  Jörg Wittwer,et al.  The double curse of misconceptions: misconceptions impair not only text comprehension but also metacomprehension in the domain of statistics , 2018 .

[12]  Chandralekha Singh,et al.  Do advanced physics students learn from their mistakes without explicit intervention , 2010 .

[13]  Alexander Renkl,et al.  Finding and fixing errors in worked examples: Can this foster learning outcomes? ☆ , 2007 .

[14]  Joseph Weizenbaum,et al.  ELIZA—a computer program for the study of natural language communication between man and machine , 1966, CACM.