Evaluating ChatGPT-Generated Linear Algebra Formative Assessments

Author
Keywords
Abstract
This research explored Large Language Models potential uses on formative assessment for mathematical problem-solving process. The study provides a conceptual analysis of feedback and how the use of these models is related in the context of formative assessment for Linear Algebra problems. Particularly, the performance of a popular model known as ChatGPT in mathematical problems fails on reasoning, proofs, model construction, among others. Formative assessment is a process used by teachers and students during instruction that provides feedback to adjust ongoing teaching and learning to improve student’s achievement of intended instructional outcomes. The study analyzed and evaluated feedback provided to engineering students in their solutions, from both, instructors and ChatGPT, against fine-grained criteria of a formative feedback model that includes affective aspects. Considering preliminary outputs, and to improve performance of feedback from both agents’ instructors and ChatGPT, we developed a framework for formative assessment in mathematical problemsolving using a Large Language Model (LLM). We designed a framework to generate prompts, supported by common Linear Algebra mistakes within the context of concept development and problem-solving strategies. In this framework, the instructor acts as an agent to verify tasks in a math problem assigned to students, establishing a virtuous cycle of learning of queries supported by ChatGPT. Results revealed potentialities and challenges on how to improve feedback on graduate-level math problems, by which both educators and students adapt teaching and learning strategies.
Year of Publication
2024
Journal
International Journal of Interactive Multimedia and Artificial Intelligence
Volume
8
Start Page
75
Issue
Special issue on Generative Artificial Intelligence in Education
Number
5
Number of Pages
75-82
Date Published
03/2024
ISSN Number
1989-1660
URL
DOI
Attachment