An automated assignment grading system that leverages LLMs and AI to enhance grading efficiency and reliability. It includes modules for data input, criteria definition, AI integration, consistency checks, and comprehensive reporting, aimed at improving educational outcomes.
To experimentally determine the effectiveness of Large Language Models (LLMs) in grading various types of assignment submissions and to assess their performance relative to human graders.
Conduct tests using both the whole assignment and segmented parts.
Collect grading outcomes from LLMs and compare these with benchmarks set by human graders.
Analysis:
Use statistical methods to analyze the data collected and validate the hypotheses.
Focus on evaluating how closely LLM grading aligns with human grading standards.
Goals
Primary Goal: Determine the feasibility of replacing or augmenting human grading with LLMs.
Secondary Goal: Explore the potential for customizing the grading style of LLMs to match different grading preferences, emulating the style of various professors.
Future Work
In a subsequent phase, investigate methods to adapt LLM grading styles based on specific user preferences and grading styles of different educators.
Expected Outcomes
Comprehensive analysis report detailing the performance of LLMs in grading assignments relative to human graders.
Insights into the adaptability of LLMs for personalized grading styles.
This issue aims to methodically assess the capabilities of LLMs in an educational setting, focusing on their potential to enhance or replace traditional grading methods while maintaining or improving grading accuracy and personalization.
Objective
To experimentally determine the effectiveness of Large Language Models (LLMs) in grading various types of assignment submissions and to assess their performance relative to human graders.
Background Resources
Experimental Design
Hypothesis Formation: Develop clear hypotheses based on the potential outcomes of LLMs in assignment grading. Here is a Wiki Document for starters Hypotheses for Testing Automated Assignment Grading Software
Data Collection:
Analysis:
Goals
Future Work
Expected Outcomes
This issue aims to methodically assess the capabilities of LLMs in an educational setting, focusing on their potential to enhance or replace traditional grading methods while maintaining or improving grading accuracy and personalization.