6285641688335, 628551515511 info@scirepid.com

 
jpk - Jurnal Profesi Keguruan - Vol. 11 Issue. 1 (2025)

Can We Trust AI to Assess Writing? An Analysis of Scoring Reliability and Feedback Consistency

Fitriani Fitriani, Puput Zuli Eko Rini,



Abstract

This study analyzes AI-generated writing assessments' scoring reliability and feedback consistency using ChatGPT. Adopting a mixed-methods approach, 23 student descriptive texts were evaluated across three assessment rounds. Quantitative findings showed high scoring reliability, with an Intraclass Correlation Coefficient (ICC) of 0.93, indicating excellent consistency across repeated evaluations. Qualitative analysis revealed that ChatGPT consistently addressed five core writing criteria—content, organization, vocabulary, language use, and mechanics. However, the feedback varied in focus and detail across rounds, and the absence of reference to prior feedback limited its support for revision as a recursive process. The findings suggest that although ChatGPT demonstrates reliable scoring and generally stable feedback themes, it lacks the continuity to facilitate sustained writing development. To enhance its pedagogical value, AI-based feedback systems should be designed to build upon previous responses, thereby enabling more effective support for students' progressive improvement in writing.







DOI :


Sitasi :

0

PISSN :

2460-4399

EISSN :

2528-7214

Date.Create Crossref:

04-Jul-2025

Date.Issue :

30-Jun-2025

Date.Publish :

30-Jun-2025

Date.PublishOnline :

30-Jun-2025



PDF File :

Resource :

Open

License :