Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/107742
PIRA download icon_1.1View/Download Full Text
Title: Evaluating the effectiveness of AI-based essay grading tools in the summative assessment of higher education
Authors: Yeung, WE 
Qi, C 
Xiao, JL 
Wong, FR 
Issue Date: 2023
Source: ICERI2023 Proceedings : 16th International Conference of Education, Research and Innovation, November 13th–15th, 2023, Seville, Spain, p. 8069-8073
Abstract: Recently, new developments in Artificial Intelligence (AI)-related educational assessment are attracting increasing interest from educators. AI-based marking or grading tools are able to greatly ease teachers’ workload, especially when marking individual essays that involve thousands of university students. Compared with other assessment methods, essays are notably time consuming to be marked manually. Implementing AI algorithms for automated grading of essays not only reduces the time for assessment, but also provides a mechanism to test the robustness of human grading itself (Ramalingam, et al., 2018). There are three categories of assessment in higher education: diagnostic assessment, formative assessment, and summative assessment. AI tools for essay grading are more used in formative and summative assessments than for diagnostic assessment. Compared with those of formative assessments, essays in summative assessments are usually longer and more challenging to mark. It is therefore imperative to understand how to use AI-based automated essay scoring (AES) tools to complete the marking tasks effectively. Technology-enhanced assessment have been widely discussed in pedagogy literature, however, the study on AI assessment, especially AI assessment for summative essay grading in higher education is rare. This research-in-progress paper provides solid theoretical foundation and study plan to assess and compare the effectiveness of various existing AI-based essay grading tools in higher education. The AI-based essay grading tools include Intelligent Essay Assessor, Intellimetric, e-Rater, Copyleaks, progressay, and ASC. Specifically, in our research design, reflective essay assignments from a freshmen course that involve thousands of students in university will be used as the summative assessments. 800 student assignments with human-generated grades will be used as training data for the selected AI grading platforms. Based on platform requirements/functions, the assessment rubrics and marking rules will be entered into each AI grader platforms. Grades of another 200 student assignments generated by the selected AI graders will be used as testing data to examine the predictive accuracy and effectiveness of these AI algorithms. A high level of accuracy and effectiveness means closer scores to those of human marked essays. Finally, we are going to compare the accuracy and reliability of the grades generated by different platforms and select the one with highest prediction rate for future practice or further research. Statistics such as correlation and T-test will be used to test the differences between platforms. For future research plan, upon selecting the best assessment platform, we intend to investigate on how to generate tailored feedback to each individual student in other formative assessments. In the last stage, after gaining experiences with the commercial platforms, we will further develop our own advanced AI programs to enhance the overall effectiveness of the AI-based assessment tool in both summative and formative assessments.
Keywords: Artificial Intelligence
Essay grading
Higher education
Summative assessment
Publisher: International Academy of Technology, Education and Development (IATED)
DOI: 10.21125/iceri.2023.2063
Rights: © Copyright 2023, IATED Academy. This work is subject to copyright. All rights reserved.
Posted with permission of the publisher.
The following paper W.E. Yeung, C. Qi, J.L. Xiao, F.R. Wong (2023) Evaluating the effectiveness of AI-based essay grading tools in the summative assessment of higher education, ICERI2023 Proceedings, pp. 8069-8073 is available at https://library.iated.org/view/YEUNG2023EVA.
Appears in Collections:Conference Paper

Files in This Item:
File Description SizeFormat 
Yeung_Evaluating_Effectiveness_AI-based.pdfPre-Published version766.5 kBAdobe PDFView/Open
Open Access Information
Status open access
File Version Final Accepted Manuscript
Access
View full-text via PolyU eLinks SFX Query
Show full item record

Page views

190
Citations as of Apr 13, 2025

Downloads

329
Citations as of Apr 13, 2025

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.