Investigating the feasibility of automatic assessment of programming tasks
Abstract
Aim/Purpose The aims of this study were to investigate the feasibility of automatic assessment
of programming tasks and to compare manual assessment with automatic assessment in terms of the effect of the different assessment methods on the
marks of the students.
Background Manual assessment of programs written by students can be tedious. The assistance of automatic assessment methods might possibly assist in reducing the
assessment burden, but there may be drawbacks diminishing the benefits of applying automatic assessment. The paper reports on the experience of a lecturer
trying to introduce automated grading. Students’ solutions to a practical Java programming test were assessed both manually and automatically and the lecturer
tied the experience to the unified theory of acceptance and use of technology
(UTAUT).
Methodology The participants were 226 first-year students registered for a Java programming
course. Of the tests the participants submitted, 214 were assessed both manually
and automatically. Various statistical methods were used to compare the manual
assessment of student’s solutions with the automatic assessment of the same
solutions. A detailed investigation of reasons for differences was also carried out.
A further data collection method was the lecturer’s reflection on the feasibility of
automatic assessment of programming tasks based on the UTAUT.
Contribution This study enhances the knowledge regarding benefits and drawbacks of automatic assessment of students’ programming tasks. The research contributes to
the UTAUT by applying it in a context where it has hardly been used. Furthermore, the study is a confirmation of previous work stating that automatic assessment may be less reliable for students with lower marks, but more trustworthy for
the high achieving students
URI
http://hdl.handle.net/10394/31927http://www.jite.org/documents/Vol17/JITEv17IIPp201-223Liebenberg4853.pdf
https://doi.org/10.28945/4150