Skip to main content
eScholarship
Open Access Publications from the University of California

Fallible feedback: Judgments of "novice" automated and human writing tutors

Abstract

Automated Writing Evaluation (AWE) tools are steadily gaining prominence in educational settings given the ease and scale in which they can be deployed. Nevertheless, despite producing a similar performance as human raters, the accuracy of automated systems is often met with skepticism. In the current study we explore whether such skepticism extends to writing feedback believed to be generated by human tutors in training or an AI tutor under development. When both sources are fallible, are critical judgments mitigated? Participants (N=477) judged the accuracy of feedback on writing samples given by human or AI tutors where the feedback was normed to be accurate, inaccurate, and ambiguous. Results showed participants were still more likely to deem identical feedback provided by an AI, as compared to human tutors, as less accurate, and they were more confident in these evaluations. It appears that ÒnoviceÓ AWE systems are not completely immune to negative bias.

Main Content
For improved accessibility of PDF content, download the file to your device.
Current View