A Comparison of Peer- and Tutor-Grading of an Introductory R Coding Assessment

We investigate the level of agreement between tutor and peer grading of an introductory R programming assessment. Comparing peer and tutor grades we find a strong correlation of 0.848, 95%CI = (0.809, 0.880). Using standard multivariate data analysis techniques we find that tutors and peers grade si...

Full description

Saved in:
Bibliographic Details
Main Authors: Charlotte M. Jones-Todd, Amy Renelle
Format: Article
Language:English
Published: Taylor & Francis Group 2025-08-01
Series:Journal of Statistics and Data Science Education
Subjects:
Online Access:https://www.tandfonline.com/doi/10.1080/26939169.2025.2520205
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:We investigate the level of agreement between tutor and peer grading of an introductory R programming assessment. Comparing peer and tutor grades we find a strong correlation of 0.848, 95%CI = (0.809, 0.880). Using standard multivariate data analysis techniques we find that tutors and peers grade similarly given a prescriptive criterion. However, when given a subjective criterion, tutors and peers use different schemas to grade. We find that tutors grade the subjective criterion autonomously from the other rubric criteria, whereas peers grade in line with the prescriptive criteria that evaluates the components and structure of the code. In addition, we estimate between-assessor and between-submission variation using a discrete-Beta mixed model and show that between-submission is greater than grader submission for both peers and tutors. Finally, we advocate for the use of peer assessment as a learning exercise and encourage readers to adapt the activity accordingly.
ISSN:2693-9169