Back to Search Start Over

Utilizing Generalizability Theory to Investigate the Reliability of Grades Assigned to Undergraduate Research Papers

Authors :
Mihaiela R. Gugiu
Paul C. Gugiu
Robert Baldus
Source :
Journal of MultiDisciplinary Evaluation, Vol 8, Iss 19 (2012)
Publication Year :
2012
Publisher :
The Evaluation Center at Western Michigan University, 2012.

Abstract

Background: Educational researchers have long espoused the virtues of writing with regard to student cognitive skills. However, research on the reliability of the grades assigned to written papers reveals a high degree of contradiction, with some researchers concluding that the grades assigned are very reliable whereas others suggesting that they are so unreliable that random assignment of grades would have been almost as helpful. Purpose: The primary purpose of the study was to investigate the reliability of grades assigned to written reports. The secondary purpose was to illustrate the use of Generalizability Theory, specifically the fully-crossed two-facet model, for computing interrater reliability coefficients. Setting: The participants for this study were 29 undergraduate students enrolled in an introductory-level course on Political Behavior in Spring 2011 at a Midwest university. Intervention: Not applicable. Research Design: Students were randomly assigned to one of nine groups. Two-facet fully crossed G-study and D-study designs were used wherein two raters graded four assignments for 9 student groups—72 evaluations in total. The universe of admissible observations was deemed to be random for both raters and assignments, whereas the universe of generalization was deemed to be mixed (random for two raters but fixed for four assignments). Data Collection and Analysis: The semester-long project was assigned to groups consisting of an annotated bibliography, survey development, sampling design, and analysis and final report. Four grading rubrics were developed and utilized to evaluate the quality of each written report. Two-facet generalizability analyses were conducted to assess interrater reliability using software developed by one of the authors. Findings: This study found a very high interrater reliability coefficient (0.929) for only two raters who received no training in how to use the four grading rubrics. Keywords: grading; reliability; Generalizability Theory; writing

Subjects

Subjects :
Education
Social Sciences

Details

Language :
English
ISSN :
15568180
Volume :
8
Issue :
19
Database :
Directory of Open Access Journals
Journal :
Journal of MultiDisciplinary Evaluation
Publication Type :
Academic Journal
Accession number :
edsdoj.b0e703975a84648a5912076a5d6b488
Document Type :
article
Full Text :
https://doi.org/10.56645/jmde.v8i19.362