Global Times Call for Global Measures: Investigating Automated Essay Scoring in Linguistically-Diverse MOOCs

Authors

  • Erin Dawna Reilly University of Texas at Austin
  • Kyle M Wiliams University of Texas at Austin
  • Rose E Stafford University of Texas at Austin
  • Stephanie B. Corliss University of Texas at Austin
  • Janet C. Walkow University of Texas at Austin
  • Donna K. Kidwell University of Texas at Austin

DOI:

https://doi.org/10.24059/olj.v20i2.638

Keywords:

Massive open online courses, assessment, automated essay scoring, technology-enhanced language learning, multicultural assessment

Abstract

This paper utilizes a case-study design to discuss global aspects of massive open online course (MOOC) assessment. Drawing from the literature on open-course models and linguistic gatekeeping in education, we position freeform assessment in MOOCs as both challenging and valuable, with an emphasis on current practices and student resources. We report on the findings from a linguistically-diverse pharmacy MOOC, taught by a native English speaker, which utilized an automated essay scoring (AES) assignment to engage students in the application of course content. Native English speakers performed better on the assignment overall, across both automated- and human-graders. Additionally, our results suggest that the use of an AES system may disadvantage non-native English speakers, with agreement between instructor and AES scoring being significantly lower for non-native English speakers. Survey responses also revealed that students often utilized online translators, though analyses showed that this did not detrimentally affect essay grades. Pedagogical and future assignment suggestions are then outlined, utilizing a multicultural-lens and acknowledging the possibility of certain assessments disadvantaging non-native English speakers within an English-based MOOC system.

Downloads

Published

2016-05-16

Issue

Section

Section Two: MOOCs, Psychological Constructs, Communication Behaviors