Powergrading: a Clustering Approach to Amplify Human Effort for Short Answer Grading
- Sumit Basu ,
- Chuck Jacobs ,
- Lucy Vanderwende
Transactions of the ACL |
We introduce a new approach to the machine-assisted grading of short answer questions. We follow past work in automated grading by first training a similarity metric between student responses, but then go on to use this metric to group responses into clusters and subclusters. The resulting groupings allow teachers to grade multiple responses with a single action, provide rich feedback to groups of similar answers, and discover modalities of misunderstanding among students; we refer to this amplification of grader effort as “powergrading.” We develop the means to further reduce teacher effort by automatically performing actions when an answer key is available. We show results in terms of grading progress with a small “budget” of human actions, both from our method and an LDA-based approach, on a test corpus of 10 questions answered by 698 respondents.
Publication Downloads
Powergrading Short Answer Grading Corpus
October 4, 2013
This corpus contains the original data analyzed in the following paper: Basu, Jacobs, and Vanderwende, "Powergrading: a Clustering Approach to Amplify Human Effort for Short Answer Grading,” Transactions of the ACL, 2013. It consists of responses from 100 + 698 crowdsourced workers to each of 20 short-answer questions. These questions are taken from the 100 questions published by the United States Citizenship and Immigration Services as preparation for the citizenship test. It also contains labels of response correctness (grades) from three judges for a subset of 10 questions for the set of 698 responses (3 x 6980 labels).