@inproceedings{610bef3bc811445d8e4da3005e3a4c7e,
title = "Exploring effect of rater on prediction error in automatic text grading for open-ended question",
abstract = "This paper aims to explore the way of evaluating the automatic text grader for open-ended questions by considering the relationships among raters, grade levels, and prediction errors. The open-ended question in this study was about aurora and required knowledge of earth science and physics. Each student's response was graded from 0 to 10 points by three raters. The automatic grading systems were designed as support-vector-machine regression models with linear, quadratic, and RBF kernel respectively. The three kinds of regression models were separately trained through grades by three human raters and the average grades. The preliminary evaluation with 391 students' data shows results as the following: (1) The higher the grade-level is, the larger the prediction error is. (2) The ranks of prediction errors of human-rater-trained models at three grade levels are different. (3) The model trained through the average grades has the best performance at all three grade-levels no matter what the kind of kernel is. These results suggest that examining the prediction errors of models in detail on different grade-levels is worthwhile for finding the best matching between raters' grades and models.",
keywords = "Automatic grader, Prediction error, Rater, SVM, Science learning, Testing",
author = "Lee, {Che Di} and Chang, {Chun Yen} and Li, {Tsai Yen} and Fu, {Hsieh Hai} and Jen, {Tsung Hau} and Lee, {Kang Che}",
year = "2009",
language = "English",
isbn = "9789868473539",
series = "Proceedings of the 17th International Conference on Computers in Education, ICCE 2009",
pages = "462--466",
booktitle = "Proceedings of the 17th International Conference on Computers in Education, ICCE 2009",
note = "17th International Conference on Computers in Education, ICCE 2009 ; Conference date: 30-11-2009 Through 04-12-2009",
}