Automated Scoring of Constructed‐Response Science Items: Prospects and Obstacles

EDUCATIONAL MEASUREMENT-ISSUES AND PRACTICE(2014)

引用 71|浏览5
暂无评分
摘要
Content-based automated scoring has been applied in a variety of science domains. However, many prior applications involved simplified scoring rubrics without considering rubrics representing multiple levels of understanding. This study tested a concept-based scoring tool for content-based scoring, c-rater (TM), for four science items with rubrics aiming to differentiate among multiple levels of understanding. The items showed moderate to good agreement with human scores. The findings suggest that automated scoring has the potential to score constructed-response items with complex scoring rubrics, but in its current design cannot replace human raters. This article discusses sources of disagreement and factors that could potentially improve the accuracy of concept-based automated scoring.
更多
查看译文
关键词
automated scoring,constructed-response items,c-rater (TM),science assessment
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要