Question Difficulty Ranking for Multiple-Choice Reading Comprehension
CoRR(2024)
摘要
Multiple-choice (MC) tests are an efficient method to assess English
learners. It is useful for test creators to rank candidate MC questions by
difficulty during exam curation. Typically, the difficulty is determined by
having human test takers trial the questions in a pretesting stage. However,
this is expensive and not scalable. Therefore, we explore automated approaches
to rank MC questions by difficulty. However, there is limited data for explicit
training of a system for difficulty scores. Hence, we compare task transfer and
zero-shot approaches: task transfer adapts level classification and reading
comprehension systems for difficulty ranking while zero-shot prompting of
instruction finetuned language models contrasts absolute assessment against
comparative. It is found that level classification transfers better than
reading comprehension. Additionally, zero-shot comparative assessment is more
effective at difficulty ranking than the absolute assessment and even the task
transfer approaches at question difficulty ranking with a Spearman's
correlation of 40.4
correlation.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要