论文标题
更好的分散注意力:基于变压器的干扰物生成和多项选择问题过滤
Better Distractions: Transformer-based Distractor Generation and Multiple Choice Question Filtering
论文作者
论文摘要
对于教育领域,能够产生语义上正确且与教育相关的多项选择问题(MCQ)可能会产生很大的影响。虽然问题产生本身是一个积极的研究主题,但产生干扰因素(不正确的多项选择选项)受到了较少的关注。错过的机会,因为该领域仍然有很大的改进空间。在这项工作中,我们使用Race数据集训练GPT-2语言模型为给定的问题和文本上下文生成三个干扰素。接下来,我们训练BERT语言模型回答MCQ,并使用此模型作为过滤器,仅选择可以回答的问题,因此可能是有道理的。为了评估我们的工作,我们首先使用文本生成指标,这表明我们的模型表现优于较早的干扰物生成工作(DG),并实现最先进的性能。同样,通过计算问答答案能力,我们表明较大的基本模型会带来更好的性能。此外,我们进行了一项人类评估研究,该研究证实了生成的问题的质量,但没有显示质量检查过滤器的统计学意义。
For the field of education, being able to generate semantically correct and educationally relevant multiple choice questions (MCQs) could have a large impact. While question generation itself is an active research topic, generating distractors (the incorrect multiple choice options) receives much less attention. A missed opportunity, since there is still a lot of room for improvement in this area. In this work, we train a GPT-2 language model to generate three distractors for a given question and text context, using the RACE dataset. Next, we train a BERT language model to answer MCQs, and use this model as a filter, to select only questions that can be answered and therefore presumably make sense. To evaluate our work, we start by using text generation metrics, which show that our model outperforms earlier work on distractor generation (DG) and achieves state-of-the-art performance. Also, by calculating the question answering ability, we show that larger base models lead to better performance. Moreover, we conducted a human evaluation study, which confirmed the quality of the generated questions, but showed no statistically significant effect of the QA filter.