论文标题
作弊自动简短答案分级:形容词和副词的对抗性
Cheating Automatic Short Answer Grading: On the Adversarial Usage of Adjectives and Adverbs
论文作者
论文摘要
自动分级模型是在大型学生身体指导中节省的时间和精力的重视。特别是随着对大规模标准化测试的教育数字化和兴趣的日益增加,自动分级的普及已上升到广泛可用和使用商业解决方案的地步。但是,对于简短的答案格式,由于自然语言的歧义和多功能性,自动分级是具有挑战性的。虽然自动简短的答案分级模型开始与某些数据集的人类绩效进行比较,但它们的鲁棒性,尤其是对对抗性操纵数据的稳健性是值得怀疑的。分级模型中可利用的漏洞可能会产生深远的后果,从作弊的学生获得不当信用的作弊到完全破坏自动分级,即使大多数预测是有效的。在本文中,我们设计了针对教育简短答案等级方案量身定制的黑盒对抗攻击,以研究分级模型的鲁棒性。在我们的攻击中,我们将形容词和副词插入不正确的学生答案的自然位置,使该模型预测它们是正确的。我们观察到使用最先进的模型BERT和T5在10到22个百分点之间丧失了预测准确性。尽管我们的攻击使答案对于人类在我们的实验中显得不那么自然,但它并没有显着增加毕业生对作弊的怀疑。根据我们的实验,我们提供了建议在实践中更安全地利用自动分级系统的建议。
Automatic grading models are valued for the time and effort saved during the instruction of large student bodies. Especially with the increasing digitization of education and interest in large-scale standardized testing, the popularity of automatic grading has risen to the point where commercial solutions are widely available and used. However, for short answer formats, automatic grading is challenging due to natural language ambiguity and versatility. While automatic short answer grading models are beginning to compare to human performance on some datasets, their robustness, especially to adversarially manipulated data, is questionable. Exploitable vulnerabilities in grading models can have far-reaching consequences ranging from cheating students receiving undeserved credit to undermining automatic grading altogether - even when most predictions are valid. In this paper, we devise a black-box adversarial attack tailored to the educational short answer grading scenario to investigate the grading models' robustness. In our attack, we insert adjectives and adverbs into natural places of incorrect student answers, fooling the model into predicting them as correct. We observed a loss of prediction accuracy between 10 and 22 percentage points using the state-of-the-art models BERT and T5. While our attack made answers appear less natural to humans in our experiments, it did not significantly increase the graders' suspicions of cheating. Based on our experiments, we provide recommendations for utilizing automatic grading systems more safely in practice.