论文标题
部分可观测时空混沌系统的无模型预测
Towards an Enhanced Understanding of Bias in Pre-trained Neural Language Models: A Survey with Special Emphasis on Affective Bias
论文作者
论文摘要
随着一些研究开始讨论和报告NLP应用程序中的潜在偏见,深入审查了自然语言处理(NLP)的显着进步(NLP),尤其是在最近大型预训练的神经语言模型的出现中带来的显着进步。发现NLP中的偏见源自由人类编码的潜在历史偏见,这些偏见被NLP算法所延续甚至扩大的文本数据。我们提出了一项调查,以理解大型预训练的语言模型,分析它们在这些模型中发生的阶段,以及可以量化和减轻这些偏见的各种方式。考虑到在诸如商业,医疗保健,教育等现实世界系统中基于文本情感计算的下游任务的广泛适用性,我们特别强调在情感(情感)的背景下,即情感偏见,在大型预训练的语言模型中。我们介绍了各种偏见评估语料库的摘要,这些概述有助于帮助未来的研究,并讨论预训练的语言模型中偏见研究中的挑战。我们认为,我们试图在预训练的语言模型中对偏见进行全面看法,尤其是对情感偏见的探索将对对这个不断发展的领域感兴趣的研究人员非常有利。
The remarkable progress in Natural Language Processing (NLP) brought about by deep learning, particularly with the recent advent of large pre-trained neural language models, is brought into scrutiny as several studies began to discuss and report potential biases in NLP applications. Bias in NLP is found to originate from latent historical biases encoded by humans into textual data which gets perpetuated or even amplified by NLP algorithm. We present a survey to comprehend bias in large pre-trained language models, analyze the stages at which they occur in these models, and various ways in which these biases could be quantified and mitigated. Considering wide applicability of textual affective computing based downstream tasks in real-world systems such as business, healthcare, education, etc., we give a special emphasis on investigating bias in the context of affect (emotion) i.e., Affective Bias, in large pre-trained language models. We present a summary of various bias evaluation corpora that help to aid future research and discuss challenges in the research on bias in pre-trained language models. We believe that our attempt to draw a comprehensive view of bias in pre-trained language models, and especially the exploration of affective bias will be highly beneficial to researchers interested in this evolving field.