论文标题

跨模型公平:在模型多重性下公平与伦理的经验研究

Cross-model Fairness: Empirical Study of Fairness and Ethics Under Model Multiplicity

论文作者

Sokol, Kacper, Kull, Meelis, Chan, Jeffrey, Salim, Flora

论文摘要

尽管数据驱动的预测模型是一种严格的技术结构,但它们可能在社会环境中运作,在这种社会环境中,良性工程选择需要隐式,间接和意外的现实生活后果。这种系统的公平性 - 与个人和群体有关 - 在这个领域是一个相关的考虑因素。算法可以区分各种受保护特征的人,而不管这些特性是否包含在数据中或通过代理变量可辨别。迄今为止,该概念主要是针对固定模型的,通常是在不同的分类阈值下,努力识别和消除其操作的不良,歧视性和可能是非法的方面。在这里,我们回溯了这个固定模型假设,以提出和探索一个新的跨模型公平定义,当从一组同样性能良好的模型(即基于公用事业的模型多重性)中选择一个预测变量时,可能会损害个体。由于一个人可能会在其他被认为是等效的模型中对某人进行不同的分类,因此该人可能会主张预测者授予他们最有利的结果,并采用可能对他人产生不利影响。我们通过二维示例和线性分类介绍了这种情况。然后,我们提出了一项基于现实生活中的预测模型和数据集的全面实证研究,这些模型和数据集受到算法公平社区的流行;最后,我们研究了跨模型公平性的分析特性及其在更广泛的背景下的分析特性。我们的发现表明,这种不公平在现实生活中很容易发现,仅凭技术手段就很难减轻这种不公平,因此很可能会降低预测性能。

While data-driven predictive models are a strictly technological construct, they may operate within a social context in which benign engineering choices entail implicit, indirect and unexpected real-life consequences. Fairness of such systems -- pertaining both to individuals and groups -- is one relevant consideration in this space; algorithms can discriminate people across various protected characteristics regardless of whether these properties are included in the data or discernible through proxy variables. To date, this notion has predominantly been studied for a fixed model, often under different classification thresholds, striving to identify and eradicate undesirable, discriminative and possibly unlawful aspects of its operation. Here, we backtrack on this fixed model assumption to propose and explore a novel definition of cross-model fairness where individuals can be harmed when one predictor is chosen ad hoc from a group of equally well performing models, i.e., in view of utility-based model multiplicity. Since a person may be classified differently across models that are otherwise considered equivalent, this individual could argue for a predictor granting them the most favourable outcome, employing which may have adverse effects on other people. We introduce this scenario with a two-dimensional example and linear classification; then, we present a comprehensive empirical study based on real-life predictive models and data sets that are popular with the algorithmic fairness community; finally, we investigate analytical properties of cross-model fairness and its ramifications in a broader context. Our findings suggest that such unfairness can be readily found in real life and it may be difficult to mitigate by technical means alone as doing so is likely to degrade predictive performance.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源