喵ID:MWwwBm免责声明

Examining risks of racial biases in NLP tools for child protective services

基本信息

DOI:
10.1145/3593013.3594094
发表时间:
2023-05
期刊:
Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency
影响因子:
--
通讯作者:
Anjalie Field;Amanda Coston;Nupoor Gandhi;A. Chouldechova;Emily Putnam-Hornstein;David Steier;Yulia Tsvetkov
中科院分区:
其他
文献类型:
--
作者: Anjalie Field;Amanda Coston;Nupoor Gandhi;A. Chouldechova;Emily Putnam-Hornstein;David Steier;Yulia Tsvetkov研究方向: -- MeSH主题词: --
关键词: --
来源链接:pubmed详情页地址

文献摘要

Although much literature has established the presence of demographic bias in natural language processing (NLP) models, most work relies on curated bias metrics that may not be reflective of real-world applications. At the same time, practitioners are increasingly using algorithmic tools in high-stakes settings, with particular recent interest in NLP. In this work, we focus on one such setting: child protective services (CPS). CPS workers often write copious free-form text notes about families they are working with, and CPS agencies are actively seeking to deploy NLP models to leverage these data. Given well-established racial bias in this setting, we investigate possible ways deployed NLP is liable to increase racial disparities. We specifically examine word statistics within notes and algorithmic fairness in risk prediction, coreference resolution, and named entity recognition (NER). We document consistent algorithmic unfairness in NER models, possible algorithmic unfairness in coreference resolution models, and little evidence of exacerbated racial bias in risk prediction. While there is existing pronounced criticism of risk prediction, our results expose previously undocumented risks of racial bias in realistic information extraction systems, highlighting potential concerns in deploying them, even though they may appear more benign. Our work serves as a rare realistic examination of NLP algorithmic fairness in a potential deployed setting and a timely investigation of a specific risk associated with deploying NLP in CPS settings.
尽管许多文献已经证实自然语言处理(NLP)模型中存在人口统计学偏差,但大多数研究依赖于精心设计的偏差指标,这些指标可能无法反映现实世界的应用情况。与此同时,从业者在高风险环境中越来越多地使用算法工具,近期对NLP尤其感兴趣。在这项工作中,我们关注这样一个环境:儿童保护服务(CPS)。CPS工作人员经常撰写大量关于他们所服务家庭的自由格式文本记录,并且CPS机构正在积极寻求部署NLP模型以利用这些数据。鉴于在这种环境中已确定存在种族偏差,我们研究已部署的NLP可能会增加种族差异的方式。我们专门检查记录中的词汇统计以及风险预测、共指消解和命名实体识别(NER)中的算法公平性。我们记录了NER模型中持续存在的算法不公平性、共指消解模型中可能存在的算法不公平性,以及在风险预测中种族偏差加剧的证据很少。虽然对风险预测已有明显的批评,但我们的结果揭示了现实信息提取系统中以前未记录的种族偏差风险,凸显了部署这些系统时的潜在担忧,尽管它们可能看起来更无害。我们的工作是对潜在部署环境中NLP算法公平性的一次罕见的现实检验,也是对在CPS环境中部署NLP相关特定风险的一次及时调查。
参考文献(68)
被引文献(6)

数据更新时间:{{ references.updateTime }}

Anjalie Field;Amanda Coston;Nupoor Gandhi;A. Chouldechova;Emily Putnam-Hornstein;David Steier;Yulia Tsvetkov
通讯地址:
--
所属机构:
--
电子邮件地址:
--
免责声明免责声明
1、猫眼课题宝专注于为科研工作者提供省时、高效的文献资源检索和预览服务;
2、网站中的文献信息均来自公开、合规、透明的互联网文献查询网站,可以通过页面中的“来源链接”跳转数据网站。
3、在猫眼课题宝点击“求助全文”按钮,发布文献应助需求时求助者需要支付50喵币作为应助成功后的答谢给应助者,发送到用助者账户中。若文献求助失败支付的50喵币将退还至求助者账户中。所支付的喵币仅作为答谢,而不是作为文献的“购买”费用,平台也不从中收取任何费用,
4、特别提醒用户通过求助获得的文献原文仅用户个人学习使用,不得用于商业用途,否则一切风险由用户本人承担;
5、本平台尊重知识产权,如果权利所有者认为平台内容侵犯了其合法权益,可以通过本平台提供的版权投诉渠道提出投诉。一经核实,我们将立即采取措施删除/下架/断链等措施。
我已知晓