Mitigating the impact of biased artificial intelligence in emergency decision-making

被引:34
作者
Adam, Hammaad [1 ]
Balagopalan, Aparna [2 ]
Alsentzer, Emily [3 ,4 ,5 ]
Christia, Fotini [6 ,7 ]
Ghassemi, Marzyeh [4 ,8 ]
机构
[1] MIT, Inst Data Syst & Soc, Cambridge, MA 02139 USA
[2] MIT, Dept Elect Engn & Comp Sci, Cambridge, MA 02139 USA
[3] MIT, Harvard MIT Program Hlth Sci & Technol, Cambridge, MA 02139 USA
[4] MIT, Inst Med Engn Sci, Cambridge, MA 02139 USA
[5] Brigham & Womens Hosp, Div Gen Internal Med, Boston, MA 02115 USA
[6] MIT, Sociotechn Syst Res Ctr, Cambridge, MA 02139 USA
[7] MIT, Dept Polit Sci, Cambridge, MA 02139 USA
[8] Vector Inst, CIFAR AI Chair, Toronto, ON M5G IM1, Canada
来源
COMMUNICATIONS MEDICINE | 2022年 / 2卷 / 01期
关键词
RACIAL BIAS; MUSLIMS; HEALTH;
D O I
10.1038/s43856-022-00214-4
中图分类号
R-3 [医学研究方法]; R3 [基础医学];
学科分类号
1001 ;
摘要
Background Prior research has shown that artificial intelligence (AI) systems often encode biases against minority subgroups. However, little work has focused on ways to mitigate the harm discriminatory algorithms can cause in high-stakes settings such as medicine. Methods In this study, we experimentally evaluated the impact biased AI recommendations have on emergency decisions, where participants respond to mental health crises by calling for either medical or police assistance. We recruited 438 clinicians and 516 non-experts to participate in our web-based experiment. We evaluated participant decision-making with and without advice from biased and unbiased AI systems. We also varied the style of the AI advice, framing it either as prescriptive recommendations or descriptive flags. Results Participant decisions are unbiased without AI advice. However, both clinicians and non-experts are influenced by prescriptive recommendations from a biased algorithm, choosing police help more often in emergencies involving African-American or Muslim men. Crucially, using descriptive flags rather than prescriptive recommendations allows respondents to retain their original, unbiased decision-making. Conclusions Our work demonstrates the practical danger of using biased models in health contexts, and suggests that appropriately framing decision support can mitigate the effects of AI bias. These findings must be carefully considered in the many real-world clinical scenarios where inaccurate or biased models may be used to inform important decisions.
引用
收藏
页数:6
相关论文
共 38 条
[1]   Large language models associate Muslims with violence [J].
Abid, Abubakar ;
Farooqi, Maheen ;
Zou, James .
NATURE MACHINE INTELLIGENCE, 2021, 3 (06) :461-463
[2]   Pivotal trial of an autonomous AI-based diagnostic system for detection of diabetic retinopathy in primary care offices [J].
Abramoff, Michael D. ;
Lavin, Philip T. ;
Birch, Michele ;
Shah, Nilay ;
Folk, James C. .
NPJ DIGITAL MEDICINE, 2018, 1
[3]  
Angwin J., 2014, PROPUBLICA
[4]   The Road to Explainability is Paved with Bias: Measuring the Fairness of Explanations [J].
Balagopalan, Aparna ;
Zhang, Haoran ;
Hamidieh, Kimia ;
Hartvigsen, Thomas ;
Rudzicz, Frank ;
Ghassemi, Marzyeh .
PROCEEDINGS OF 2022 5TH ACM CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY, FACCT 2022, 2022, :1194-1206
[5]   Are Emily and Greg more employable than Lakisha and Jamal? A field experiment on labor market discrimination [J].
Bertrand, M ;
Mullainathan, S .
AMERICAN ECONOMIC REVIEW, 2004, 94 (04) :991-1013
[6]   Should I Add Recommendations to My Warning System? The RCRAFT Framework Can Answer This and Other Questions About Supporting the Assessment of Automation Designs [J].
Bouzekri, Elodie ;
Martinie, Celia ;
Palanque, Philippe ;
Atwood, Katrina ;
Gris, Christine .
HUMAN-COMPUTER INTERACTION, INTERACT 2021, PT IV, 2021, 12935 :405-429
[7]  
Brown K., 2021, The New York Times
[8]  
Brown TB, 2020, ADV NEUR IN, V33
[9]   The reduction of race and gender bias in clinical treatment recommendations using clinician peer networks in an experimental setting [J].
Centola, Damon ;
Guilbeault, Douglas ;
Sarkar, Urmimala ;
Khoong, Elaine ;
Zhang, Jingwen .
NATURE COMMUNICATIONS, 2021, 12 (01)
[10]  
Chohlas-Wood A., 2020, Brookings Institution's Series on AI and Bias