Large language models associate Muslims with violence

被引:82
作者
Abid, Abubakar [1 ]
Farooqi, Maheen [2 ]
Zou, James [3 ]
机构
[1] Stanford Univ, Dept Elect Engn, Stanford, CA 94305 USA
[2] McMaster Univ, Dept Hlth Res Methods Evidence & Impact, Hamilton, ON, Canada
[3] Stanford Univ, Dept Biomed Data Sci, Stanford, CA 94305 USA
基金
美国国家科学基金会;
关键词
D O I
10.1038/s42256-021-00359-2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large language models, which are increasingly used in AI applications, display undesirable stereotypes such as persistent associations between Muslims and violence. New approaches are needed to systematically reduce the harmful bias of language models in deployment.
引用
收藏
页码:461 / 463
页数:3
相关论文
共 14 条
[1]  
Bender E. M., 2021, P ACM C FAIRN ACC TR, P610
[2]  
Bolukbasi T, 2016, ADV NEUR IN, V29
[3]  
Bordia S., 2019, P C N AM CHAPT ASS C
[4]  
Brown Tom B., 2020, NEURIPS 2020
[5]  
Dai A.M., 2015, Adv. Neural Inf. Process. Syst.
[6]  
Kitaev N., 2020, P INT C LEARN REPR I
[7]  
Lewis M., 2020, P 58 ANN M ASS COMP, DOI 10.18653/v1/2020.acl-main.703
[8]  
Li X., PREPRINT
[9]  
Lu Kaiji, 2020, LOGIC LANGUAGE SECUR, P189
[10]  
Mikolov T., 2013, P WORKSHOP ICLR