BriefGPT.xyz
Aug, 2023
大型语言模型中的性别偏见和刻板印象
Gender bias and stereotypes in Large Language Models
HTML
PDF
Hadas Kotek, Rikker Dockum, David Q. Sun
TL;DR
通过对四个最近发表的大型语言模型进行测试,我们发现大型语言模型在性别刻板印象、职业偏见和句子结构等方面表现出偏见,它们更倾向于选择与性别刻板印象相关的职业,并且在提供自己选择的解释时常常存在事实错误,对于这些模型的偏见行为,我们需要谨慎测试以确保它们对待边缘化个体和社区是公平的。
Abstract
large language models
(LLMs) have made substantial progress in the past several months, shattering state-of-the-art benchmarks in many domains. This paper investigates LLMs' behavior with respect to
gender stereotypes
→