AIDB Daily Papers
LLMにおけるアイデンティティを巡るユーモアの反事実的不公平性の調査
※ 日本語タイトル・ポイントはAIによる自動生成です。正確な内容は原論文をご確認ください。
ポイント
- LLMがユーモアにどう反応するかを、話者や対象を入れ替えて調査した。
- この研究は、LLMが学習データから社会的な偏見を内包していることを明らかにする点で重要である。
- 特権的な話者からのジョークは拒否されやすく、悪意があると判断されやすく、社会的危害が大きいと評価された。
Abstract
Humor holds up a mirror to social perception: what we find funny often reflects who we are and how we judge others. When language models engage with humor, their reactions expose the social assumptions they have internalized from training data. In this paper, we investigate counterfactual unfairness through humor by observing how the model's responses change when we swap who speaks and who is addressed while holding other factors constant. Our framework spans three tasks: humor generation refusal, speaker intention inference, and relational/societal impact prediction, covering both identity-agnostic humor and identity-specific disparagement humor. We introduce interpretable bias metrics that capture asymmetric patterns under identity swaps. Experiments across state-of-the-art models reveal consistent relational disparities: jokes told by privileged speakers are refused up to 67.5% more often, judged as malicious 64.7% more frequently, and rated up to 1.5 points higher in social harm on a 5-point scale. These patterns highlight how sensitivity and stereotyping coexist in generative models, complicating efforts toward fairness and cultural alignment.
Paper AI Chat
この論文のPDF全文を対象にAIに質問できます。
質問の例: