AIDB Daily Papers
LLMの落とし穴:常識よりも倫理を優先?物語の焦点バイアスの謎
※ 日本語タイトル・ポイントはAIによる自動生成です。正確な内容は原論文をご確認ください。
ポイント
- LLMが倫理的推論を優先し、常識的な理解を軽視する傾向を検証しました。
- 道徳的ジレンマに常識的な矛盾を埋め込んだ新しいベンチマークデータセットCoMoralを導入しました。
- LLMは、物語の主人公よりも脇役の矛盾を検出しやすいという物語焦点バイアスを示しました。
Abstract
Large Language Models (LLMs) are increasingly deployed across diverse real-world applications and user communities. As such, it is crucial that these models remain both morally grounded and knowledge-aware. In this work, we uncover a critical limitation of current LLMs -- their tendency to prioritize moral reasoning over commonsense understanding. To investigate this phenomenon, we introduce CoMoral, a novel benchmark dataset containing commonsense contradictions embedded within moral dilemmas. Through extensive evaluation of ten LLMs across different model sizes, we find that existing models consistently struggle to identify such contradictions without prior signal. Furthermore, we observe a pervasive narrative focus bias, wherein LLMs more readily detect commonsense contradictions when they are attributed to a secondary character rather than the primary (narrator) character. Our comprehensive analysis underscores the need for enhanced reasoning-aware training to improve the commonsense robustness of large language models.
Paper AI Chat
この論文のPDF全文を対象にAIに質問できます。
質問の例: