AIDB Daily Papers
裏の意味を読み解く:LLMは行間を理解できるのか?
※ 日本語タイトル・ポイントはAIによる自動生成です。正確な内容は原論文をご確認ください。
ポイント
- 本研究では、LLMが隠喩や暗示といった「行間」を理解し、コミュニケーションに活用できるかを検証した。
- LLMは字義通りの解釈に偏る傾向が強く、ニュアンスを考慮したコミュニケーションが苦手であることが明らかになった。
- 共通認識の利用やペルソナ設定により改善が見られるものの、LLMの創造的なコミュニケーション能力には課題が残る。
Abstract
Human communication is fundamentally creative, and often makes use of subtext -- implied meaning that goes beyond the literal content of the text. Here, we systematically study whether language models can use subtext in communicative settings, and introduce four new evaluation suites to assess these capabilities. Our evaluation settings range from writing & interpreting allegories to playing multi-agent and multi-modal games inspired by the rules of board games like Dixit. We find that frontier models generally exhibit a strong bias towards overly literal, explicit communication, and thereby fail to account for nuanced constraints -- even the best performing models generate literal clues 60% of times in one of our environments -- Visual Allusions. However, we find that some models can sometimes make use of common ground with another party to help them communicate with subtext, achieving 30%-50% reduction in overly literal clues; but they struggle at inferring presence of a common ground when not explicitly stated. For allegory understanding, we find paratextual and persona conditions to significantly shift the interpretation of subtext. Overall, our work provides quantifiable measures for an inherently complex and subjective phenomenon like subtext and reveals many weaknesses and idiosyncrasies of current LLMs. We hope this research to inspire future work towards socially grounded creative communication and reasoning.
Paper AI Chat
この論文のPDF全文を対象にAIに質問できます。
質問の例: