AIDB Daily Papers
LLMは意図の不一致により複数ターンの会話で道に迷う
※ 日本語タイトル・ポイントはAIによる自動生成です。正確な内容は原論文をご確認ください。
ポイント
- LLMが複数ターンの会話で性能低下する原因を、モデルの能力不足ではなく意図のずれであると指摘した。
- 会話の構造的な曖昧さが意図のずれを生むため、モデルをスケールアップするだけでは解決できない点が重要である。
- Mediator-Assistantアーキテクチャにより意図理解とタスク実行を分離し、性能低下を軽減できることを示した。
Abstract
Multi-turn conversation has emerged as a predominant interaction paradigm for Large Language Models (LLMs). Users often employ follow-up questions to refine their intent, expecting LLMs to adapt dynamically. However, recent research reveals that LLMs suffer a substantial performance drop in multi-turn settings compared to single-turn interactions with fully specified instructions, a phenomenon termed ``Lost in Conversation'' (LiC). While this prior work attributes LiC to model unreliability, we argue that the root cause lies in an intent alignment gap rather than intrinsic capability deficits. In this paper, we first demonstrate that LiC is not a failure of model capability but rather a breakdown in interaction between users and LLMs. We theoretically show that scaling model size or improving training alone cannot resolve this gap, as it arises from structural ambiguity in conversational context rather than representational limitations. To address this, we propose to decouple intent understanding from task execution through a Mediator-Assistant architecture. By utilizing an experience-driven Mediator to explicate user inputs into explicit, well-structured instructions based on historical interaction patterns, our approach effectively bridges the gap between vague user intent and model interpretation. Experimental results demonstrate that this method significantly mitigates performance degradation in multi-turn conversations across diverse LLMs.
Paper AI Chat
この論文のPDF全文を対象にAIに質問できます。
質問の例: