AIDB Daily Papers
高性能LLMは協調性も高いとは限らない?コストゼロの協調における失敗要因
※ 日本語タイトル・ポイントはAIによる自動生成です。正確な内容は原論文をご確認ください。
ポイント
- LLMエージェントの協調性を、戦略的複雑さを排除した環境で検証する新たな実験環境を構築した。
- 能力が高いLLMほど協調性が高いとは限らず、明示的な指示があっても協調を怠る場合があることが判明した。
- 協調性の低いモデルにはプロトコル明示、協調性の弱いモデルにはインセンティブ付与が有効であることが示唆された。
Abstract
Large language model (LLM) agents increasingly coordinate in multi-agent systems, yet we lack an understanding of where and why cooperation failures may arise. In many real-world coordination problems, from knowledge sharing in organizations to code documentation, helping others carries negligible personal cost while generating substantial collective benefits. However, whether LLM agents cooperate when helping neither benefits nor harms the helper, while being given explicit instructions to do so, remains unknown. We build a multi-agent setup designed to study cooperative behavior in a frictionless environment, removing all strategic complexity from cooperation. We find that capability does not predict cooperation: OpenAI o3 achieves only 17% of optimal collective performance while OpenAI o3-mini reaches 50%, despite identical instructions to maximize group revenue. Through a causal decomposition that automates one side of agent communication, we separate cooperation failures from competence failures, tracing their origins through agent reasoning analysis. Testing targeted interventions, we find that explicit protocols double performance for low-competence models, and tiny sharing incentives improve models with weak cooperation. Our findings suggest that scaling intelligence alone will not solve coordination problems in multi-agent systems and will require deliberate cooperative design, even when helping others costs nothing.
Paper AI Chat
この論文のPDF全文を対象にAIに質問できます。
質問の例: