AIDB Daily Papers
CoopEval:社会的なジレンマにおける協力維持メカニズムとLLMエージェントのベンチマーク
※ 日本語タイトル・ポイントはAIによる自動生成です。正確な内容は原論文をご確認ください。
ポイント
- LLMエージェントが他のエージェントと効果的かつ安全に相互作用するためのメカニズムを比較研究しました。
- 囚人のジレンマなどのゲームにおいて、推論能力の高いLLMが協力しなくなる傾向に対処することが重要です。
- 契約と仲介がLLMモデル間の協調を促進する上で最も効果的であり、進化的な圧力下でさらに効果を発揮することが示されました。
Abstract
It is increasingly important that LLM agents interact effectively and safely with other goal-pursuing agents, yet, recent works report the opposite trend: LLMs with stronger reasoning capabilities behave _less_ cooperatively in mixed-motive games such as the prisoner's dilemma and public goods settings. Indeed, our experiments show that recent models -- with or without reasoning enabled -- consistently defect in single-shot social dilemmas. To tackle this safety concern, we present the first comparative study of game-theoretic mechanisms that are designed to enable cooperative outcomes between rational agents _in equilibrium_. Across four social dilemmas testing distinct components of robust cooperation, we evaluate the following mechanisms: (1) repeating the game for many rounds, (2) reputation systems, (3) third-party mediators to delegate decision making to, and (4) contract agreements for outcome-conditional payments between players. Among our findings, we establish that contracting and mediation are most effective in achieving cooperative outcomes between capable LLM models, and that repetition-induced cooperation deteriorates drastically when co-players vary. Moreover, we demonstrate that these cooperation mechanisms become _more effective_ under evolutionary pressures to maximize individual payoffs.
Paper AI Chat
この論文のPDF全文を対象にAIに質問できます。
質問の例: