AIDB Daily Papers
LLM推論のための脳にヒントを得たグラフ型マルチエージェントシステム
※ 日本語タイトル・ポイントはAIによる自動生成です。正確な内容は原論文をご確認ください。
ポイント
- 複雑な推論タスクにおいてLLMの限界を克服するため、脳のグローバルワークスペース理論に着想を得た新しいアーキテクチャBIGMASを提案した。
- BIGMASは、動的に構築されるグラフ構造で連携する専門LLMエージェント群により、タスク固有の最適化と全体的な視点での推論を可能にする。
- Game24などのタスクで実験した結果、BIGMASは既存のマルチエージェント手法を凌駕し、モデルレベルの推論能力を補完することを示した。
Abstract
Large Language Models (LLMs) have demonstrated remarkable capabilities across a wide range of language tasks, yet complex multi-step reasoning remains a fundamental challenge. While Large Reasoning Models (LRMs) equipped with extended chain-of-thought mechanisms demonstrate improved performance over standard LLMs, both model types still suffer from accuracy collapse on sufficiently complex tasks, suggesting that scaling model-level reasoning alone is insufficient. Inspired by the global workspace theory of human cognition, we propose Brain-Inspired Graph Multi-Agent Systems (BIGMAS), in which specialized LLM agents are organized as nodes in a dynamically constructed directed graph and coordinate exclusively through a centralized shared workspace. A problem-adaptive GraphDesigner constructs task-specific agent topologies, while a global Orchestrator leverages the complete shared state for routing decisions, overcoming the local-view bottleneck of reactive approaches. Experiments on Game24, Six Fives, and Tower of London across six frontier LLMs demonstrate that BIGMAS consistently improves reasoning performance for both standard LLMs and LRMs, outperforming existing multi-agent baselines including ReAct and Tree of Thoughts, showing that multi-agent architectural design provides complementary gains orthogonal to model-level reasoning enhancements.
Paper AI Chat
この論文のPDF全文を対象にAIに質問できます。
質問の例: