AIDB Daily Papers
スーパーリサーチ:超深層・超広範な調査による大規模言語モデルを用いた高度に複雑な質問への回答
※ 日本語タイトル・ポイントはAIによる自動生成です。正確な内容は原論文をご確認ください。
ポイント
- 大規模言語モデル(LLM)が複雑な質問に答える能力を評価するSuper Researchタスクを導入し、研究計画への構造的な分解を統合した。
- 多様な視点を得るための超広範な検索と、反復的な質問を通じて不確実性を解消する超深層な調査を組み合わせる点が新しい。
- 専門家が作成した300の質問で評価した結果、Super Researchは検証可能なレポートを生成し、LLMの一般的な研究能力の指標となることが示された。
Abstract
While Large Language Models (LLMs) have demonstrated proficiency in Deep Research or Wide Search, their capacity to solve highly complex questions-those requiring long-horizon planning, massive evidence gathering, and synthesis across heterogeneous sources-remains largely unexplored. We introduce Super Research, a task for complex autonomous research tasks that integrates (i) structured decomposition into a research plan, (ii) super wide retrieval for diverse perspectives, and (iii) super deep investigation to resolve uncertainties through iterative queries. To evaluate this capability, we curated a benchmark of 300 expert-written questions across diverse domains, each requiring up to 100+ retrieval steps and 1,000+ web pages to reconcile conflicting evidence. Super Research produces verifiable reports with fine-grained citations and intermediate artifacts (e.g., outlines and tables) to ensure traceable reasoning. Furthermore, we present a graph-anchored auditing protocol that evaluates Super Research along five dimensions: Coverage, Logical Consistency, Report Utility, Objectivity and Citation Health. While super-complex questions may be infrequent in standard applications, Super Research serves as a critical ceiling evaluation and stress test for LLM capabilities. A model's proficiency within Super Research acts as a powerful proxy for its general research competence; success here suggests the robustness necessary to navigate nearly any subordinate research task. Leaderboard is available at: https://cnsdqd-dyb.github.io/Super-Research-Benchmark/
Paper AI Chat
この論文のPDF全文を対象にAIに質問できます。
質問の例: