AIDB Daily Papers
Co-FactChecker:大規模推論モデルを活用した人間とAIの協調型主張検証フレームワーク
※ 日本語タイトル・ポイントはAIによる自動生成です。正確な内容は原論文をご確認ください。
ポイント
- 専門家による知識とLLMの推論能力を組み合わせ、主張検証のミスマッチを解消するフレームワークを提案した。
- 既存のLLMは自然言語でのフィードバックに弱いため、思考過程を共有し編集する新しいインタラクションを導入した。
- 自動評価と人間評価の両方で、Co-FactCheckerが既存手法を上回り、高品質な推論と判断を実現することを示した。
Abstract
Professional fact-checkers rely on domain knowledge and deep contextual understanding to verify claims. Large language models (LLMs) and large reasoning models (LRMs) lack such grounding and primarily reason from available evidence alone, creating a mismatch between expert-led and fully automated claim verification. To mitigate this gap, we posit human-AI collaboration as a more promising path forward, where expert feedback, grounded in real-world knowledge and domain expertise, guides the model's reasoning. However, existing LRMs are hard to calibrate to natural language feedback, particularly in a multi-turn interaction setup. We propose Co-FactChecker, a framework for human-AI collaborative claim verification. We introduce a new interaction paradigm that treats the model's thinking trace as a shared scratchpad. Co-FactChecker translates expert feedback into trace-edits that introduce targeted modifications to the trace, sidestepping the shortcomings of dialogue-based interaction. We provide theoretical results showing that trace-editing offers advantages over multi-turn dialogue, and our automatic evaluations demonstrate that Co-FactChecker outperforms existing autonomous and human-AI collaboration approaches. Human evaluations further show that Co-FactChecker is preferred over multi-turn dialogue, producing higher quality reasoning and verdicts along with relatively easier to interpret and more useful thinking traces.
Paper AI Chat
この論文のPDF全文を対象にAIに質問できます。
質問の例: