AIDB Daily Papers
CoAct:人間とAIの協調によるLLM選好学習の新フレームワーク
※ 日本語タイトル・ポイントはAIによる自動生成です。正確な内容は原論文をご確認ください。
ポイント
- 人間とAIの協調により、LLMの選好学習を効率化するCoActフレームワークを提案した。
- AIによる自己評価と人間の専門家による評価を組み合わせ、データコストと信頼性の課題を解決した。
- 3つの推論ベンチマークで既存手法を大幅に上回る性能向上を達成した。
Abstract
Learning from preference-based feedback has become an effective approach for aligning LLMs across diverse tasks. However, high-quality human-annotated preference data remains expensive and scarce. Existing methods address this challenge through either self-rewarding, which scales by using purely AI-generated labels but risks unreliability, or active learning, which ensures quality through oracle annotation but cannot fully leverage unlabeled data. In this paper, we present CoAct, a novel framework that synergistically combines self-rewarding and active learning through strategic human-AI collaboration. CoAct leverages self-consistency to identify both reliable self-labeled data and samples that require oracle verification. Additionally, oracle feedback guides the model to generate new instructions within its solvable capability. Evaluated on three reasoning benchmarks across two model families, CoAct achieves average improvements of +13.25% on GSM8K, +8.19% on MATH, and +13.16% on WebInstruct, consistently outperforming all baselines.
Paper AI Chat
この論文のPDF全文を対象にAIに質問できます。
質問の例: