AIDB Daily Papers
LLM拡張調査における修正難易度と最適なサンプル配分
※ 日本語タイトル・ポイントはAIによる自動生成です。正確な内容は原論文をご確認ください。
ポイント
- 大規模言語モデル(LLM)を用いた調査で、人間による回答者数を最適に配分する手法を提案した。
- LLMの予測精度が質問ごとにばらつくため、人間による修正が必要な難易度を算出し、LLMの信頼性が低い質問に人間を多く割り当てる。
- 過去のデータから修正難易度を予測するメタ学習により、パイロット調査なしで効率的なサンプル配分を実現し、平均11%のMSE削減を達成した。
Abstract
Large Language Models can generate synthetic survey responses at low cost, but their accuracy varies unpredictably across questions. We study the design problem of allocating a fixed budget of human respondents across estimation tasks when cheap LLM predictions are available for every task. Our framework combines three components. First, building on Prediction-Powered Inference, we characterize a question-specific rectification difficulty that governs how quickly the estimator's variance decreases with human sample size. Second, we derive a closed-form optimal allocation rule that directs more human labels to tasks where the LLM is least reliable. Third, since rectification difficulty depends on unobserved human responses for new surveys, we propose a meta-learning approach, trained on historical data, that predicts it for entirely new tasks without pilot data. The framework extends to general M-estimation, covering regression coefficients and multinomial logit partworths for conjoint analysis. We validate the framework on two datasets spanning different domains, question types, and LLMs, showing that our approach captures 61-79% of the theoretically attainable efficiency gains, achieving 11.4% and 10.5% MSE reductions without requiring any pilot human data for the target survey.
Paper AI Chat
この論文のPDF全文を対象にAIに質問できます。
質問の例: