AIDB Daily Papers
HumanLM:状態整合によるユーザーシミュレーションは応答模倣を凌駕する
※ 日本語タイトル・ポイントはAIによる自動生成です。正確な内容は原論文をご確認ください。
ポイント
- 本研究では、ユーザーの潜在状態を反映した応答を生成する新しいフレームワークHumanLMを提案した。
- 既存のユーザーシミュレータの限界を克服し、心理学に基づいた潜在状態を導入することで、よりリアルなユーザーを再現する。
- HumanLMは、大規模ベンチマークHumanualで既存手法を大幅に上回り、実時間シミュレーションでも高い性能を示した。
Abstract
Large Language Models (LLMs) are increasingly used to simulate how specific users respond to a given context, enabling more user-centric applications that rely on user feedback. However, existing user simulators mostly imitate surface-level patterns and language styles, which fail to reflect the underlying states of real users (e.g., beliefs and emotions). To address these limitations, we propose a novel training framework, HumanLM, which builds user simulators that accurately reflect real users. Our key insight is that, in addition to generating responses, the model should generate natural-language latent states that align with ground-truth responses through reinforcement learning. These latent states correspond to a set of psychologically grounded state dimensions that drive how real users respond. HumanLM further synthesizes these aligned latent states into responses that accurately represent real users. For extensive evaluation, we develop Humanual, a comprehensive benchmark for simulating real users based on public data. Humanual consists of six large-scale datasets with 26k users and 216k responses in total, spanning diverse tasks such as generating user responses to daily life issues, political blogs, and chat sessions with LLM assistants. Across datasets, HumanLM significantly outperforms alternative approaches, achieving an average relative improvement of 16.3% in alignment scores from an LLM judge. In a real-time simulation study with 111 participants, HumanLM achieves the highest similarity to real user responses and competitive human-likeness scores.
Paper AI Chat
この論文のPDF全文を対象にAIに質問できます。
質問の例: