AIDB Daily Papers
LLMの推論は連鎖ではなく潜在空間に宿る
※ 日本語タイトル・ポイントはAIによる自動生成です。正確な内容は原論文をご確認ください。
ポイント
- 本研究は、LLMの推論を表面的な思考連鎖ではなく、潜在空間の軌跡形成として捉えるべきだと主張する。
- この視点の転換は、解釈性や推論ベンチマーク、推論時の介入方法に関する議論に不可欠である。
- 現在の証拠は、LLMの推論が潜在状態のダイナミクスによって最もよく説明されるという仮説を支持する。
Abstract
This position paper argues that large language model (LLM) reasoning should be studied as latent-state trajectory formation rather than as faithful surface chain-of-thought (CoT). This matters because claims about faithfulness, interpretability, reasoning benchmarks, and inference-time intervention all depend on what the field takes the primary object of reasoning to be. We ask what that object should be once three often-confounded factors are separated and formalize three competing hypotheses: H1, reasoning is primarily mediated by latent-state trajectories; H2, reasoning is primarily mediated by explicit surface CoT; and H0, most apparent reasoning gains are better explained by generic serial compute than by any privileged representational object. Reorganizing recent empirical, mechanistic, and survey work under this framework, and adding compute-audited worked exemplars that factorize surface traces, latent interventions, and matched budget expansions, we find that current evidence most strongly supports H1 as a default working hypothesis rather than as a task-independent verdict. We therefore make two recommendations: the field should treat latent-state dynamics as the default object of study for LLM reasoning, and it should evaluate reasoning with designs that explicitly disentangle surface traces, latent states, and serial compute.
Paper AI Chat
この論文のPDF全文を対象にAIに質問できます。
質問の例: