AIDB Daily Papers
LatentPilot:潜在的視覚推論による先読みでシーンを認識する視覚言語ナビゲーション
※ 日本語タイトル・ポイントはAIによる自動生成です。正確な内容は原論文をご確認ください。
ポイント
- 行動がもたらす未来の視覚的変化を予測する新しいVLNモデル、LatentPilotを提案しました。
- 過去と現在の視覚情報だけでなく、行動と視覚世界の因果関係を学習し、よりロバストな意思決定を可能にします。
- R2R-CE, RxR-CE, R2R-PEベンチマークで最高性能を達成し、実ロボット実験でも優れた環境理解を示しました。
Abstract
Existing vision-and-language navigation (VLN) models primarily reason over past and current visual observations, while largely ignoring the future visual dynamics induced by actions. As a result, they often lack an effective understanding of the causal relationship between actions and how the visual world changes, limiting robust decision-making. Humans, in contrast, can imagine the near future by leveraging action-dynamics causality, which improves both environmental understanding and navigation choices. Inspired by this capability, we propose LatentPilot, a new paradigm that exploits future observations during training as a valuable data source to learn action-conditioned visual dynamics, while requiring no access to future frames at inference. Concretely, we propose a flywheel-style training mechanism that iteratively collects on-policy trajectories and retrains the model to better match the agent's behavior distribution, with an expert takeover triggered when the agent deviates excessively. LatentPilot further learns visual latent tokens without explicit supervision; these latent tokens attend globally in a continuous latent space and are carried across steps, serving as both the current output and the next input, thereby enabling the agent to dream ahead and reason about how actions will affect subsequent observations. Experiments on R2R-CE, RxR-CE, and R2R-PE benchmarks achieve new SOTA results, and real-robot tests across diverse environments demonstrate LatentPilot's superior understanding of environment-action dynamics in scene. Project page:https://abdd.top/latentpilot/
Paper AI Chat
この論文のPDF全文を対象にAIに質問できます。
質問の例: