AIDB Daily Papers
Orion-Lite:LLMの推論能力を効率的な視覚のみの運転モデルへ蒸留
※ 日本語タイトル・ポイントはAIによる自動生成です。正確な内容は原論文をご確認ください。
ポイント
- 大規模言語モデル(LLM)の知識を自動運転システムに活用し、複雑な状況への対応力向上を目指した研究。
- LLMを軽量な運転モデルに蒸留することで、高い推論能力を維持しつつ、計算コストを抑える新しいアプローチ。
- 視覚情報のみを用いるOrion-Liteは、教師モデルであるORIONを上回る性能を示し、Bench2Driveで最高スコアを達成。
Abstract
Leveraging the general world knowledge of Large Language Models (LLMs) holds significant promise for improving the ability of autonomous driving systems to handle rare and complex scenarios. While integrating LLMs into Vision-Language-Action (VLA) models has yielded state-of-the-art performance, their massive parameter counts pose severe challenges for latency-sensitive and energy-efficient deployment. Distilling LLM knowledge into a compact driving model offers a compelling solution to retain these reasoning capabilities while maintaining a manageable computational footprint. Although previous works have demonstrated the efficacy of distillation, these efforts have primarily focused on relatively simple scenarios and open-loop evaluations. Therefore, in this work, we investigate LLM distillation in more complex, interactive scenarios under closed-loop evaluation. We demonstrate that through a combination of latent feature distillation and ground-truth trajectory supervision, an efficient vision-only student model textbf{Orion-Lite} can even surpass the performance of its massive VLA teacher, ORION. Setting a new state-of-the-art on the rigorous Bench2Drive benchmark, with a Driving Score of 80.6. Ultimately, this reveals that vision-only architectures still possess significant, untapped potential for high-performance reactive planning.
Paper AI Chat
この論文のPDF全文を対象にAIに質問できます。
質問の例: