Multimodal Learning 相关度: 9/10

SToRM: Supervised Token Reduction for Multi-modal LLMs toward efficient end-to-end autonomous driving

Seo Hyun Kim, Jin Bok Park, Do Yeon Koo, Ho Gun Park, Il Yong Chun
arXiv: 2602.11656v1 发布: 2026-02-12 更新: 2026-02-12

AI 摘要

SToRM通过监督式Token缩减,在保证性能的同时显著降低多模态LLM在自动驾驶中的计算成本。

主要贡献

  • 提出Supervised Token Reduction框架SToRM
  • 设计轻量级的重要性预测器
  • 引入anchor-context merging模块

方法论

使用重要性预测器估计token重要性,通过监督学习获取伪标签,利用anchor-context合并减少冗余。

原文摘要

In autonomous driving, end-to-end (E2E) driving systems that predict control commands directly from sensor data have achieved significant advancements. For safe driving in unexpected scenarios, these systems may additionally rely on human interventions such as natural language instructions. Using a multi-modal large language model (MLLM) facilitates human-vehicle interaction and can improve performance in such scenarios. However, this approach requires substantial computational resources due to its reliance on an LLM and numerous visual tokens from sensor inputs, which are limited in autonomous vehicles. Many MLLM studies have explored reducing visual tokens, but often suffer end-task performance degradation compared to using all tokens. To enable efficient E2E driving while maintaining performance comparable to using all tokens, this paper proposes the first Supervised Token Reduction framework for multi-modal LLMs (SToRM). The proposed framework consists of three key elements. First, a lightweight importance predictor with short-term sliding windows estimates token importance scores. Second, a supervised training approach uses an auxiliary path to obtain pseudo-supervision signals from an all-token LLM pass. Third, an anchor-context merging module partitions tokens into anchors and context tokens, and merges context tokens into relevant anchors to reduce redundancy while minimizing information loss. Experiments on the LangAuto benchmark show that SToRM outperforms state-of-the-art E2E driving MLLMs under the same reduced-token budget, maintaining all-token performance while reducing computational cost by up to 30x.

标签

自动驾驶 多模态LLM Token缩减 计算效率

arXiv 分类

cs.CV cs.AI cs.RO