Multimodal Learning 相关度: 9/10

Feeling the Space: Egomotion-Aware Video Representation for Efficient and Accurate 3D Scene Understanding

Shuyao Shi, Kang G. Shin
arXiv: 2603.17980v1 发布: 2026-03-18 更新: 2026-03-18

AI 摘要

Motion-MLLM利用运动数据增强MLLM,提升3D场景理解的效率和准确性。

主要贡献

  • 提出Motion-MLLM框架,融合运动数据和视觉信息
  • 设计级联运动-视觉关键帧过滤模块
  • 构建非对称跨模态融合模块

方法论

利用IMU数据和视觉特征选择关键帧,通过运动tokens连接视觉和运动信息,实现高效的跨模态融合。

原文摘要

Recent Multimodal Large Language Models (MLLMs) have shown high potential for spatial reasoning within 3D scenes. However, they typically rely on computationally expensive 3D representations like point clouds or reconstructed Bird's-Eye View (BEV) maps, or lack physical grounding to resolve ambiguities in scale and size. This paper significantly enhances MLLMs with egomotion modality data, captured by Inertial Measurement Units (IMUs) concurrently with the video. In particular, we propose a novel framework, called Motion-MLLM, introducing two key components: (1) a cascaded motion-visual keyframe filtering module that leverages both IMU data and visual features to efficiently select a sparse yet representative set of keyframes, and (2) an asymmetric cross-modal fusion module where motion tokens serve as intermediaries that channel egomotion cues and cross-frame visual context into the visual representation. By grounding visual content in physical egomotion trajectories, Motion-MLLM can reason about absolute scale and spatial relationships across the scene. Our extensive evaluation shows that Motion-MLLM makes significant improvements in various tasks related to 3D scene understanding and spatial reasoning. Compared to state-of-the-art (SOTA) methods based on video frames and explicit 3D data, Motion-MLLM exhibits similar or even higher accuracy with significantly less overhead (i.e., 1.40$\times$ and 1.63$\times$ higher cost-effectiveness, respectively).

标签

Multimodal Learning 3D Scene Understanding Egomotion MLLM

arXiv 分类

cs.CV