Utonia: Toward One Encoder for All Point Clouds
AI 摘要
提出Utonia,一个统一的自监督点云Transformer编码器,适用于多个领域。
主要贡献
- 提出一个统一的跨域点云编码器Utonia
- 证明了Utonia在不同领域之间的迁移能力
- 展示了Utonia在具身智能和多模态推理中的应用潜力
方法论
使用自监督学习训练一个Transformer编码器,使其能够处理来自不同领域的点云数据,并学习通用的表示空间。
原文摘要
We dream of a future where point clouds from all domains can come together to shape a single model that benefits them all. Toward this goal, we present Utonia, a first step toward training a single self-supervised point transformer encoder across diverse domains, spanning remote sensing, outdoor LiDAR, indoor RGB-D sequences, object-centric CAD models, and point clouds lifted from RGB-only videos. Despite their distinct sensing geometries, densities, and priors, Utonia learns a consistent representation space that transfers across domains. This unification improves perception capability while revealing intriguing emergent behaviors that arise only when domains are trained jointly. Beyond perception, we observe that Utonia representations can also benefit embodied and multimodal reasoning: conditioning vision-language-action policies on Utonia features improves robotic manipulation, and integrating them into vision-language models yields gains on spatial reasoning. We hope Utonia can serve as a step toward foundation models for sparse 3D data, and support downstream applications in AR/VR, robotics, and autonomous driving.