SceneAssistant: A Visual Feedback Agent for Open-Vocabulary 3D Scene Generation
AI 摘要
SceneAssistant通过视觉反馈迭代优化,实现开放词汇的3D场景生成。
主要贡献
- 提出基于视觉反馈的3D场景生成框架
- 利用VLM进行空间推理和规划
- 实现开放词汇和高质量的3D场景生成
方法论
利用3D对象生成模型和VLM,通过预定义的原子操作和视觉反馈,迭代优化场景的空间布局和文本对齐。
原文摘要
Text-to-3D scene generation from natural language is highly desirable for digital content creation. However, existing methods are largely domain-restricted or reliant on predefined spatial relationships, limiting their capacity for unconstrained, open-vocabulary 3D scene synthesis. In this paper, we introduce SceneAssistant, a visual-feedback-driven agent designed for open-vocabulary 3D scene generation. Our framework leverages modern 3D object generation model along with the spatial reasoning and planning capabilities of Vision-Language Models (VLMs). To enable open-vocabulary scene composition, we provide the VLMs with a comprehensive set of atomic operations (e.g., Scale, Rotate, FocusOn). At each interaction step, the VLM receives rendered visual feedback and takes actions accordingly, iteratively refining the scene to achieve more coherent spatial arrangements and better alignment with the input text. Experimental results demonstrate that our method can generate diverse, open-vocabulary, and high-quality 3D scenes. Both qualitative analysis and quantitative human evaluations demonstrate the superiority of our approach over existing methods. Furthermore, our method allows users to instruct the agent to edit existing scenes based on natural language commands. Our code is available at https://github.com/ROUJINN/SceneAssistant