Multimodal Learning 相关度: 9/10

Ontology-Guided Diffusion for Zero-Shot Visual Sim2Real Transfer

Mohamed Youssef, Mayar Elfares, Anna-Maria Meer, Matteo Bortoletto, Andreas Bulling
arXiv: 2603.18719v1 发布: 2026-03-19 更新: 2026-03-19

AI 摘要

OGD利用知识图谱引导扩散模型,实现了零样本Sim2Real图像转换,提升了图像的真实感和可解释性。

主要贡献

  • 提出Ontology-Guided Diffusion (OGD) 框架
  • 使用知识图谱表示图像真实感
  • 利用符号规划器进行视觉编辑

方法论

OGD通过知识图谱编码图像特征,指导扩散模型生成更真实的图像,并结合符号规划器进行视觉编辑。

原文摘要

Bridging the simulation-to-reality (sim2real) gap remains challenging as labelled real-world data is scarce. Existing diffusion-based approaches rely on unstructured prompts or statistical alignment, which do not capture the structured factors that make images look real. We introduce Ontology- Guided Diffusion (OGD), a neuro-symbolic zero-shot sim2real image translation framework that represents realism as structured knowledge. OGD decomposes realism into an ontology of interpretable traits -- such as lighting and material properties -- and encodes their relationships in a knowledge graph. From a synthetic image, OGD infers trait activations and uses a graph neural network to produce a global embedding. In parallel, a symbolic planner uses the ontology traits to compute a consistent sequence of visual edits needed to narrow the realism gap. The graph embedding conditions a pretrained instruction-guided diffusion model via cross-attention, while the planned edits are converted into a structured instruction prompt. Across benchmarks, our graph-based embeddings better distinguish real from synthetic imagery than baselines, and OGD outperforms state-of-the-art diffusion methods in sim2real image translations. Overall, OGD shows that explicitly encoding realism structure enables interpretable, data-efficient, and generalisable zero-shot sim2real transfer.

标签

Sim2Real Diffusion Model Knowledge Graph Image Translation

arXiv 分类

cs.CV cs.AI