Multimodal Learning 相关度: 9/10

Unsafe2Safe: Controllable Image Anonymization for Downstream Utility

Mih Dinh, SouYoung Jin
arXiv: 2603.28605v1 发布: 2026-03-30 更新: 2026-03-30

AI 摘要

Unsafe2Safe提出了一种自动化的图像匿名化流程,保证隐私的同时维持图像效用。

主要贡献

  • 提出了一种全自动的图像匿名化pipeline
  • 设计了一个综合的匿名化质量评估标准
  • 通过微调扩散模型进一步提升隐私保护和语义保真度

方法论

使用视觉语言模型检测隐私风险,并利用大语言模型生成编辑指令,通过指令驱动的扩散编辑器实现匿名化。

原文摘要

Large-scale image datasets frequently contain identifiable or sensitive content, raising privacy risks when training models that may memorize and leak such information. We present Unsafe2Safe, a fully automated pipeline that detects privacy-prone images and rewrites only their sensitive regions using multimodally guided diffusion editing. Unsafe2Safe operates in two stages. Stage 1 uses a vision-language model to (i) inspect images for privacy risks, (ii) generate paired private and public captions that respectively include and omit sensitive attributes, and (iii) prompt a large language model to produce structured, identity-neutral edit instructions conditioned on the public caption. Stage 2 employs instruction-driven diffusion editors to apply these dual textual prompts, producing privacy-safe images that preserve global structure and task-relevant semantics while neutralizing private content. To measure anonymization quality, we introduce a unified evaluation suite covering Quality, Cheating, Privacy, and Utility dimensions. Across MS-COCO, Caltech101, and MIT Indoor67, Unsafe2Safe reduces face similarity, text similarity, and demographic predictability by large margins, while maintaining downstream model accuracy comparable to training on raw data. Fine-tuning diffusion editors on our automatically generated triplets (private caption, public caption, edit instruction) further improves both privacy protection and semantic fidelity. Unsafe2Safe provides a scalable, principled solution for constructing large, privacy-safe datasets without sacrificing visual consistency or downstream utility.

标签

图像匿名化 隐私保护 扩散模型 视觉语言模型

arXiv 分类

cs.CV cs.CY cs.LG