HarassGuard: Detecting Harassment Behaviors in Social Virtual Reality with Vision-Language Models
AI 摘要
HarassGuard利用视觉-语言模型检测社交VR中的骚扰行为,保护用户隐私。
主要贡献
- 构建了基于视觉的骚扰行为数据集
- 提出了基于VLM的骚扰行为检测系统HarassGuard
- 验证了HarassGuard在保护隐私前提下的有效性
方法论
使用视觉-语言模型,通过prompt工程和微调,结合VR环境上下文信息检测骚扰行为。
原文摘要
Social Virtual Reality (VR) platforms provide immersive social experiences but also expose users to serious risks of online harassment. Existing safety measures are largely reactive, while proactive solutions that detect harassment behavior during an incident often depend on sensitive biometric data, raising privacy concerns. In this paper, we present HarassGuard, a vision-language model (VLM) based system that detects physical harassment in social VR using only visual input. We construct an IRB-approved harassment vision dataset, apply prompt engineering, and fine-tune VLMs to detect harassment behavior by considering contextual information in social VR. Experimental results demonstrate that HarassGuard achieves competitive performance compared to state-of-the-art baselines (i.e., LSTM/CNN, Transformer), reaching an accuracy of up to 88.09% in binary classification and 68.85% in multi-class classification. Notably, HarassGuard matches these baselines while using significantly fewer fine-tuning samples (200 vs. 1,115), offering unique advantages in contextual reasoning and privacy-preserving detection.