LLM Reasoning 相关度: 8/10

They Said Memes Were Harmless-We Found the Ones That Hurt: Decoding Jokes, Symbols, and Cultural References

Sahil Tripathi, Gautam Siddharth Kashyap, Mehwish Nasim, Jian Yang, Jiechao Gao, Usman Naseem
arXiv: 2602.03822v1 发布: 2026-02-03 更新: 2026-02-03

AI 摘要

提出了CROSS-ALIGN+框架,提升基于meme的社交恶意信息检测效果,并增强模型可解释性。

主要贡献

  • 缓解文化盲区
  • 减少边界模糊
  • 增强可解释性

方法论

三阶段框架:通过知识库增强表示、LoRA适配器锐化边界、生成级联解释。

原文摘要

Meme-based social abuse detection is challenging because harmful intent often relies on implicit cultural symbolism and subtle cross-modal incongruence. Prior approaches, from fusion-based methods to in-context learning with Large Vision-Language Models (LVLMs), have made progress but remain limited by three factors: i) cultural blindness (missing symbolic context), ii) boundary ambiguity (satire vs. abuse confusion), and iii) lack of interpretability (opaque model reasoning). We introduce CROSS-ALIGN+, a three-stage framework that systematically addresses these limitations: (1) Stage I mitigates cultural blindness by enriching multimodal representations with structured knowledge from ConceptNet, Wikidata, and Hatebase; (2) Stage II reduces boundary ambiguity through parameter-efficient LoRA adapters that sharpen decision boundaries; and (3) Stage III enhances interpretability by generating cascaded explanations. Extensive experiments on five benchmarks and eight LVLMs demonstrate that CROSS-ALIGN+ consistently outperforms state-of-the-art methods, achieving up to 17% relative F1 improvement while providing interpretable justifications for each decision.

标签

Meme Social Abuse Detection Multimodal Learning Explainability

arXiv 分类

cs.CL