Multimodal Learning 相关度: 9/10

Physics-based phenomenological characterization of cross-modal bias in multimodal models

Hyeongmo Kim, Sohyun Kang, Yerin Choi, Seungyeon Ji, Junhyuk Woo, Hyunsuk Chung, Soyeon Caren Han, Kyungreem Han
arXiv: 2602.20624v1 发布: 2026-02-24 更新: 2026-02-24

AI 摘要

该论文提出了一种基于物理现象的解释性方法,用于分析多模态LLM中的跨模态偏差和公平性问题。

主要贡献

  • 提出了基于物理现象的解释性方法来分析MLLM偏差
  • 使用物理代理模型描述Transformer动态,分析跨模态偏差
  • 通过实验证明多模态输入可能加剧模态主导性

方法论

构建基于物理的代理模型描述Transformer动态,进行多输入诊断实验,包括扰动分析和动态分析。

原文摘要

The term 'algorithmic fairness' is used to evaluate whether AI models operate fairly in both comparative (where fairness is understood as formal equality, such as "treat like cases as like") and non-comparative (where unfairness arises from the model's inaccuracy, arbitrariness, or inscrutability) contexts. Recent advances in multimodal large language models (MLLMs) are breaking new ground in multimodal understanding, reasoning, and generation; however, we argue that inconspicuous distortions arising from complex multimodal interaction dynamics can lead to systematic bias. The purpose of this position paper is twofold: first, it is intended to acquaint AI researchers with phenomenological explainable approaches that rely on the physical entities that the machine experiences during training/inference, as opposed to the traditional cognitivist symbolic account or metaphysical approaches; second, it is to state that this phenomenological doctrine will be practically useful for tackling algorithmic fairness issues in MLLMs. We develop a surrogate physics-based model that describes transformer dynamics (i.e., semantic network structure and self-/cross-attention) to analyze the dynamics of cross-modal bias in MLLM, which are not fully captured by conventional embedding- or representation-level analyses. We support this position through multi-input diagnostic experiments: 1) perturbation-based analyses of emotion classification using Qwen2.5-Omni and Gemma 3n, and 2) dynamical analysis of Lorenz chaotic time-series prediction through the physical surrogate. Across two architecturally distinct MLLMs, we show that multimodal inputs can reinforce modality dominance rather than mitigate it, as revealed by structured error-attractor patterns under systematic label perturbation, complemented by dynamical analysis.

标签

多模态学习 公平性 物理信息 Transformer 偏差分析

arXiv 分类

cs.AI cond-mat.stat-mech