Multimodal Learning 相关度: 9/10

CARE Drive A Framework for Evaluating Reason-Responsiveness of Vision Language Models in Automated Driving

Lucas Elbert Suryana, Farah Bierenga, Sanne van Buuren, Pepijn Kooij, Elsefien Tulleners, Federico Scari, Simeon Calvert, Bart van Arem, Arkady Zgonnikov
arXiv: 2602.15645v1 发布: 2026-02-17 更新: 2026-02-17

AI 摘要

提出CARE Drive框架,评估自动驾驶视觉语言模型对人类理由的响应性,提高决策可解释性。

主要贡献

  • 提出CARE Drive框架,评估视觉语言模型在自动驾驶中的理由响应性
  • 通过上下文扰动测量决策对人类理由的敏感度
  • 验证了人类理由对模型决策的影响,并发现了模型对不同理由的敏感度差异

方法论

CARE Drive通过prompt校准和上下文扰动,比较基线和增强模型决策,评估模型对安全、社会压力和效率等理由的响应。

原文摘要

Foundation models, including vision language models, are increasingly used in automated driving to interpret scenes, recommend actions, and generate natural language explanations. However, existing evaluation methods primarily assess outcome based performance, such as safety and trajectory accuracy, without determining whether model decisions reflect human relevant considerations. As a result, it remains unclear whether explanations produced by such models correspond to genuine reason responsive decision making or merely post hoc rationalizations. This limitation is especially significant in safety critical domains because it can create false confidence. To address this gap, we propose CARE Drive, Context Aware Reasons Evaluation for Driving, a model agnostic framework for evaluating reason responsiveness in vision language models applied to automated driving. CARE Drive compares baseline and reason augmented model decisions under controlled contextual variation to assess whether human reasons causally influence decision behavior. The framework employs a two stage evaluation process. Prompt calibration ensures stable outputs. Systematic contextual perturbation then measures decision sensitivity to human reasons such as safety margins, social pressure, and efficiency constraints. We demonstrate CARE Drive in a cyclist overtaking scenario involving competing normative considerations. Results show that explicit human reasons significantly influence model decisions, improving alignment with expert recommended behavior. However, responsiveness varies across contextual factors, indicating uneven sensitivity to different types of reasons. These findings provide empirical evidence that reason responsiveness in foundation models can be systematically evaluated without modifying model parameters.

标签

自动驾驶 视觉语言模型 理由响应性 可解释性

arXiv 分类

cs.AI cs.CV