Human-Centred LLM Privacy Audits: Findings and Frictions
AI 摘要
研究LLM隐私审计,发现LLM会泄露个人信息,并提出改进隐私审计的建议。
主要贡献
- 提出了LMP2隐私审计工具
- 揭示了LLM隐私评估的挑战和摩擦
- 针对以人为本的LLM隐私审计提出了建议
方法论
通过用户研究(N=458)和对不同LLM的评估,分析LLM对个人信息的预测准确性和用户对隐私的感知。
原文摘要
Large language models (LLMs) learn statistical associations from massive training corpora and user interactions, and deployed systems can surface or infer information about individuals. Yet people lack practical ways to inspect what a model associates with their name. We report interim findings from an ongoing study and introduce LMP2, a browser-based self-audit tool. In two user studies ($N_{total}{=}458$), GPT-4o predicts 11 of 50 features for everyday people with $\ge$60\% accuracy, and participants report wanting control over LLM-generated associations despite not considering all outputs privacy violations. To validate our probing method, we evaluate eight LLMs on public figures and non-existent names, observing clear separation between stable name-conditioned associations and model defaults. Our findings also contribute to exposing a broader generative AI evaluation crisis: when outputs are probabilistic, context-dependent, and user-mediated through elicitation, what model--individual associations even include is under-specified and operationalisation relies on crafting probes and metrics that are hard to validate or compare. To move towards reliable, actionable human-centred LLM privacy audits, we identify nine frictions that emerged in our study and offer recommendations for future work and the design of human-centred LLM privacy audits.