📄
Abstract - What Do LLMs Associate with Your Name? A Human-Centered Black-Box Audit of Personal Data
Large language models (LLMs), and conversational agents based on them, are exposed to personal data (PD) during pre-training and during user interactions. Prior work shows that PD can resurface, yet users lack insight into how strongly models associate specific information to their identity. We audit PD across eight LLMs (3 open-source; 5 API-based, including GPT-4o), introduce LMP2 (Language Model Privacy Probe), a human-centered, privacy-preserving audit tool refined through two formative studies (N=20), and run two studies with EU residents to capture (i) intuitions about LLM-generated PD (N1=155) and (ii) reactions to tool output (N2=303). We show empirically that models confidently generate multiple PD categories for well-known individuals. For everyday users, GPT-4o generates 11 features with 60% or more accuracy (e.g., gender, hair color, languages). Finally, 72% of participants sought control over model-generated associations with their name, raising questions about what counts as PD and whether data privacy rights should extend to LLMs.
大语言模型将什么信息与你的名字关联?一项以人为中心的个人数据黑盒审计 /
What Do LLMs Associate with Your Name? A Human-Centered Black-Box Audit of Personal Data
1️⃣ 一句话总结
这篇论文通过开发一个名为LMP2的隐私保护审计工具,实证研究发现大语言模型(如GPT-4o)能高准确度地从人名推断出多种个人特征(如性别、发色),并揭示了大多数用户希望对模型生成的此类关联进行控制,从而引发了对个人数据定义及隐私权是否应延伸至大语言模型的新讨论。