菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-01-10
📄 Abstract - Are LLMs Vulnerable to Preference-Undermining Attacks (PUA)? A Factorial Analysis Methodology for Diagnosing the Trade-off between Preference Alignment and Real-World Validity

Large Language Model (LLM) training often optimizes for preference alignment, rewarding outputs that are perceived as helpful and interaction-friendly. However, this preference-oriented objective can be exploited: manipulative prompts can steer responses toward user-appeasing agreement and away from truth-oriented correction. In this work, we investigate whether aligned models are vulnerable to Preference-Undermining Attacks (PUA), a class of manipulative prompting strategies designed to exploit the model's desire to please user preferences at the expense of truthfulness. We propose a diagnostic methodology that provides a finer-grained and more directive analysis than aggregate benchmark scores, using a factorial evaluation framework to decompose prompt-induced shifts into interpretable effects of system objectives (truth- vs. preference-oriented) and PUA-style dialogue factors (directive control, personal derogation, conditional approval, reality denial) within a controlled $2 \times 2^4$ design. Surprisingly, more advanced models are sometimes more susceptible to manipulative prompts. Beyond the dominant reality-denial factor, we observe model-specific sign reversals and interactions with PUA-style factors, suggesting tailored defenses rather than uniform robustness. These findings offer a novel, reproducible factorial evaluation methodology that provides finer-grained diagnostics for post-training processes like RLHF, enabling better trade-offs in the product iteration of LLMs by offering a more nuanced understanding of preference alignment risks and the impact of manipulative prompts.

顶级标签: llm model evaluation natural language processing
详细标签: preference alignment adversarial attacks factorial analysis robustness truthfulness 或 搜索:

大语言模型是否易受偏好破坏攻击?一种用于诊断偏好对齐与现实有效性权衡的因子分析方法 / Are LLMs Vulnerable to Preference-Undermining Attacks (PUA)? A Factorial Analysis Methodology for Diagnosing the Trade-off between Preference Alignment and Real-World Validity


1️⃣ 一句话总结

这篇论文发现,为了让大语言模型显得更“乐于助人”而进行的偏好对齐训练,反而可能让它们更容易被一种叫做‘偏好破坏攻击’的诱导性提问所操控,从而为了讨好用户而牺牲事实准确性;为此,作者提出了一种新的因子分析方法,可以更精细地诊断这种风险,帮助开发者在模型迭代中做出更好的权衡。

源自 arXiv: 2601.06596