菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-14
📄 Abstract - Meet Dynamic Individual Preferences: Resolving Conflicting Human Value with Paired Fine-Tuning

Recent advances in large language models (LLMs) have significantly improved the alignment of models with general human preferences. However, a major challenge remains in adapting LLMs to individual preferences, which are not only diverse but also dynamic. In this paper, we introduce a novel framework, Preference-Paired Fine-Tuning (PFT), designed to align models with contradictory and evolving individual preferences. We present a new dataset, Value Conflict Dilemma (VCD), which includes scenarios that involve conflicting human preferences, facilitating the evaluation of our approach. Our experiments demonstrate that PFT outperforms single-preference training methods, achieving up to 96.6% accuracy in multi-choice classification tasks and the highest open-ended generation score of 8.69. PFT also shows significant improvements over DPO, SFT and some traditional training methods, especially when handling conflicting preferences. Additionally, with limited user history data, models can inferring preference vector rapidly, achieving a 44.76% improvement in user-specific preference alignment in comparison to single-preference models.

顶级标签: llm model training agents
详细标签: preference alignment fine-tuning value conflict personalization dynamic preferences 或 搜索:

满足动态个人偏好:通过配对微调解决冲突的人类价值观 / Meet Dynamic Individual Preferences: Resolving Conflicting Human Value with Paired Fine-Tuning


1️⃣ 一句话总结

这篇论文提出了一个名为‘偏好配对微调’的新方法,让大型语言模型不仅能理解大众的普遍偏好,还能学习和适应每个用户独特且可能相互矛盾、不断变化的个人偏好,从而提供更个性化的服务。

源自 arXiv: 2604.12479