菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-11
📄 Abstract - MERIT Feedback Elicits Better Bargaining in LLM Negotiators

Bargaining is often regarded as a logical arena rather than an art or a matter of intuition, yet Large Language Models (LLMs) still struggle to navigate it due to limited strategic depth and difficulty adapting to complex human factors. Current benchmarks rarely capture this limitation. To bridge this gap, we present an utility feedback centric framework. Our contributions are: (i) AgoraBench, a new benchmark spanning nine challenging settings (e.g., deception, monopoly) that supports diverse strategy modeling; (ii) human-aligned, economically grounded metrics derived from utility theory. This is operationalized via agent utility, negotiation power, and acquisition ratio that implicitly measure how well the negotiation aligns with human preference and (iii) a human preference grounded dataset with learning pipeline that strengthens LLMs' bargaining ability through both prompting and finetuning. Empirical results indicate that baseline LLM strategies often diverge from human preferences, while our mechanism substantially improves negotiation performance, yielding deeper strategic behavior and stronger opponent awareness.

顶级标签: llm agents benchmark
详细标签: negotiation utility feedback human preference alignment strategic reasoning agent evaluation 或 搜索:

基于效用反馈的机制提升大语言模型谈判者的议价能力 / MERIT Feedback Elicits Better Bargaining in LLM Negotiators


1️⃣ 一句话总结

这篇论文提出了一个结合新基准、经济指标和人类偏好数据集的框架,通过提供基于效用的反馈,有效提升了大语言模型在复杂谈判中的策略深度和对手适应性,使其行为更贴近人类偏好。

源自 arXiv: 2602.10467