基于情境-价值-行动架构的价值驱动大语言模型智能体 / Context-Value-Action Architecture for Value-Driven Large Language Model Agents
1️⃣ 一句话总结
这篇论文针对现有大语言模型智能体行为僵化、容易产生价值极化的问题,提出了一种新的情境-价值-行动架构,通过一个基于真实人类数据训练的价值验证器来模拟动态价值激活,从而显著提升了智能体行为的真实性、多样性和可解释性。
Large Language Models (LLMs) have shown promise in simulating human behavior, yet existing agents often exhibit behavioral rigidity, a flaw frequently masked by the self-referential bias of current "LLM-as-a-judge" evaluations. By evaluating against empirical ground truth, we reveal a counter-intuitive phenomenon: increasing the intensity of prompt-driven reasoning does not enhance fidelity but rather exacerbates value polarization, collapsing population diversity. To address this, we propose the Context-Value-Action (CVA) architecture, grounded in the Stimulus-Organism-Response (S-O-R) model and Schwartz's Theory of Basic Human Values. Unlike methods relying on self-verification, CVA decouples action generation from cognitive reasoning via a novel Value Verifier trained on authentic human data to explicitly model dynamic value activation. Experiments on CVABench, which comprises over 1.1 million real-world interaction traces, demonstrate that CVA significantly outperforms baselines. Our approach effectively mitigates polarization while offering superior behavioral fidelity and interpretability.
基于情境-价值-行动架构的价值驱动大语言模型智能体 / Context-Value-Action Architecture for Value-Driven Large Language Model Agents
这篇论文针对现有大语言模型智能体行为僵化、容易产生价值极化的问题,提出了一种新的情境-价值-行动架构,通过一个基于真实人类数据训练的价值验证器来模拟动态价值激活,从而显著提升了智能体行为的真实性、多样性和可解释性。
源自 arXiv: 2604.05939