投资者偏好的无模型推断:一种基于相对熵逆强化学习的方法 / Model-Free Inference of Investor Preferences: A Relative Entropy IRL Approach
1️⃣ 一句话总结
本文提出了一种无需知道市场变化规律的新方法,通过观察投资者的实际行动来反向推导其背后的投资偏好和决策目标,并使用邻近点技术解决了实际数据不足的问题,最后还设计了一套统计检验来验证推断结果的可靠性。
We present a framework using Relative Entropy Inverse Reinforcement Learning (RE-IRL) to recover investor reward functions from observed investment actions and market conditions. Unlike traditional IRL algorithms, RE-IRL is employed to account for environments where transition probabilities are unknown or inaccessible. To address the challenge of data sparsity, we utilize a $K$-nearest neighbor approach to estimate the observed behavior policy. Furthermore, we propose a statistical testing framework to evaluate the validity and robustness of the estimated results.
投资者偏好的无模型推断:一种基于相对熵逆强化学习的方法 / Model-Free Inference of Investor Preferences: A Relative Entropy IRL Approach
本文提出了一种无需知道市场变化规律的新方法,通过观察投资者的实际行动来反向推导其背后的投资偏好和决策目标,并使用邻近点技术解决了实际数据不足的问题,最后还设计了一套统计检验来验证推断结果的可靠性。
源自 arXiv: 2604.24280