菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-07
📄 Abstract - An Actor-Critic Framework for Continuous-Time Jump-Diffusion Controls with Normalizing Flows

Continuous-time stochastic control with time-inhomogeneous jump-diffusion dynamics is central in finance and economics, but computing optimal policies is difficult under explicit time dependence, discontinuous shocks, and high dimensionality. We propose an actor-critic framework that serves as a mesh-free solver for entropy-regularized control problems and stochastic games with jumps. The approach is built on a time-inhomogeneous little q-function and an appropriate occupation measure, yielding a policy-gradient representation that accommodates time-dependent drift, volatility, and jump terms. To represent expressive stochastic policies in continuous-action spaces, we parameterize the actor using conditional normalizing flows, enabling flexible non-Gaussian policies while retaining exact likelihood evaluation for entropy regularization and policy optimization. We validate the method on time-inhomogeneous linear-quadratic control, Merton portfolio optimization, and a multi-agent portfolio game, using explicit solutions or high-accuracy benchmarks. Numerical results demonstrate stable learning under jump discontinuities, accurate approximation of optimal stochastic policies, and favorable scaling with respect to dimension and number of agents.

顶级标签: reinforcement learning financial agents
详细标签: actor-critic continuous-time control jump-diffusion normalizing flows stochastic games 或 搜索:

一种基于标准化流的连续时间跳跃扩散控制执行者-评判者框架 / An Actor-Critic Framework for Continuous-Time Jump-Diffusion Controls with Normalizing Flows


1️⃣ 一句话总结

本文提出了一种新的AI算法框架,能够有效解决金融经济领域中包含随机跳跃和复杂时间变化的高维决策难题,其核心是使用一种灵活的神经网络来学习并优化随机策略,并在投资组合管理等实际问题上验证了其准确性和高效性。

源自 arXiv: 2604.05398