Fuz-RL:一种用于不确定性下安全强化学习的模糊引导鲁棒框架 / Fuz-RL: A Fuzzy-Guided Robust Framework for Safe Reinforcement Learning under Uncertainty
1️⃣ 一句话总结
这篇论文提出了一个名为Fuz-RL的新框架,它利用模糊数学理论来帮助强化学习智能体在充满不确定性的复杂环境中,既能做出鲁棒的决策,又能有效保障自身安全,从而在性能和安全性之间取得更好的平衡。
Safe Reinforcement Learning (RL) is crucial for achieving high performance while ensuring safety in real-world applications. However, the complex interplay of multiple uncertainty sources in real environments poses significant challenges for interpretable risk assessment and robust decision-making. To address these challenges, we propose Fuz-RL, a fuzzy measure-guided robust framework for safe RL. Specifically, our framework develops a novel fuzzy Bellman operator for estimating robust value functions using Choquet integrals. Theoretically, we prove that solving the Fuz-RL problem (in Constrained Markov Decision Process (CMDP) form) is equivalent to solving distributionally robust safe RL problems (in robust CMDP form), effectively avoiding min-max optimization. Empirical analyses on safe-control-gym and safety-gymnasium scenarios demonstrate that Fuz-RL effectively integrates with existing safe RL baselines in a model-free manner, significantly improving both safety and control performance under various types of uncertainties in observation, action, and dynamics.
Fuz-RL:一种用于不确定性下安全强化学习的模糊引导鲁棒框架 / Fuz-RL: A Fuzzy-Guided Robust Framework for Safe Reinforcement Learning under Uncertainty
这篇论文提出了一个名为Fuz-RL的新框架,它利用模糊数学理论来帮助强化学习智能体在充满不确定性的复杂环境中,既能做出鲁棒的决策,又能有效保障自身安全,从而在性能和安全性之间取得更好的平衡。
源自 arXiv: 2602.20729