菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-25
📄 Abstract - Primary-Fine Decoupling for Action Generation in Robotic Imitation

Multi-modal distribution in robotic manipulation action sequences poses critical challenges for imitation learning. To this end, existing approaches often model the action space as either a discrete set of tokens or a continuous, latent-variable distribution. However, both approaches present trade-offs: some methods discretize actions into tokens and therefore lose fine-grained action variations, while others generate continuous actions in a single stage tend to produce unstable mode transitions. To address these limitations, we propose Primary-Fine Decoupling for Action Generation (PF-DAG), a two-stage framework that decouples coarse action consistency from fine-grained variations. First, we compress action chunks into a small set of discrete modes, enabling a lightweight policy to select consistent coarse modes and avoid mode bouncing. Second, a mode conditioned MeanFlow policy is learned to generate high-fidelity continuous actions. Theoretically, we prove PF-DAG's two-stage design achieves a strictly lower MSE bound than single-stage generative policies. Empirically, PF-DAG outperforms state-of-the-art baselines across 56 tasks from Adroit, DexArt, and MetaWorld benchmarks. It further generalizes to real-world tactile dexterous manipulation tasks. Our work demonstrates that explicit mode-level decoupling enables both robust multi-modal modeling and reactive closed-loop control for robotic manipulation.

顶级标签: robotics model training agents
详细标签: imitation learning action generation multi-modal distribution two-stage framework dexterous manipulation 或 搜索:

用于机器人模仿动作生成的主-细解耦方法 / Primary-Fine Decoupling for Action Generation in Robotic Imitation


1️⃣ 一句话总结

这篇论文提出了一种名为PF-DAG的两阶段新方法,它通过先将机器人动作分解为粗略模式和精细变化来生成动作,从而在模仿学习中更稳定、更准确地复现复杂的多模态操作任务,并在多个基准测试和真实任务中表现出色。

源自 arXiv: 2602.21684