📄
Abstract - MOMO: A framework for seamless physical, verbal, and graphical robot skill learning and adaptation
Industrial robot applications require increasingly flexible systems that non-expert users can easily adapt for varying tasks and environments. However, different adaptations benefit from different interaction modalities. We present an interactive framework that enables robot skill adaptation through three complementary modalities: kinesthetic touch for precise spatial corrections, natural language for high-level semantic modifications, and a graphical web interface for visualizing geometric relations and trajectories, inspecting and adjusting parameters, and editing via-points by drag-and-drop. The framework integrates five components: energy-based human-intention detection, a tool-based LLM architecture (where the LLM selects and parameterizes predefined functions rather than generating code) for safe natural language adaptation, Kernelized Movement Primitives (KMPs) for motion encoding, probabilistic Virtual Fixtures for guided demonstration recording, and ergodic control for surface finishing. We demonstrate that this tool-based LLM architecture generalizes skill adaptation from KMPs to ergodic control, enabling voice-commanded surface finishing. Validation on a 7-DoF torque-controlled robot at the Automatica 2025 trade fair demonstrates the practical applicability of our approach in industrial settings.
MOMO:一种实现机器人物理、语言和图形无缝技能学习与适应的框架 /
MOMO: A framework for seamless physical, verbal, and graphical robot skill learning and adaptation
1️⃣ 一句话总结
本文提出一个名为MOMO的交互框架,允许非专业用户通过三种方式(物理触碰、自然语言和图形界面)灵活调整工业机器人的行为,其中语言指令通过一个安全的大语言模型架构实现,仅调用预定义函数而非生成代码,从而在保证安全的同时拓展了机器人技能的应用范围。