菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-21
📄 Abstract - Learning Lifted Action Models from Unsupervised Visual Traces

Efficient construction of models capturing the preconditions and effects of actions is essential for applying AI planning in real-world domains. Extensive prior work has explored learning such models from high-level descriptions of state and/or action sequences. In this paper, we tackle a more challenging setting: learning lifted action models from sequences of state images, without action observation. We propose a deep learning framework that jointly learns state prediction, action prediction, and a lifted action model. We also introduce a mixed-integer linear program (MILP) to prevent prediction collapse and self-reinforcing errors among predictions. The MILP takes the predicted states, actions, and action model over a subset of traces and solves for logically consistent states, actions, and action model that are as close as possible to the original predictions. Pseudo-labels extracted from the MILP solution are then used to guide further training. Experiments across multiple domains show that integrating MILP-based correction helps the model escape local optima and converge toward globally consistent solutions.

顶级标签: machine learning reinforcement learning computer vision
详细标签: action model learning visual traces unsupervised mixed-integer linear program state prediction 或 搜索:

从无监督视觉轨迹中学习提升动作模型 / Learning Lifted Action Models from Unsupervised Visual Traces


1️⃣ 一句话总结

本研究提出一种深度学习框架,能够仅从图像序列(不依赖动作标签)中自动学习动作的前提条件和效果,并通过混合整数线性规划来纠正预测错误,从而构建逻辑一致的动作模型。

源自 arXiv: 2604.19043