作为ASP程序员的LLM:自我修正实现任务无关的非单调推理 / LLMs as ASP Programmers: Self-Correction Enables Task-Agnostic Nonmonotonic Reasoning
1️⃣ 一句话总结
本文提出了一种名为“LLM+ASP”的框架,让大型语言模型自动将自然语言问题转换为回答集程序(一种能处理默认规则和例外的非单调逻辑),并通过求解器的结构化反馈进行自我修正,从而在不依赖人工定制知识的情况下,在多种推理任务上显著优于传统方法。
Recent large language models (LLMs) have achieved impressive reasoning milestones but continue to struggle with high computational costs, logical inconsistencies, and sharp performance degradation on high-complexity problems. While neuro-symbolic methods attempt to mitigate these issues by coupling LLMs with symbolic reasoners, existing approaches typically rely on monotonic logics (e.g., SMT) that cannot represent defeasible reasoning -- essential components of human cognition. We present "LLM+ASP," a framework that translates natural language into Answer Set Programming (ASP), a nonmonotonic formalism based on stable model semantics. Unlike prior "LLM+ASP" approaches that require manually authored knowledge modules, domain-specific prompts, or evaluation restricted to single problem classes, our framework operates without any per-task engineering and applies uniformly across diverse reasoning tasks. Our system utilizes an automated self-correction loop where structured feedback from the ASP solver enables iterative refinement. Evaluating across six diverse benchmarks, we demonstrate that: (1) stable model semantics allow LLMs to naturally express default rules and exceptions, outperforming SMT-based alternatives by significant margins on nonmonotonic tasks; (2) iterative self-correction is the primary driver of performance, effectively replacing the need for handcrafted domain knowledge; (3) compact in-context reference guides substantially outperform verbose documentation, revealing a "context rot" phenomenon where excessive context hinders constraint adherence.
作为ASP程序员的LLM:自我修正实现任务无关的非单调推理 / LLMs as ASP Programmers: Self-Correction Enables Task-Agnostic Nonmonotonic Reasoning
本文提出了一种名为“LLM+ASP”的框架,让大型语言模型自动将自然语言问题转换为回答集程序(一种能处理默认规则和例外的非单调逻辑),并通过求解器的结构化反馈进行自我修正,从而在不依赖人工定制知识的情况下,在多种推理任务上显著优于传统方法。
源自 arXiv: 2604.27960