菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-07
📄 Abstract - QiMeng-PRepair: Precise Code Repair via Edit-Aware Reward Optimization

Large Language Models (LLMs) achieve strong program repair performance but often suffer from over-editing, where excessive modifications overwrite correct code and hinder bug localization. We systematically quantify its impact and introduce precise repair task, which maximizes reuse of correct code while fixing only buggy parts. Building on this insight, we propose PRepair, a framework that mitigates over-editing and improves repair accuracy. PRepair has two components: Self-Breaking, which generates diverse buggy programs via controlled bug injection and min-max sampling, and Self-Repairing, which trains models with Edit-Aware Group Relative Policy Optimization (EA-GRPO) using an edit-aware reward to encourage minimal yet correct edits. Experiments show that PRepair improves repair precision by up to 31.4% under $\mathrm{fix}_1@1$, a metric that jointly considers repair correctness and extent, and significantly increases decoding throughput when combined with speculative editing, demonstrating its potential for precise and practical code repair.

顶级标签: llm model training systems
详细标签: code repair program synthesis reward optimization policy optimization over-editing 或 搜索:

QiMeng-PRepair:通过编辑感知奖励优化实现精确代码修复 / QiMeng-PRepair: Precise Code Repair via Edit-Aware Reward Optimization


1️⃣ 一句话总结

这篇论文提出了一个名为PRepair的新框架,通过让大语言模型学会‘只修改错误代码、尽量保留正确代码’的方式,来解决现有代码修复工具过度修改的问题,从而更精确、高效地修复程序错误。

源自 arXiv: 2604.05963