菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-29
📄 Abstract - RTLSeek: Boosting the LLM-Based RTL Generation with Multi-Stage Diversity-Oriented Reinforcement Learning

Register Transfer Level (RTL) design translates high-level specifications into hardware using HDLs such as Verilog. Although LLM-based RTL generation is promising, the scarcity of functionally verifiable high-quality data limits both accuracy and diversity. Existing post-training typically produces a single HDL implementation per specification, lacking awareness of RTL variations needed for different design goals. We propose RTLSeek, a post-training paradigm that applies rule-based Diversity-Oriented Reinforcement Learning to improve RTL correctness and diversity. Our Diversity-Centric Multi-Objective Reward Scheduling integrates expert knowledge with EDA feedback, and a three-stage framework maximizes the utility of limited data. Experiments on the RTLLM benchmark show that RTLSeek surpasses prior methods, with ablation results confirming that encouraging broader design-space exploration improves RTL quality and achieves the principle of "the more generated, the better results." Implementation framework, including the dataset, source code, and model weights, is shown at this https URL.

顶级标签: llm model training systems
详细标签: reinforcement learning hardware design rtl generation diversity verilog 或 搜索:

RTLSeek:利用多阶段多样性导向强化学习提升基于大语言模型的RTL生成 / RTLSeek: Boosting the LLM-Based RTL Generation with Multi-Stage Diversity-Oriented Reinforcement Learning


1️⃣ 一句话总结

这篇论文提出了一个名为RTLSeek的新方法,它通过一种鼓励生成多种不同硬件设计方案的多阶段强化学习训练策略,有效解决了当前AI生成硬件设计代码时质量不高、方案单一的问题,从而显著提升了生成结果的正确性和实用性。

源自 arXiv: 2603.27630