ReGuLaR:基于渲染思维链引导的变分潜在推理 / ReGuLaR: Variational Latent Reasoning Guided by Rendered Chain-of-Thought
1️⃣ 一句话总结
这篇论文提出了一种名为ReGuLaR的新方法,它通过将思维链转化为图像来引导和压缩推理过程,从而让大语言模型在保持高准确率的同时,大幅减少了计算开销。
While Chain-of-Thought (CoT) significantly enhances the performance of Large Language Models (LLMs), explicit reasoning chains introduce substantial computational redundancy. Recent latent reasoning methods attempt to mitigate this by compressing reasoning processes into latent space, but often suffer from severe performance degradation due to the lack of appropriate compression guidance. In this study, we propose Rendered CoT-Guided variational Latent Reasoning (ReGuLaR), a simple yet novel latent learning paradigm resolving this issue. Fundamentally, we formulate latent reasoning within the Variational Auto-Encoding (VAE) framework, sampling the current latent reasoning state from the posterior distribution conditioned on previous ones. Specifically, when learning this variational latent reasoning model, we render explicit reasoning chains as images, from which we extract dense visual-semantic representations to regularize the posterior distribution, thereby achieving efficient compression with minimal information loss. Extensive experiments demonstrate that ReGuLaR significantly outperforms existing latent reasoning methods across both computational efficiency and reasoning effectiveness, and even surpasses CoT through multi-modal reasoning, providing a new and insightful solution to latent reasoning. Code: this https URL.
ReGuLaR:基于渲染思维链引导的变分潜在推理 / ReGuLaR: Variational Latent Reasoning Guided by Rendered Chain-of-Thought
这篇论文提出了一种名为ReGuLaR的新方法,它通过将思维链转化为图像来引导和压缩推理过程,从而让大语言模型在保持高准确率的同时,大幅减少了计算开销。
源自 arXiv: 2601.23184