SeLaR:大语言模型中的选择性潜在推理 / SeLaR: Selective Latent Reasoning in Large Language Models
1️⃣ 一句话总结
这篇论文提出了一种名为SeLaR的轻量级方法,通过智能地在大语言模型推理过程中切换使用确定性词和软性概率向量,解决了现有推理方法容易出错或探索不足的问题,从而在多个任务上提升了推理性能。
Chain-of-Thought (CoT) has become a cornerstone of reasoning in large language models, yet its effectiveness is constrained by the limited expressiveness of discrete token sampling. Recent latent reasoning approaches attempt to alleviate this limitation by replacing discrete tokens with soft embeddings (probability-weighted mixtures of token embeddings) or hidden states, but they commonly suffer from two issues: (1) global activation injects perturbations into high-confidence steps, impairing reasoning stability; and (2) soft embeddings quickly collapse toward the highest-probability token, limiting exploration of alternative trajectories. To address these challenges, we propose SeLaR (Selective Latent Reasoning), a lightweight and training-free framework. SeLaR introduces an entropy-gated mechanism that activates soft embeddings only at low-confidence steps, while preserving discrete decoding at high-confidence steps. Additionally, we propose an entropy-aware contrastive regularization that pushes soft embeddings away from the dominant (highest-probability) token's direction, encouraging sustained exploration of multiple latent reasoning paths. Experiments on five reasoning benchmarks demonstrate that SeLaR consistently outperforms standard CoT and state-of-the-art training-free methods.
SeLaR:大语言模型中的选择性潜在推理 / SeLaR: Selective Latent Reasoning in Large Language Models
这篇论文提出了一种名为SeLaR的轻量级方法,通过智能地在大语言模型推理过程中切换使用确定性词和软性概率向量,解决了现有推理方法容易出错或探索不足的问题,从而在多个任务上提升了推理性能。
源自 arXiv: 2604.08299