菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-09
📄 Abstract - Deterministic Differentiable Structured Pruning for Large Language Models

Structured pruning reduces LLM inference cost by removing low-importance architectural components. This can be viewed as learning a multiplicative gate for each component under an l0 sparsity constraint. Due to the discreteness of the l0 norm, prior work typically adopts stochastic hard-concrete relaxations to enable differentiable optimization; however, this stochasticity can introduce a train--test mismatch when sampled masks are discretized for deployment and restricts masks to a bounded, near-binary range. To address this, we propose Deterministic Differentiable Pruning (DDP), a mask-only optimization method that eliminates stochasticity by directly optimizing a deterministic soft surrogate of the discrete l0 objective. Compared with prior approaches, DDP offers greater expressiveness, reduced train--test mismatch, and faster convergence. We apply our method to several dense and MoE models, including Qwen3-32B and Qwen3-30B-A3B, achieving a performance loss as small as 1% on downstream tasks while outperforming previous methods at 20% sparsity. We further demonstrate end-to-end inference speedups in realistic deployment settings with vLLM.

顶级标签: llm model training systems
详细标签: structured pruning model compression differentiable optimization inference acceleration sparsity 或 搜索:

面向大语言模型的确定性可微分结构化剪枝 / Deterministic Differentiable Structured Pruning for Large Language Models


1️⃣ 一句话总结

这篇论文提出了一种名为DDP的确定性可微分剪枝方法,它能更高效、更准确地压缩大语言模型的规模,在保证模型性能损失很小的同时,显著提升推理速度。

源自 arXiv: 2603.08065