脉冲层自适应幅度剪枝 / Spiking Layer-Adaptive Magnitude-based Pruning
1️⃣ 一句话总结
这篇论文提出了一种名为SLAMP的新方法,专门用于高效地压缩脉冲神经网络,它通过考虑脉冲信号在时间上的累积效应和不同网络层的重要性,智能地裁剪掉不重要的连接,从而在保持模型准确性的同时,大幅降低了计算和能耗开销,使其更适合在实际设备上部署。
Spiking Neural Networks (SNNs) provide energy-efficient computation but their deployment is constrained by dense connectivity and high spiking operation costs. Existing magnitude-based pruning strategies, when naively applied to SNNs, fail to account for temporal accumulation, non-uniform timestep contributions, and membrane stability, often leading to severe performance degradation. This paper proposes Spiking Layer-Adaptive Magnitude-based Pruning (SLAMP), a theory-guided pruning framework that generalizes layer-adaptive magnitude pruning to temporal SNNs by explicitly controlling worst-case output distortion across layers and timesteps. SLAMP formulates sparsity allocation as a temporal distortion-constrained optimization problem, yielding time-aware layer importance scores that reduce to conventional layer-adaptive pruning in single-timestep limit. An efficient two-stage procedure is derived, combining temporal score estimation, global sparsity allocation, and magnitude pruning with retraining for stability recovery. Experiments on CIFAR10, CIFAR100, and the event-based CIFAR10-DVS datasets demonstrate that SLAMP achieves substantial connectivity and spiking operation reductions while preserving accuracy, enabling efficient and deployable SNN inference.
脉冲层自适应幅度剪枝 / Spiking Layer-Adaptive Magnitude-based Pruning
这篇论文提出了一种名为SLAMP的新方法,专门用于高效地压缩脉冲神经网络,它通过考虑脉冲信号在时间上的累积效应和不同网络层的重要性,智能地裁剪掉不重要的连接,从而在保持模型准确性的同时,大幅降低了计算和能耗开销,使其更适合在实际设备上部署。
源自 arXiv: 2603.14946