菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-16
📄 Abstract - Spiking Layer-Adaptive Magnitude-based Pruning

Spiking Neural Networks (SNNs) provide energy-efficient computation but their deployment is constrained by dense connectivity and high spiking operation costs. Existing magnitude-based pruning strategies, when naively applied to SNNs, fail to account for temporal accumulation, non-uniform timestep contributions, and membrane stability, often leading to severe performance degradation. This paper proposes Spiking Layer-Adaptive Magnitude-based Pruning (SLAMP), a theory-guided pruning framework that generalizes layer-adaptive magnitude pruning to temporal SNNs by explicitly controlling worst-case output distortion across layers and timesteps. SLAMP formulates sparsity allocation as a temporal distortion-constrained optimization problem, yielding time-aware layer importance scores that reduce to conventional layer-adaptive pruning in single-timestep limit. An efficient two-stage procedure is derived, combining temporal score estimation, global sparsity allocation, and magnitude pruning with retraining for stability recovery. Experiments on CIFAR10, CIFAR100, and the event-based CIFAR10-DVS datasets demonstrate that SLAMP achieves substantial connectivity and spiking operation reductions while preserving accuracy, enabling efficient and deployable SNN inference.

顶级标签: model training systems theory
详细标签: spiking neural networks model pruning energy efficiency temporal optimization layer-adaptive pruning 或 搜索:

脉冲层自适应幅度剪枝 / Spiking Layer-Adaptive Magnitude-based Pruning


1️⃣ 一句话总结

这篇论文提出了一种名为SLAMP的新方法,专门用于高效地压缩脉冲神经网络,它通过考虑脉冲信号在时间上的累积效应和不同网络层的重要性,智能地裁剪掉不重要的连接,从而在保持模型准确性的同时,大幅降低了计算和能耗开销,使其更适合在实际设备上部署。

源自 arXiv: 2603.14946