联邦学习中的二值神经网络:实现低成本推理 / Federated Learning of Binary Neural Networks: Enabling Low-Cost Inference
1️⃣ 一句话总结
这篇论文提出了一种名为FedBNN的新框架,它通过在联邦学习过程中直接训练二值化神经网络,大幅降低了模型在手机等边缘设备上的计算和内存开销,同时保持了与使用传统浮点数模型相近的准确率。
Federated Learning (FL) preserves privacy by distributing training across devices. However, using DNNs is computationally intensive at the low-powered edge during inference. Edge deployment demands models that simultaneously optimize memory footprint and computational efficiency, a dilemma where conventional DNNs fail by exceeding resource limits. Traditional post-training binarization reduces model size but suffers from severe accuracy loss due to quantization errors. To address these challenges, we propose FedBNN, a rotation-aware binary neural network framework that learns binary representations directly during local training. By encoding each weight as a single bit $\{+1, -1\}$ instead of a $32$-bit float, FedBNN shrinks the model footprint, significantly reducing runtime (during inference) FLOPs and memory requirements in comparison to federated methods using real models. Evaluations across multiple benchmark datasets demonstrate that FedBNN significantly reduces resource consumption while performing similarly to existing federated methods using real-valued models.
联邦学习中的二值神经网络:实现低成本推理 / Federated Learning of Binary Neural Networks: Enabling Low-Cost Inference
这篇论文提出了一种名为FedBNN的新框架,它通过在联邦学习过程中直接训练二值化神经网络,大幅降低了模型在手机等边缘设备上的计算和内存开销,同时保持了与使用传统浮点数模型相近的准确率。
源自 arXiv: 2603.15507