基于联邦强化学习的不完全信息下高效移动群智感知 / Federated Reinforcement Learning for Efficient Mobile Crowdsensing under Incomplete Information
1️⃣ 一句话总结
本文针对移动用户在不了解全局信息时如何选择感知任务以赚取报酬的问题,提出了一种名为FDRL-PPO的分散式联邦深度强化学习算法,让用户通过本地学习和共享模型来优化策略,从而在保护隐私的同时提高任务完成率、公平性和能效。
Mobile crowdsensing (MCS) is a distributed sensing architecture that utilizes existing sensors on mobile units (MUs) to perform sensing tasks. A mobile crowdsensing platform (MCSP) publishes the sensing tasks and the MUs decide whether to participate in exchange for money. The MCS system is dynamic: the task requirements, the MUs' availability, and their available resources change over time. The MUs aim to find an efficient task participation strategy to maximize their income while the MCSP focuses on maximizing the number of completed tasks. As optimal strategies require perfect non-causal information about the MCS system, which is unavailable in realistic scenarios, the main challenge is to find an efficient task participation strategy for the MUs under incomplete information. To this end, a novel fully decentralized federated deep reinforcement learning algorithm, FDRL-PPO, is proposed. FDRL-PPO enables every MU to learn its own task participation strategy based on its experiences, available resources, and preferences, without relying on perfect non-causal information about the MCS system. To replenish their batteries, the MUs rely on energy harvesting. As a result, their available energy varies over time, leading to varying availability and fragmented learning experiences. To mitigate these challenges, the proposed approach leverages federated learning, enabling MUs to collaboratively improve their models without sharing private raw data like their own experiences. By exchanging only learned models, MUs collectively compensate for individual limitations, and find more scalable, robust, and efficient task participation strategies. Comprehensive evaluations on both synthetic and real-world datasets show that FDRL-PPO consistently outperforms benchmark algorithms in terms of task completion ratio, fairness in task completion, energy consumption, and number of conflicting proposals.
基于联邦强化学习的不完全信息下高效移动群智感知 / Federated Reinforcement Learning for Efficient Mobile Crowdsensing under Incomplete Information
本文针对移动用户在不了解全局信息时如何选择感知任务以赚取报酬的问题,提出了一种名为FDRL-PPO的分散式联邦深度强化学习算法,让用户通过本地学习和共享模型来优化策略,从而在保护隐私的同时提高任务完成率、公平性和能效。
源自 arXiv: 2605.02705