ArXiv TLDR

Federated Reinforcement Learning for Efficient Mobile Crowdsensing under Incomplete Information

🐦 Tweet
2605.02705

Sumedh J. Dongare, Patrick Weber, Andrea Ortiz, Walid Saad, Oliver Hinz + 1 more

cs.LGcs.NI

TLDR

FDRL-PPO uses federated deep reinforcement learning to help mobile units efficiently participate in crowdsensing tasks despite incomplete information and varying energy.

Key contributions

  • Proposes FDRL-PPO, a decentralized federated deep reinforcement learning algorithm for mobile crowdsensing (MCS).
  • Enables mobile units (MUs) to learn optimal task participation strategies under incomplete system information.
  • Leverages federated learning to mitigate fragmented experiences and varying energy from harvesting.
  • Achieves superior performance in task completion, fairness, energy efficiency, and fewer conflicting proposals.

Why it matters

This paper tackles the challenge of efficient task participation in dynamic mobile crowdsensing systems with incomplete information. Its federated reinforcement learning approach enables mobile units to collaboratively learn robust strategies, overcoming individual limitations and preserving data privacy. This significantly improves scalability and efficiency in crowdsensing operations.

Original Abstract

Mobile crowdsensing (MCS) is a distributed sensing architecture that utilizes existing sensors on mobile units (MUs) to perform sensing tasks. A mobile crowdsensing platform (MCSP) publishes the sensing tasks and the MUs decide whether to participate in exchange for money. The MCS system is dynamic: the task requirements, the MUs' availability, and their available resources change over time. The MUs aim to find an efficient task participation strategy to maximize their income while the MCSP focuses on maximizing the number of completed tasks. As optimal strategies require perfect non-causal information about the MCS system, which is unavailable in realistic scenarios, the main challenge is to find an efficient task participation strategy for the MUs under incomplete information. To this end, a novel fully decentralized federated deep reinforcement learning algorithm, FDRL-PPO, is proposed. FDRL-PPO enables every MU to learn its own task participation strategy based on its experiences, available resources, and preferences, without relying on perfect non-causal information about the MCS system. To replenish their batteries, the MUs rely on energy harvesting. As a result, their available energy varies over time, leading to varying availability and fragmented learning experiences. To mitigate these challenges, the proposed approach leverages federated learning, enabling MUs to collaboratively improve their models without sharing private raw data like their own experiences. By exchanging only learned models, MUs collectively compensate for individual limitations, and find more scalable, robust, and efficient task participation strategies. Comprehensive evaluations on both synthetic and real-world datasets show that FDRL-PPO consistently outperforms benchmark algorithms in terms of task completion ratio, fairness in task completion, energy consumption, and number of conflicting proposals.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.