Abstract:
As a new distributed machine learning methodology, Federated Learning (FL) allows mobile devices (MDs) to collaboratively train a global model without sharing their raw d...Show MoreMetadata
Abstract:
As a new distributed machine learning methodology, Federated Learning (FL) allows mobile devices (MDs) to collaboratively train a global model without sharing their raw data in a privacy-preserving manner. However, it is a great challenge to schedule each MD and allocate various resources reasonably. This paper studies the joint optimization of computing resources used by MDs for FL training, the number of local iterations as well as WPT duration of each MD in a Wireless Power Transfer (WPT) assisted FL system, with the goal of maximizing the total utility of all MDs in the entire FL training process. Furthermore, we analyze the problem by using the Karush-Kuhn-Tucker (KKT) conditions and Lagrange dual method, and propose an improved Lagrangian subgradient method to solve this problem. Finally, extensive simulation experiments are conducted under various scenarios to verify the effectiveness of the proposed algorithm. The results show that our proposed algorithm has better performance in terms of the total utility of all MDs compared with other benchmark methods.
Published in: IEEE INFOCOM 2024 - IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS)
Date of Conference: 20-20 May 2024
Date Added to IEEE Xplore: 13 August 2024
ISBN Information: