Abstract:
In federated learning (FL) systems, the central server typically samples a subset of participating clients at each global iteration for model training. To mitigate privac...Show MoreMetadata
Abstract:
In federated learning (FL) systems, the central server typically samples a subset of participating clients at each global iteration for model training. To mitigate privacy leakage, clients may insert noise into local parameters before uploading them for global aggregation, leading to FL model performance degradation. This paper aims to design a Privacy-aware Client Sampling framework in FEDerated learning, named FedPCS, to tackle the heterogeneous client sampling issues and improve model performance. First, we obtain a pioneering upper bound for the accuracy loss of the FL model with privacy-aware client sampling probabilities. Based on this, we model the interactions between the central server and participating clients as a two-stage Stackelberg game. In Stage I, the central server designs the optimal time-dependent reward for cost minimization by considering the trade-off between the accuracy loss of the FL model and the rewards allocated. In Stage II, each client determines the correction factor that dynamically adjusts its privacy budget based on the reward allocated to maximize its utility. To surmount the obstacle of approximating other clients’ private information, we introduce the mean-field estimator to estimate the average privacy budget. We analytically demonstrate the existence and convergence of the fixed point for the mean-field estimator and derive the Stackelberg Nash Equilibrium to obtain the optimal strategy profile. Through rigorously theoretical convergence analysis, we guarantee the robustness of our proposed FedPCS. Moreover, considering the conventional sampling strategy in privacy-preserving federated learning, we prove that the random sampling approach’s price of anarchy (PoA) can be arbitrarily large. To remedy such efficiency loss, we show that the proposed privacy-aware client sampling strategy successfully reduces PoA, which is upper bounded by a reachable constant. To address the challenge of varying privacy requirements throughout diff...
Published in: IEEE Transactions on Networking ( Early Access )