Publication

On the Convergence of Hybrid Server-Clients Collaborative Training

Publication Info

Publication

On the Convergence of Hybrid Server-Clients Collaborative Training

Abstract

Modern distributed machine learning (ML) paradigms, such as federated learning (FL), utilize data distributed at different clients to train a global model. In such paradigm, local datasets never leave the clients for better privacy protection, and the parameter server (PS) only performs simple aggregation. In practice, however, there is often some amount of data available at the PS, and its computation capability is strong enough to carry out more demanding tasks than simple model aggregation. The focus of this paper is to analyze the model convergence of a new hybrid learning architecture, which leverages the PS dataset and its computation power for collaborative model training with clients. Different from FL where stochastic gradient descent (SGD) is always computed in parallel across clients, the new architecture has both parallel SGD at clients and sequential SGD at PS. We analyze the convergence rate upper bounds of this aggregate-then-advance design for both strongly convex and non-convex loss functions. We show that when the local SGD has an O(1/t) stepsize, the server SGD needs to scale its stepsize to no slower than O(1/t2) in order to strictly outperform local SGD with strongly convex loss functions. The theoretical findings are corroborated by numerical experiments, where advantages in terms of both accuracy and convergence speed over clients-only (local SGD and FED AVG) and server-only training are demonstrated.

CiTation

K. Yang, S. Chen and C. Shen, "On the Convergence of Hybrid Server-Clients Collaborative Training," in IEEE Journal on Selected Areas in Communications, vol. 41, no. 3, pp. 802-819, March 2023, doi: 10.1109/JSAC.2022.3229443.

Contributors

Info

Date:
December 28, 2022
Type:
Publication
DOI:
0.1109/JSAC.2022.3229443