FedProc: Prototypical contrastive federated learning on non-IID data

Federated learning (FL) enables multiple clients to jointly train high-performance deep learning models while maintaining the training data locally. However, it is challenging to accomplish this form of efficient collaborative learning when all the clients’ local data are not independent and identic...

Full description

Saved in:
Bibliographic Details
Published inFuture generation computer systems Vol. 143; pp. 93 - 104
Main Authors Mu, Xutong, Shen, Yulong, Cheng, Ke, Geng, Xueli, Fu, Jiaxuan, Zhang, Tao, Zhang, Zhiwei
Format Journal Article
LanguageEnglish
Published Elsevier B.V 01.06.2023
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Federated learning (FL) enables multiple clients to jointly train high-performance deep learning models while maintaining the training data locally. However, it is challenging to accomplish this form of efficient collaborative learning when all the clients’ local data are not independent and identically distributed (i.e., non-IID). Despite extensive efforts to address this challenge, the results for image classification tasks remain inadequate. In this paper, we propose FedProc: prototypical contrastive federated learning. The core idea of this approach is to utilize the prototypes as global knowledge to correct the drift of each client’s local training. Specifically, we designed a local network structure and a global prototype contrast loss to regulate the training of the local model. These efforts make the direction of local optimization consistent with the global optimum such that the global model achieves good performance on non-IID data. Evaluative studies supported by theoretical significance demonstrate that FedProc improves accuracy by 1.6% to 7.9% with an acceptable computational cost compared to state-of-the-art federated learning methods. •This paper proposes a novel federated learning framework (FedProc) to solve the non-IID data problem. The framework innovatively introduces a global class prototype to correct for local training, making the direction of local optimization consistent with the global optimization goal.•In this paper, we elaborate a generic hybrid local network architecture such that the local network takes full advantage of the potential knowledge provided by the global class prototype. The superior performance of this architecture is due to the design of a hybrid loss function.•We theoretically analyze the convergence of FedProc and obtain an upper bound on the convergence, which provides convergence guarantees for this work. In addition, experimental results show that FedProc is significantly better than the state-of-the-art methods in terms of accuracy and computational efficiency.
ISSN:0167-739X
1872-7115
DOI:10.1016/j.future.2023.01.019