Please use this identifier to cite or link to this item: http://bura.brunel.ac.uk/handle/2438/29341
Title: Representative Kernels-based CNN for Faster Transmission in Federated Learning
Authors: Li, W
Shen, Z
Liu, X
Wang, M
Ma, C
Ding, C
Cao, J
Keywords: federated learning;convolution neural network;representative kernels;kernel generation function;parameter reduction;module selection
Issue Date: 4-Jul-2024
Publisher: Institute of Electrical and Electronics Engineers (IEEE)
Citation: Li, W. et al. (2024) 'Representative Kernels-based CNN for Faster Transmission in Federated Learning', IEEE Transactions on Mobile Computing, 0 (early access), pp. 1 - 14. doi: 10.1109/TMC.2024.3423448.
Abstract: Federated Learning (FL) has attracted many attentions because of its ability to ensure data privacy and security. In FL, due to the contradiction between limited bandwidth and huge transmission parameters, it has been an ongoing challenge to reduce the model parameters that need to be transmitted to the server in the clients for fast transmission. Existing works that attempt to reduce the amount of transmitted parameters have limitations: 1) the reduced number of parameters is not significant; 2) the performance of the global model is limited. In this paper, we propose a novel method called Fed-KGF that significantly reduces the amount of model parameters that need to be transferred while improving the performance of the global model in FL. Since convolution kernels in a Convolution Neural Network (CNN) account for most parameters, our goal is to reduce the parameter transmission between the clients and the server by reducing the number of convolution kernels. Specifically, we construct an incomplete model with a few representative kernels, referred to the convolution kernels, that are solely updated during training. We propose Kernel Generation Function (KGF) to generate convolution kernels to render the incomplete model to be a complete one, and discard those generated kernels after training local models. The parameters that need to be transmitted only reside in the representative kernels, which are significantly reduced. Furthermore, there is a client-drift in the traditional FL because it adopts the averaging method, which hurts the global model performance. We innovatively select one or few modules (a module indicates a convolution function + several non-convolution functions) from all client models in a permutation way, and only aggregate the uploaded modules rather than averaging them in server to reduce client-drift, thereby improving the performance of the global model and further reducing the transmitted parameters. Experimental results on both non-Independent and Identically Distributed (non-IID) and IID scenarios for image classification and object detection tasks demonstrate that our Fed-KGF outperforms the SOTA methods. Fed-KGF achieves approximately 11% higher classification accuracy and roughly 33% fewer parameters than the recent FedCAMS model on CIFAR-10, and gains approximately 3.64% higher detection precision and around 37% fewer parameters than the SOTA SmartIdx model on COCO2017 datasets.
URI: https://bura.brunel.ac.uk/handle/2438/29341
DOI: https://doi.org/10.1109/TMC.2024.3423448
ISSN: 1536-1233
Other Identifiers: ORCiD: Wei Li https://orcid.org/0000-0002-3135-0447
ORCiD: Xiulong Liu https://orcid.org/0000-0002-4746-5599
ORCiD: Mingfeng Wang https://orcid.org/0000-0001-6551-0325
ORCiD: Chao Ma https://orcid.org/0000-0002-7443-6267
ORCiD: Chuntao Ding https://orcid.org/0000-0001-8362-8407
ORCiD: Jiannong Cao https://orcid.org/0000-0002-2725-2529
Appears in Collections:Dept of Mechanical and Aerospace Engineering Research Papers

Files in This Item:
File Description SizeFormat 
FullText.pdfCopyright © 2024 Institute of Electrical and Electronics Engineers (IEEE). Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. See: https://journals.ieeeauthorcenter.ieee.org/become-an-ieee-journal-author/publishing-ethics/guidelines-and-policies/post-publication-policies/11.6 MBAdobe PDFView/Open


Items in BURA are protected by copyright, with all rights reserved, unless otherwise indicated.