The computing power of various Internet of Things (IoT) devices is quite different. To enable IoT devices with lower computing power to perform machine learning, all nodes can only train smaller models, which results in the waste of computing power for high-performance devices. In this article, a heterogeneous model fusion federated learning (HFL) mechanism is proposed. Each node trains learning models of different scales according to its own computing capabilities. After receiving the gradient trained by each node, the parameter server (PS) corrects the received gradient with the repeat matrix, and then update the corresponding region of the global model according to the mapping matrix. After all update operations are over, the PS assigns the compressed model to the corresponding node. This article uses a variety of experimental schemes to evaluate the proposed method, including three data sets, two model structures, and three computational complexity levels. The proposed method have been proved it not only maximizes the use of the unbalanced computing power of edge nodes, but also enables different structural models to compensate for the shortcomings of others, improving the overall performance.
Heterogeneous Model Fusion Federated Learning Mechanism Based on Model Mapping / Lu, X.; Liao, Y.; Liu, C.; Lio, P.; Hui, P.. - In: IEEE INTERNET OF THINGS JOURNAL. - ISSN 2327-4662. - 9:8(2022), pp. 6058-6068. [10.1109/JIOT.2021.3110908]
Heterogeneous Model Fusion Federated Learning Mechanism Based on Model Mapping
Lio P.;
2022
Abstract
The computing power of various Internet of Things (IoT) devices is quite different. To enable IoT devices with lower computing power to perform machine learning, all nodes can only train smaller models, which results in the waste of computing power for high-performance devices. In this article, a heterogeneous model fusion federated learning (HFL) mechanism is proposed. Each node trains learning models of different scales according to its own computing capabilities. After receiving the gradient trained by each node, the parameter server (PS) corrects the received gradient with the repeat matrix, and then update the corresponding region of the global model according to the mapping matrix. After all update operations are over, the PS assigns the compressed model to the corresponding node. This article uses a variety of experimental schemes to evaluate the proposed method, including three data sets, two model structures, and three computational complexity levels. The proposed method have been proved it not only maximizes the use of the unbalanced computing power of edge nodes, but also enables different structural models to compensate for the shortcomings of others, improving the overall performance.File | Dimensione | Formato | |
---|---|---|---|
lU_Heterogeneous_2022.pdf
solo gestori archivio
Tipologia:
Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza:
Tutti i diritti riservati (All rights reserved)
Dimensione
6.96 MB
Formato
Adobe PDF
|
6.96 MB | Adobe PDF | Contatta l'autore |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.