Federated learning (FedL) is a machine learning (ML) technique utilized to train deep neural networks (DeepNNs) in a distributed way without the need to share data among the federated training clients. FedL was proposed for edge computing and Internet of things (IoT) tasks in which a centralized server was responsible for coordinating and governing the training process. To remove the design limitation implied by the centralized entity, this work proposes two different solutions to decentralize existing FedL algorithms, enabling the application of FedL on networks with arbitrary communication topologies, and thus extending the domain of application of FedL to more complex scenarios and new tasks. Of the two proposed algorithms, one, called FedLCon, is developed based on results from discrete-time weighted average consensus theory and is able to reconstruct the performances of the standard centralized FedL solutions, as also shown by the reported validation tests.

A Weighted Average Consensus Approach for Decentralized Federated Learning / Giuseppi, A.; Manfredi, S.; Pietrabissa, A.. - In: MACHINE INTELLIGENCE RESEARCH. - ISSN 2731-5398. - 19:4(2022), pp. 319-330. [10.1007/s11633-022-1338-z]

A Weighted Average Consensus Approach for Decentralized Federated Learning

Giuseppi A.
;
Pietrabissa A.
2022

Abstract

Federated learning (FedL) is a machine learning (ML) technique utilized to train deep neural networks (DeepNNs) in a distributed way without the need to share data among the federated training clients. FedL was proposed for edge computing and Internet of things (IoT) tasks in which a centralized server was responsible for coordinating and governing the training process. To remove the design limitation implied by the centralized entity, this work proposes two different solutions to decentralize existing FedL algorithms, enabling the application of FedL on networks with arbitrary communication topologies, and thus extending the domain of application of FedL to more complex scenarios and new tasks. Of the two proposed algorithms, one, called FedLCon, is developed based on results from discrete-time weighted average consensus theory and is able to reconstruct the performances of the standard centralized FedL solutions, as also shown by the reported validation tests.
2022
artificial intelligence; deep learning; discrete-time consensus; distributed systems; federated averaging (FedAvg); federated learning (FedL); machine learning (ML)
01 Pubblicazione su rivista::01a Articolo in rivista
A Weighted Average Consensus Approach for Decentralized Federated Learning / Giuseppi, A.; Manfredi, S.; Pietrabissa, A.. - In: MACHINE INTELLIGENCE RESEARCH. - ISSN 2731-5398. - 19:4(2022), pp. 319-330. [10.1007/s11633-022-1338-z]
File allegati a questo prodotto
File Dimensione Formato  
Giuseppi_preprint_A-weighted_2021.pdf

accesso aperto

Note: 10.1007/s11633-022-1338-z
Tipologia: Documento in Pre-print (manoscritto inviato all'editore, precedente alla peer review)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 1.37 MB
Formato Adobe PDF
1.37 MB Adobe PDF
Giuseppi_A-weighted_2022.pdf

solo gestori archivio

Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 1.11 MB
Formato Adobe PDF
1.11 MB Adobe PDF   Contatta l'autore

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1654509
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 11
  • ???jsp.display-item.citation.isi??? 9
social impact