One of the main characteristics in many real-world big data scenarios is their distributed nature. In a machine learning context, distributed data, together with the requirements of preserving privacy and scaling up to large networks, brings the challenge of designing fully decentralized training protocols. In this paper, we explore the problem of distributed learning when the features of every pattern are available throughout multiple agents (as is happening, for example, in a distributed database scenario). We propose an algorithm for a particular class of neural networks, known as Random Vector Functional-Link (RVFL), which is based on the Alternating Direction Method of Multipliers optimization algorithm. The proposed algorithm allows to learn an RVFL network from multiple distributed data sources, while restricting communication to the unique operation of computing a distributed average. Our experimental simulations show that the algorithm is able to achieve a generalization accuracy comparable to a fully centralized solution, while at the same time being extremely efficient.
Learning from distributed data sources using random vector functional-link networks / Scardapane, Simone; Panella, Massimo; Comminiello, Danilo; Uncini, Aurelio. - In: PROCEDIA COMPUTER SCIENCE. - ISSN 1877-0509. - STAMPA. - 53:(2015), pp. 468-477. (Intervento presentato al convegno INNS Conference on Big Data 2015 tenutosi a San Francisco, CA, U.S.A. nel 8-10 agosto 2015) [10.1016/j.procs.2015.07.324].
Learning from distributed data sources using random vector functional-link networks
SCARDAPANE, SIMONE;PANELLA, Massimo;COMMINIELLO, DANILO;UNCINI, Aurelio
2015
Abstract
One of the main characteristics in many real-world big data scenarios is their distributed nature. In a machine learning context, distributed data, together with the requirements of preserving privacy and scaling up to large networks, brings the challenge of designing fully decentralized training protocols. In this paper, we explore the problem of distributed learning when the features of every pattern are available throughout multiple agents (as is happening, for example, in a distributed database scenario). We propose an algorithm for a particular class of neural networks, known as Random Vector Functional-Link (RVFL), which is based on the Alternating Direction Method of Multipliers optimization algorithm. The proposed algorithm allows to learn an RVFL network from multiple distributed data sources, while restricting communication to the unique operation of computing a distributed average. Our experimental simulations show that the algorithm is able to achieve a generalization accuracy comparable to a fully centralized solution, while at the same time being extremely efficient.File | Dimensione | Formato | |
---|---|---|---|
Scardapane_Learning_2015.pdf
accesso aperto
Tipologia:
Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza:
Creative commons
Dimensione
305.86 kB
Formato
Adobe PDF
|
305.86 kB | Adobe PDF |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.