Neural networks are generally built by interleaving (adaptable) linear layers with (fixed) nonlinear activation functions. To increase their flexibility, several authors have proposed methods for adapting the activation functions themselves, endowing them with varying degrees of flexibility. None of these approaches, however, have gained wide acceptance in practice, and research in this topic remains open. In this paper, we introduce a novel family of flexible activation functions that are based on an inexpensive kernel expansion at every neuron. Leveraging several properties of kernel-based models, we propose multiple variations for designing and initializing these kernel activation functions (KAFs), including a multidimensional scheme allowing to nonlinearly combine information from different paths in the network. The resulting KAFs can approximate any mapping defined over a subset of the real line, either convex or non-convex. Furthermore, they are smooth over their entire domain, linear in their parameters, and they can be regularized using any known scheme, including the use of ℓ 1 penalties to enforce sparseness. To the best of our knowledge, no other known model satisfies all these properties simultaneously. In addition, we provide an overview on alternative techniques for adapting the activation functions, which is currently lacking in the literature. A large set of experiments validates our proposal.

Kafnets: Kernel-based non-parametric activation functions for neural networks / Scardapane, S.; Van Vaerenbergh, S.; Totaro, S.; Uncini, A.. - In: NEURAL NETWORKS. - ISSN 0893-6080. - 110:(2019), pp. 19-32. [10.1016/j.neunet.2018.11.002]

Kafnets: Kernel-based non-parametric activation functions for neural networks

Scardapane S.;Uncini A.
2019

Abstract

Neural networks are generally built by interleaving (adaptable) linear layers with (fixed) nonlinear activation functions. To increase their flexibility, several authors have proposed methods for adapting the activation functions themselves, endowing them with varying degrees of flexibility. None of these approaches, however, have gained wide acceptance in practice, and research in this topic remains open. In this paper, we introduce a novel family of flexible activation functions that are based on an inexpensive kernel expansion at every neuron. Leveraging several properties of kernel-based models, we propose multiple variations for designing and initializing these kernel activation functions (KAFs), including a multidimensional scheme allowing to nonlinearly combine information from different paths in the network. The resulting KAFs can approximate any mapping defined over a subset of the real line, either convex or non-convex. Furthermore, they are smooth over their entire domain, linear in their parameters, and they can be regularized using any known scheme, including the use of ℓ 1 penalties to enforce sparseness. To the best of our knowledge, no other known model satisfies all these properties simultaneously. In addition, we provide an overview on alternative techniques for adapting the activation functions, which is currently lacking in the literature. A large set of experiments validates our proposal.
2019
Activation functions; Kernel methods; neural networks; algorithms; databases, factual; Neural Networks (Computer); neurons
01 Pubblicazione su rivista::01a Articolo in rivista
Kafnets: Kernel-based non-parametric activation functions for neural networks / Scardapane, S.; Van Vaerenbergh, S.; Totaro, S.; Uncini, A.. - In: NEURAL NETWORKS. - ISSN 0893-6080. - 110:(2019), pp. 19-32. [10.1016/j.neunet.2018.11.002]
File allegati a questo prodotto
File Dimensione Formato  
Scardapane_Kafnets_2019.pdf

solo gestori archivio

Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 827.61 kB
Formato Adobe PDF
827.61 kB Adobe PDF   Contatta l'autore

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1290928
Citazioni
  • ???jsp.display-item.citation.pmc??? 2
  • Scopus 44
  • ???jsp.display-item.citation.isi??? 37
social impact