This article summarizes principles and ideas from the emerging area of applying conditional computation methods to the design of neural networks. In particular, we focus on neural networks that can dynamically activate or de-activate parts of their computational graph conditionally on their input. Examples include the dynamic selection of, e.g., input tokens, layers (or sets of layers), and sub-modules inside each layer (e.g., channels in a convolutional filter). We first provide a general formalism to describe these techniques in an uniform way. Then, we introduce three notable implementations of these principles: mixture-of-experts (MoEs) networks, token selection mechanisms, and early-exit neural networks. The paper aims to provide a tutorial-like introduction to this growing field. To this end, we analyze the benefits of these modular designs in terms of efficiency, explainability, and transfer learning, with a focus on emerging applicative areas ranging from automated scientific discovery to semantic communication.
Conditional computation in neural networks: Principles and research trends / Scardapane, Simone; Baiocchi, Alessandro; Devoto, Alessio; Marsocci, Valerio; Minervini, Pasquale; Pomponi, Jary. - In: INTELLIGENZA ARTIFICIALE. - ISSN 1724-8035. - 18:1(2024), pp. 175-190. [10.3233/ia-240035]
Conditional computation in neural networks: Principles and research trends
Scardapane, Simone
;Baiocchi, Alessandro;Devoto, Alessio;Marsocci, Valerio;Pomponi, Jary
2024
Abstract
This article summarizes principles and ideas from the emerging area of applying conditional computation methods to the design of neural networks. In particular, we focus on neural networks that can dynamically activate or de-activate parts of their computational graph conditionally on their input. Examples include the dynamic selection of, e.g., input tokens, layers (or sets of layers), and sub-modules inside each layer (e.g., channels in a convolutional filter). We first provide a general formalism to describe these techniques in an uniform way. Then, we introduce three notable implementations of these principles: mixture-of-experts (MoEs) networks, token selection mechanisms, and early-exit neural networks. The paper aims to provide a tutorial-like introduction to this growing field. To this end, we analyze the benefits of these modular designs in terms of efficiency, explainability, and transfer learning, with a focus on emerging applicative areas ranging from automated scientific discovery to semantic communication.| File | Dimensione | Formato | |
|---|---|---|---|
|
Scardapane_Conditional_2024.pdf
solo gestori archivio
Tipologia:
Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza:
Tutti i diritti riservati (All rights reserved)
Dimensione
116.11 kB
Formato
Adobe PDF
|
116.11 kB | Adobe PDF | Contatta l'autore |
|
Scardapane_preprint_Conditional_2024.pdf
accesso aperto
Note: DOI 10.3233/IA-240035
Tipologia:
Documento in Pre-print (manoscritto inviato all'editore, precedente alla peer review)
Licenza:
Creative commons
Dimensione
511.37 kB
Formato
Adobe PDF
|
511.37 kB | Adobe PDF |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.


