Among Bayesian methods, Monte Carlo dropout provides principled tools for evaluating the epistemic uncertainty of neural networks. Its popularity recently led to seminal works that proposed activating the dropout layers only during inference for evaluating epistemic uncertainty. This approach, which we call dropout injection, provides clear benefits over its traditional counterpart (which we call embedded dropout) since it allows one to obtain a post hoc uncertainty measure for any existing network previously trained without dropout, avoiding an additional, time-consuming training process. Unfortunately, no previous work thoroughly analyzed injected dropout and compared it with embedded dropout; therefore, we provide a first comprehensive investigation, focusing on regression problems. We show that the effectiveness of dropout injection strongly relies on a suitable scaling of the corresponding uncertainty measure, and propose an alternative method to implement it. We also considered the trade-off between negative log-likelihood and calibration error as a function of the scale factor. Experimental results on benchmark data sets from several regression tasks, including crowd counting, support our claim that dropout injection can effectively behave as a competitive post hoc alternative to embedded dropout.

Dropout injection at test time for post hoc uncertainty quantification in neural networks / Ledda, Emanuele; Fumera, Giorgio; Roli, Fabio. - In: INFORMATION SCIENCES. - ISSN 0020-0255. - 645:(2023), p. 119356. [10.1016/j.ins.2023.119356]

Dropout injection at test time for post hoc uncertainty quantification in neural networks

Emanuele Ledda
;
2023

Abstract

Among Bayesian methods, Monte Carlo dropout provides principled tools for evaluating the epistemic uncertainty of neural networks. Its popularity recently led to seminal works that proposed activating the dropout layers only during inference for evaluating epistemic uncertainty. This approach, which we call dropout injection, provides clear benefits over its traditional counterpart (which we call embedded dropout) since it allows one to obtain a post hoc uncertainty measure for any existing network previously trained without dropout, avoiding an additional, time-consuming training process. Unfortunately, no previous work thoroughly analyzed injected dropout and compared it with embedded dropout; therefore, we provide a first comprehensive investigation, focusing on regression problems. We show that the effectiveness of dropout injection strongly relies on a suitable scaling of the corresponding uncertainty measure, and propose an alternative method to implement it. We also considered the trade-off between negative log-likelihood and calibration error as a function of the scale factor. Experimental results on benchmark data sets from several regression tasks, including crowd counting, support our claim that dropout injection can effectively behave as a competitive post hoc alternative to embedded dropout.
2023
Crowd counting; Epistemic uncertainty; Monte Carlo dropout; Trustworthy AI; Uncertainty quantification
01 Pubblicazione su rivista::01a Articolo in rivista
Dropout injection at test time for post hoc uncertainty quantification in neural networks / Ledda, Emanuele; Fumera, Giorgio; Roli, Fabio. - In: INFORMATION SCIENCES. - ISSN 0020-0255. - 645:(2023), p. 119356. [10.1016/j.ins.2023.119356]
File allegati a questo prodotto
File Dimensione Formato  
Ledda_Dropout_2023.pdf

solo gestori archivio

Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 2.76 MB
Formato Adobe PDF
2.76 MB Adobe PDF   Contatta l'autore
Ledda_Preprint_Dropout_2023.pdf

accesso aperto

Note: https://doi.org/10.1016/j.ins.2023.119356
Tipologia: Documento in Pre-print (manoscritto inviato all'editore, precedente alla peer review)
Licenza: Creative commons
Dimensione 5.34 MB
Formato Adobe PDF
5.34 MB Adobe PDF

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1690455
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 6
  • ???jsp.display-item.citation.isi??? 3
social impact