In this paper, we propose a new approach to train a deep neural network with multiple intermediate auxiliary classifiers, branching from it. These ‘multi-exits’ models can be used to reduce the inference time by performing early exit on the intermediate branches, if the confidence of the prediction is higher than a threshold. They rely on the assumption that not all the samples require the same amount of processing to yield a good prediction. In this paper, we propose a way to train jointly all the branches of a multi-exit model without hyper-parameters, by weighting the predictions from each branch with a trained confidence score. Each confidence score is an approximation of the real one produced by the branch, and it is calculated and regularized while training the rest of the model. We evaluate our proposal on a set of image classification benchmarks, using different neural models and early-exit stopping criteria.

A probabilistic re-intepretation of confidence scores in multi-exit models / Pomponi, J.; Scardapane, S.; Uncini, A.. - In: ENTROPY. - ISSN 1099-4300. - 24:1(2022), pp. 1-14. [10.3390/e24010001]

A probabilistic re-intepretation of confidence scores in multi-exit models

Pomponi J.
;
Scardapane S.;Uncini A.
2022

Abstract

In this paper, we propose a new approach to train a deep neural network with multiple intermediate auxiliary classifiers, branching from it. These ‘multi-exits’ models can be used to reduce the inference time by performing early exit on the intermediate branches, if the confidence of the prediction is higher than a threshold. They rely on the assumption that not all the samples require the same amount of processing to yield a good prediction. In this paper, we propose a way to train jointly all the branches of a multi-exit model without hyper-parameters, by weighting the predictions from each branch with a trained confidence score. Each confidence score is an approximation of the real one produced by the branch, and it is calculated and regularized while training the rest of the model. We evaluate our proposal on a set of image classification benchmarks, using different neural models and early-exit stopping criteria.
2022
adaptive computation; branch neural networks; deep learning; deep neural networks; fast inference
01 Pubblicazione su rivista::01a Articolo in rivista
A probabilistic re-intepretation of confidence scores in multi-exit models / Pomponi, J.; Scardapane, S.; Uncini, A.. - In: ENTROPY. - ISSN 1099-4300. - 24:1(2022), pp. 1-14. [10.3390/e24010001]
File allegati a questo prodotto
File Dimensione Formato  
Pomponi_Probabilistic-re-intepretationentropy_2022.pdf

accesso aperto

Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza: Creative commons
Dimensione 587.29 kB
Formato Adobe PDF
587.29 kB Adobe PDF

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1612482
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 3
  • ???jsp.display-item.citation.isi??? 3
social impact