Residual deep neural networks (ResNets) are mathematically described as interacting particle systems. In the case of infinitely many layers the ResNet leads to a system of coupled system of ordinary differential equations known as neural differential equations. For large scale input data we derive a mean-field limit and show well-posedness of the resulting description. Further, we analyze the existence of solutions to the training process by using both a controllability and an optimal control point of view. Numerical investigations based on the solution of a formal optimality system illustrate the theoretical findings.
Continuous limits of residual neural networks in case of large input data / Herty, M; Thunen, A; Trimborn, T; Visconti, G. - In: COMMUNICATIONS IN APPLIED AND INDUSTRIAL MATHEMATICS. - ISSN 2038-0909. - 13:1(2022), pp. 96-120. [10.2478/caim-2022-0008]
Continuous limits of residual neural networks in case of large input data
Herty, M;Visconti, G
2022
Abstract
Residual deep neural networks (ResNets) are mathematically described as interacting particle systems. In the case of infinitely many layers the ResNet leads to a system of coupled system of ordinary differential equations known as neural differential equations. For large scale input data we derive a mean-field limit and show well-posedness of the resulting description. Further, we analyze the existence of solutions to the training process by using both a controllability and an optimal control point of view. Numerical investigations based on the solution of a formal optimality system illustrate the theoretical findings.File | Dimensione | Formato | |
---|---|---|---|
Herty_Continuous-limits_2022.pdf
accesso aperto
Tipologia:
Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza:
Creative commons
Dimensione
945.95 kB
Formato
Adobe PDF
|
945.95 kB | Adobe PDF |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.