The paper presents an overview of global issues in optimizationmethods for training feedforward neural networks (FNN) in a regression setting.We first recall the learning optimization paradigm for FNN and we briefly discuss global scheme for the joint choice of the network topologies and of the network parameters. The main part of the paper focuses on the core subproblem which is the continuous unconstrained (regularized) weights optimization problem with the aim of reviewing global methods specifically arising both in multi layer perceptron/deep networks and in radial basis networks.We review some recent results on the existence of non-global stationary points of the unconstrained nonlinear problem and the role of determining a global solution in a supervised learning paradigm. Local algorithms that are widespread used to solve the continuous unconstrained problems are addressed with focus on possible improvements to exploit the global properties. Hybrid global methods specifically devised for FNN training optimization problems which embed local algorithms are discussed too.
Global optimization issues in deep network regression: an overview / Palagi, Laura. - In: JOURNAL OF GLOBAL OPTIMIZATION. - ISSN 0925-5001. - 73:2(2019), pp. 239-277.
|Titolo:||Global optimization issues in deep network regression: an overview|
PALAGI, Laura (Corresponding author)
|Data di pubblicazione:||2019|
|Citazione:||Global optimization issues in deep network regression: an overview / Palagi, Laura. - In: JOURNAL OF GLOBAL OPTIMIZATION. - ISSN 0925-5001. - 73:2(2019), pp. 239-277.|
|Appartiene alla tipologia:||01a Articolo in rivista|