In this thesis, we focus on problems in which the derivative of the objective function is either unavailable or unreliable, which can occur in a variety of situations including the presence of legacy codes (codes written in the past but not maintained), problems of parameter tuning for simulation or optimization algorithms and engineering problems where the objective functions are the output of black-box simulation software. Despite the absence or the unreliability of the derivatives, our interest is in the resolution of the optimization problem using gradient-based methods, which take advantage of the rich and relevant information normally included in the gradient of the objective function. We address the lack of derivatives considering two different scenarios. In the first one, we consider smooth problems with additive noise affecting objective function evaluations. We assume that objective function evaluations can be obtained in a cheap and fast way and we focus on gradient approximation methods that use objective function evaluations to somehow filter the noise and build an estimate of the gradient. In the second scenario, we consider potentially non-smooth simulation-based optimization problems in which neither the objective function nor its (eventual) derivative have an explicit expression. Assuming the expensiveness of the evaluations of objective functions, we enable the usage of gradient-based methods by following an approach that is based on the creation of a neural network model that replaces the simulation software used for computing the objective function. In this way, the smooth function obtained with the neural network model and its gradient are considered in the optimization procedure.

Enabling gradient-based optimization methods in problems with unreliable or absent derivatives / Boresta, Marco. - (2022 May 20).

Enabling gradient-based optimization methods in problems with unreliable or absent derivatives

BORESTA, MARCO
20/05/2022

Abstract

In this thesis, we focus on problems in which the derivative of the objective function is either unavailable or unreliable, which can occur in a variety of situations including the presence of legacy codes (codes written in the past but not maintained), problems of parameter tuning for simulation or optimization algorithms and engineering problems where the objective functions are the output of black-box simulation software. Despite the absence or the unreliability of the derivatives, our interest is in the resolution of the optimization problem using gradient-based methods, which take advantage of the rich and relevant information normally included in the gradient of the objective function. We address the lack of derivatives considering two different scenarios. In the first one, we consider smooth problems with additive noise affecting objective function evaluations. We assume that objective function evaluations can be obtained in a cheap and fast way and we focus on gradient approximation methods that use objective function evaluations to somehow filter the noise and build an estimate of the gradient. In the second scenario, we consider potentially non-smooth simulation-based optimization problems in which neither the objective function nor its (eventual) derivative have an explicit expression. Assuming the expensiveness of the evaluations of objective functions, we enable the usage of gradient-based methods by following an approach that is based on the creation of a neural network model that replaces the simulation software used for computing the objective function. In this way, the smooth function obtained with the neural network model and its gradient are considered in the optimization procedure.
20-mag-2022
File allegati a questo prodotto
File Dimensione Formato  
Tesi_dottorato_Boresta.pdf

accesso aperto

Tipologia: Tesi di dottorato
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 1.36 MB
Formato Adobe PDF
1.36 MB Adobe PDF

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1635357
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
social impact