This paper includes a twofold result for the Nonlinear Conjugate Gradient (NCG) method, in large scale unconstrained optimization. First we consider a theoretical analysis, where preconditioning is embedded in a strong convergence framework of an NCG method from the literature. Mild conditions to be satisfied by the preconditioners are defined, in order to preserve NCG convergence. As a second task, we also detail the use of novel matrix-free preconditioners for NCG. Our proposals are based on quasi-Newton updates, and either satisfy the secant equation or a secant-like condition at some of the previous iterates. We show that, in some sense, the preconditioners we propose also approximate the inverse of the Hessian matrix. In particular, the structures of our preconditioners depend on low-rank updates used, along with different choices of specific parameters. The low-rank updates are obtained as by product of NCG iterations. The results of an extended numerical experience using large scale CUTEst problems is reported, showing that our preconditioners can considerably improve the performance of NCG methods.
Preconditioned nonlinear conjugate gradient methods based on a modified secant equation / Caliciotti, Andrea; Fasano, Giovanni; Roma, Massimo. - In: APPLIED MATHEMATICS AND COMPUTATION. - ISSN 0096-3003. - STAMPA. - 318:(2018), pp. 196-214. [10.1016/j.amc.2017.08.029]
Preconditioned nonlinear conjugate gradient methods based on a modified secant equation
CALICIOTTI, ANDREA
;ROMA, Massimo
2018
Abstract
This paper includes a twofold result for the Nonlinear Conjugate Gradient (NCG) method, in large scale unconstrained optimization. First we consider a theoretical analysis, where preconditioning is embedded in a strong convergence framework of an NCG method from the literature. Mild conditions to be satisfied by the preconditioners are defined, in order to preserve NCG convergence. As a second task, we also detail the use of novel matrix-free preconditioners for NCG. Our proposals are based on quasi-Newton updates, and either satisfy the secant equation or a secant-like condition at some of the previous iterates. We show that, in some sense, the preconditioners we propose also approximate the inverse of the Hessian matrix. In particular, the structures of our preconditioners depend on low-rank updates used, along with different choices of specific parameters. The low-rank updates are obtained as by product of NCG iterations. The results of an extended numerical experience using large scale CUTEst problems is reported, showing that our preconditioners can considerably improve the performance of NCG methods.File | Dimensione | Formato | |
---|---|---|---|
Caliciotti_Preconditioned_2018.pdf
solo gestori archivio
Note: https://doi.org/10.1016/j.amc.2017.08.029
Tipologia:
Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza:
Tutti i diritti riservati (All rights reserved)
Dimensione
1.23 MB
Formato
Adobe PDF
|
1.23 MB | Adobe PDF | Contatta l'autore |
Caliciotti_preprint_Preconditioned_2018.pdf
accesso aperto
Note: https://doi.org/10.1016/j.amc.2017.08.029
Tipologia:
Documento in Pre-print (manoscritto inviato all'editore, precedente alla peer review)
Licenza:
Tutti i diritti riservati (All rights reserved)
Dimensione
542.42 kB
Formato
Adobe PDF
|
542.42 kB | Adobe PDF |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.