A limited-memory optimization method using the infinitely many times repeated BNS update and conjugate directions

Abstract
To improve the performance of the limited-memory variable metric L-BFGS method for large scale unconstrained optimization, repeating of some BFGS updates was proposed e.g. in Al-Baali (1999, 2002). Since the repeating process can be time consuming, the suitable extra updates need to be selected carefully. We show that for the limited-memory variable metric BNS method, matrix updating can be efficiently repeated infinitely many times under some conditions, with only a small increase of the number of arithmetic operations. The limit matrix can be written as a block BFGS update (Vlcek and Luksan, 2018), which can be obtained by solving of some low-order Lyapunov matrix equation. The resulting method can be advantageously combined with methods based on vector corrections for conjugacy, see e.g. Vlcek and Luksan (2015). Global convergence of the proposed algorithm is established for convex and sufficiently smooth functions. Numerical experiments demonstrate the efficiency of the new method.
Description
Subject(s)
Numerical results, Global convergence, The Lyapunov matrix equation, The repeated Byrd-Nocedal-Schnabel update, Limited-memory variable metric methods, Unconstrained minimization
Citation
ISSN
ISBN
Collections