Two modified scaled nonlinear conjugate gradient methods
作者:
Highlights:
•
摘要
Following the scaled conjugate gradient methods proposed by Andrei, we hybridize the memoryless BFGS preconditioned conjugate gradient method suggested by Shanno and the spectral conjugate gradient method suggested by Birgin and Martínez based on a modified secant equation suggested by Yuan, and propose two modified scaled conjugate gradient methods. The interesting features of our methods are applying the function values in addition to the gradient values and satisfying the sufficient descent condition for the generated search directions which leads to the global convergence for uniformly convex functions. Numerical comparisons between the implementations of one of our methods which generates descent search directions for general functions and an efficient scaled conjugate gradient method proposed by Andrei are made on a set of unconstrained optimization test problems from the CUTEr collection, using the performance profile introduced by Dolan and Moré.
论文关键词:65K05,90C53,49M37,Unconstrained optimization,Scaled nonlinear conjugate gradient method,BFGS update,Modified secant equation,Sufficient descent condition,Global convergence
论文评审过程:Received 4 December 2011, Revised 31 May 2013, Available online 13 November 2013.
论文官网地址:https://doi.org/10.1016/j.cam.2013.11.001