Two modified scaled nonlinear conjugate gradient methods

  • Authors:
  • Saman Babaie-Kafaki

  • Affiliations:
  • -

  • Venue:
  • Journal of Computational and Applied Mathematics
  • Year:
  • 2014

Quantified Score

Hi-index 7.29

Visualization

Abstract

Following the scaled conjugate gradient methods proposed by Andrei, we hybridize the memoryless BFGS preconditioned conjugate gradient method suggested by Shanno and the spectral conjugate gradient method suggested by Birgin and Martinez based on a modified secant equation suggested by Yuan, and propose two modified scaled conjugate gradient methods. The interesting features of our methods are applying the function values in addition to the gradient values and satisfying the sufficient descent condition for the generated search directions which leads to the global convergence for uniformly convex functions. Numerical comparisons between the implementations of one of our methods which generates descent search directions for general functions and an efficient scaled conjugate gradient method proposed by Andrei are made on a set of unconstrained optimization test problems from the CUTEr collection, using the performance profile introduced by Dolan and More.