Several Guaranteed Descent Conjugate Gradient Methods for Unconstrained Optimization
San-Yang Liu and
Yuan-Yuan Huang
Journal of Applied Mathematics, 2014, vol. 2014, issue 1
Abstract:
This paper investigates a general form of guaranteed descent conjugate gradient methods which satisfies the descent condition gkTdk≤-1-1/4θkgk2 θk>1/4 and which is strongly convergent whenever the weak Wolfe line search is fulfilled. Moreover, we present several specific guaranteed descent conjugate gradient methods and give their numerical results for large‐scale unconstrained optimization.
Date: 2014
References: Add references at CitEc
Citations:
Downloads: (external link)
https://doi.org/10.1155/2014/825958
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:wly:jnljam:v:2014:y:2014:i:1:n:825958
Access Statistics for this article
More articles in Journal of Applied Mathematics from John Wiley & Sons
Bibliographic data for series maintained by Wiley Content Delivery ().