Abstract
The Broyden–Fletcher–Goldfarb–Shanno (BFGS) method plays an important role among the quasi-Newton algorithms for nonconvex and unconstrained optimization problems. However, in the proof of global convergence, BFGS-type methods generally need to assume that the gradient of the objective function is Lipschitz continuous. This issue prompts us to try to find quasi-Newton method for gradient non-Lipschitz continuous and nonconvex optimization based on the classical BFGS formula. In this paper, we propose an adaptive scaling damped BFGS method for gradient non-Lipschitz continuous and nonconvex problems. With Armijo or Weak Wolfe–Powell (WWP) line search, global convergence can be obtained. Under suitable conditions the convergence rate is superlinear with WWP-type line search. Applications of the given algorithms include the tested optimization problems, which turn out the proposed method is powerful and promising.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.