Abstract

Conjugate gradient (CG) algorithms are widely applied to machine learning problems owing to their low calculation cost compared with second-order methods and better convergence properties relative to gradient descent methods. This study proposes a biased stochastic conjugate gradient (BSCG) algorithm with an adaptive step size. Specifically, the BSCG algorithm integrates both the stochastic recursive gradient method (SARAH) and the modified Barzilai–Borwein (BB) technique into the typical stochastic gradient algorithm. Compared with most of the available stochastic gradient methods, which obtain step size by selecting a fixed size elaborately and may lead to unfeasibility in finding an optimal solution, the proposed algorithm uses second-order information to gain an appropriate step size without increasing the computational cost. We not only prove that the proposed algorithm converges to global optimum, but also builds up its linear convergence rate for nonconvex objective functions. The numerical results of two machine learning models demonstrate show the superiority of the BSCG algorithm.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.