Abstract

Sharp convergence rates of stochastic approximation algorithms are given for the case where the derivative of the unknown regression function at the sought-for root is zero. The convergence rates obtained are sharp for the general step size used in the algorithms in contrast to the previous work where they are not sharp for slowly decreasing step sizes; all possible limit points are found for the case where the first matrix coefficient in the expansion of the regression function is normal; and the estimation upper bound is shown to be achieved for the multi-dimensional case in contrast to the previous work where only the one-dimensional result is proved.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.