Abstract
This paper presents a new approach to learning in recurrent neural networks, based on the descent of the error functional in the space of the linear outputs of the neurons (neuron space approach). At each step of the learning process a linear system is solved for the weights using a recursive least squares technique. This approach, with respect to traditional gradient-based algorithms, guarantees better performances from the point of view of both the speed of convergence and the numerical robustness.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.