Abstract
Based on the approximation theory of Fourier-series working in square integrable space, a Fourier neuronal network was constructed by using activation functions of the complex exponential form. Then a weights-direct-determination method was derived to decide the neural-network weights immediately, which remedied the weaknesses of conventional BP neural networks such as small convergence rate, easily converging to local minimum and possibly lengthy or oscillatory learning process. A hidden-neurons-growing algorithm was presented to adjust the neural-network structure adaptively. Theoretical analysis and simulation results substantiate further that the presented Fourier neural network and algorithm could have good properties of high-precision learning, noise-suppressing and discontinuous-function approximating.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.