Abstract
We give upper bounds on rates of approximation of real-valued functions of d Boolean variables by one-hidden-layer perceptron networks. Our bounds are of the form c/ n where c depends on certain norms of the function being approximated and n is the number of hidden units. We describe sets of functions where these norms grow either polynomially or exponentially with d.
Full Text
Sign-in/Register to access full text options
Published version (Free)
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have