Abstract
This paper introduces a new formal test of the significance of neural network inputs. It is simple, accurate, and powerful and is based on a linear relationship between the output of a neural network when all of the input variables are fixed at their mean values—other than the input variable, which is subject to significance testing—and the target values of the network. Simulation results show that as the number of observations increases, the power of the test tends to 1 in all cases, and that the empirical size approaches the nominal size in some cases. The results, based on the ordinary least squares (OLS) estimation of parameters, are very encouraging, but using a heteroscedasticity and autocorrelation consistent covariance matrix and fast double bootstrap improves the speed of the convergence to the nominal size. The test can also be used for nonlinear models with nuisance parameters.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.