Abstract
Abstract In cases of modest correlation, parameters calculated from a standard least squares linear regression can vary depending on the selection of dependent and independent variates. A neutral regression that addresses this problem is proposed. The eigenvector corresponding to the smallest eigenvalue of the cross-correlation matrix of the two variates is used as a set of regression coefficients. Error bars are calculated for the eigenvalues and eigenvectors by means of a perturbation expansion of the cross-correlation matrix and are then verified by Monte Carlo simulation. A procedure is suggested for extension of the technique to the multivariate case. Examples of a linear fit for low-correlation and a quadratic fit for high-correlation cases are given. Conclusions are presented regarding the strengths and weaknesses of both the least squares and the neutral regression.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.