Abstract
This study delves into the challenges faced by the ordinary least square (OLS) estimator, traditionally regarded as the Best Linear Unbiased Estimator in classical linear regression models. Despite its reliability under specific conditions, OLS falters in the face of multicollinearity, a problem frequently encountered in regression analyses. To combat this issue, various ridge regression estimators have been developed, characterized as one-parameter and two-parameter ridge-type estimators. In this context, our research introduces novel two-parameter estimators, building upon a recently developed one-parameter ridge estimator to mitigate the impact of multicollinearity in linear regression models. Theoretical analysis and simulation experiments were conducted to assess the performance of the proposed estimators. Remarkably, our results reveal that, under certain conditions, these new estimators outperform existing estimators, displaying a significantly reduced mean square error. To validate these findings, real-life data was employed, aligning with the outcomes derived from theoretical analysis and simulations.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.