Abstract
This paper compares design-based properties of the classical two-phase regression estimator with several nonparametric kernel-based estimators of which k nearest neighbors (kNN) is a special case. Metrics are based on the Euclidean distance applied to either a multidimensional space of explanatory variables or to a one-dimensional space of predictions obtained from a linear model. The main concepts of kernel-based regression estimators are reformulated in the design-based Monte Carlo approach to forest inventory. The results, based on a case study of a forest inventory in Switzerland and extensive simulations, suggest that the commonly used analytical external variance formula may systematically underestimate the true variance for a variety of kernel-based estimators including kNN but that it is still adequate for the classical regression estimator. Although using a bootstrap variance can help to correct this underestimation, it was also found that the bootstrap variance estimates could be unstable if the optimal bandwidth is recalculated in each bootstrap sample. These findings suggest that if the model captures the main features of the underlying process, then it is advisable to use the classical regression estimator, because it performs at least as well as the other techniques and is by far simpler to implement.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.