Abstract

Machine learning (ML) methods, such as artificial neural networks (ANN), k-nearest neighbors (kNN), random forests (RF), support vector machines (SVM), and boosted decision trees (DTs), may offer stronger predictive performance than more traditional, parametric methods, such as linear regression, multiple linear regression, and logistic regression (LR), for specific mapping and modeling tasks. However, this increased performance is often accompanied by increased model complexity and decreased interpretability, resulting in critiques of their “black box” nature, which highlights the need for algorithms that can offer both strong predictive performance and interpretability. This is especially true when the global model and predictions for specific data points need to be explainable in order for the model to be of use. Explainable boosting machines (EBM), an augmentation and refinement of generalize additive models (GAMs), has been proposed as an empirical modeling method that offers both interpretable results and strong predictive performance. The trained model can be graphically summarized as a set of functions relating each predictor variable to the dependent variable along with heat maps representing interactions between selected pairs of predictor variables. In this study, we assess EBMs for predicting the likelihood or probability of slope failure occurrence based on digital terrain characteristics in four separate Major Land Resource Areas (MLRAs) in the state of West Virginia, USA and compare the results to those obtained with LR, kNN, RF, and SVM. EBM provided predictive accuracies comparable to RF and SVM and better than LR and kNN. The generated functions and visualizations for each predictor variable and included interactions between pairs of predictor variables, estimation of variable importance based on average mean absolute scores, and provided scores for each predictor variable for new predictions add interpretability, but additional work is needed to quantify how these outputs may be impacted by variable correlation, inclusion of interaction terms, and large feature spaces. Further exploration of EBM is merited for geohazard mapping and modeling in particular and spatial predictive mapping and modeling in general, especially when the value or use of the resulting predictions would be greatly enhanced by improved interpretability globally and availability of prediction explanations at each cell or aggregating unit within the mapped or modeled extent.

Highlights

  • We explore the use of explainable boosting machine (EBM) [6,7] for predicting the probability of slope failure occurrence based on topographic predictor variables calculated from a light detection and ranging (LiDAR)-derived digital terrain model (DTM)

  • KNN algorithms showed the weakest performance while the Explainable boosting machines (EBM), random forests (RF), and support vector machines (SVM)

  • Across the four Major Land Resource Areas (MLRAs) study areas, EBM outperformed logistic regression (LR) and k-nearest neighbors (kNN) and performed comparably to RF and SVM for predicting slope failure occurrence based on multiple metrics (i.e., overall accuracy (OA), precision, recall, F1 score, specificity, negative predictive value (NPV), area under the receiver operating characteristic curve (AUC receiver operating characteristic (ROC)), and AUC PR)

Read more

Summary

Introduction

2021, 13, 4991 is generally attributed to the ability of ML algorithms to characterize patterns in noisy, large, and/or complex datasets and feature spaces without having to make distribution assumptions that are often violated [2] This increased predictive power is generally accompanied by increased model complexity and reduced interpretability, leading practitioners and researchers to critique the “black box” nature of these methods and call for the use of more interpretable or “glass box” models. This is especially true when there is an interest in or need to explain what factors contribute most to the prediction and how the response variable is impacted by specific predictor variables. P represents the probability of the sample belonging to the positive class, which is assigned a value of 1 while the negative class is assigned a value of

Methods
Results
Discussion
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call