Abstract

Abstract Models of natural processes necessarily sacrifice some realism for the sake of tractability. Detailed, parameter‐rich models often provide accurate estimates of system behaviour but can be data‐hungry and difficult to operationalize. Moreover, complexity increases the danger of ‘over‐fitting’, which leads to poor performance when models are applied to novel conditions. This challenge is typically described in terms of a trade‐off between bias and variance (i.e. low accuracy vs. low precision). In studies of ecological communities, this trade‐off often leads to an argument about the level of detail needed to describe interactions among species. Here, we used data from a grassland biodiversity experiment containing nine locally abundant plant species (the Jena ‘dominance experiment’) to parameterize models representing six increasingly complex hypotheses about interactions. For each model, we calculated goodness‐of‐fit across different subsets of the data based on sown species richness levels, and tested how performance changed depending on whether or not the same data were used to parameterize and test the model (i.e. within vs. out‐of‐sample), and whether the range of diversity treatments being predicted fell inside or outside of the range used for parameterization. As expected, goodness‐of‐fit improved as a function of model complexity for all within‐sample tests. In contrast, the best out‐of‐sample performance generally resulted from models of intermediate complexity (i.e. with only two interaction coefficients per species—an intraspecific effect and a single pooled interspecific effect), especially for predictions that fell outside the range of diversity treatments used for parameterization. In accordance with other studies, our results also demonstrate that commonly used selection methods based on AIC of models fitted to the full dataset correspond more closely to within‐sample than out‐of‐sample performance. Synthesis. Our results demonstrate that models which include only general intra and interspecific interaction coefficients can be sufficient for estimating species‐level abundances across a wide range of contexts and may provide better out‐of‐sample performance than do more complex models. These findings serve as a reminder that simpler models may often provide a better trade‐off between bias and variance in ecological systems, particularly when applying models beyond the conditions used to parameterize them.

Highlights

  • | INTRODUCTION“What a useful thing a pocket-map is!" I remarked. "That's another thing we've learned from your Nation," said Mein Herr, "map-making

  • We calculated goodness-of-fit across different subsets of the data based on sown species richness levels, and tested how performance changed depending on whether or not the same data were used to parameterize and test the model, and whether the range of diversity treatments being predicted fell inside or outside of the range used for parameterization

  • Because we are interested in community dynamics, we focus on predictions of species-level above-ground biomass rather than total biomass summed across species, which is the subject of several other studies of species interactions in biodiversity experiments (Connolly et al, 2011; Kirwan et al, 2009)

Read more

Summary

| INTRODUCTION

“What a useful thing a pocket-map is!" I remarked. "That's another thing we've learned from your Nation," said Mein Herr, "map-making. When different subsets of data are used to parameterize the model versus to test model performance, increased complexity can lead to poorer performance for the testing subset This phenomenon is known as ‘over-fitting’, and occurs when parameter tuning during the fitting process draws predictions towards these peculiarities (i.e., bias) (Wenger & Olden, 2012). Note that AIC is conceptually and theoretically related to cross-validation at the limit where a single observation is retained for testing, and the remainder of the data are used to parameterize a model (‘leave-oneout cross-validation’) (Stone, 1977) These methods are very efficient from the perspective of data requirements, but are often less useful for assessing model transferability (Brewer et al, 2016; Wenger & Olden, 2012). We expected that AIC would not necessarily be an effective indicator of out-of-sample model performance, due to its strong correspondence to within-sample estimates

| MATERIALS AND METHODS
Findings
| DISCUSSION
| CONCLUSIONS
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call