Abstract

Abstract Wildfire risk is greatest during high winds after sustained periods of dry and hot conditions. This paper is a statistical extreme-event risk attribution study that aims to answer whether extreme wildfire seasons are more likely now than under past climate. This requires modeling temporal dependence at extreme levels. We propose the use of transformed-linear time series models, which are constructed similarly to traditional autoregressive–moving-average (ARMA) models while having a dependence structure that is tied to a widely used framework for extremes (regular variation). We fit the models to the extreme values of the seasonally adjusted fire weather index (FWI) time series to capture the dependence in the upper tail for past and present climate. We simulate 10 000 fire seasons from each fitted model and compare the proportion of simulated high-risk fire seasons to quantify the increase in risk. Our method suggests that the risk of experiencing an extreme wildfire season in Grand Lake, Colorado, under current climate has increased dramatically relative to the risk under the climate of the mid-twentieth century. Our method also finds some evidence of increased risk of extreme wildfire seasons in Quincy, California, but large uncertainties do not allow us to reject a null hypothesis of no change.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call