Abstract
Deep-learning-based convection schemes have garnered significant attention for their notable improvements in simulating precipitation distribution and tropical convection in Earth system models. However, these schemes struggle to capture the stochastic nature of moist physics, which can degrade the simulation of large-scale circulations, climate means, and variability. To address this issue, a stochastic parameterization scheme called DIFF-MP, based on a probabilistic diffusion model, is developed. Cloud-resolving data are coarse-grained into resolved-scale variables and subgrid contributions, which serve as conditional inputs and outputs for DIFF-MP. The performance of DIFF-MP is compared with that of generative adversarial networks and variational autoencoders. The results demonstrate that DIFF-MP consistently outperforms these models in terms of prediction error, coverage ratio, and spread–skill correlation. Furthermore, the standard deviation, skewness, and kurtosis of the subgrid contributions generated by DIFF-MP more closely match the test data than those produced by the other models. Interpretability experiments confirm that DIFF-MP’s parameterization of moist physics is physically consistent.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.