AbstractThe use of emergent constraints to quantify uncertainty for policy-relevant quantities such as equilibrium climate sensitivity (ECS) has become increasingly widespread in recent years. Many researchers, however, claim that emergent constraints are inappropriate or even underreport uncertainty. In this paper we contribute to this discussion by examining the emergent constraints methodology in terms of its underpinning statistical assumptions. We argue that the statistical assumptions required to underpin existing frameworks are strong, hard to defend, and lead to an underreporting of uncertainty. We show how weakening them leads to a more transparent Bayesian framework wherein hitherto-ignored sources of uncertainty, such as how reality might differ from models, can be quantified. We present a guided framework for the quantification of additional uncertainties that is linked to the confidence we can have in the underpinning physical arguments for using linear constraints. We provide a software tool for implementing our framework for emergent constraints and use it to illustrate the methods on a number of recent emergent constraints for ECS. We find that the robustness of any constraint to additional uncertainties depends strongly on the confidence we have in the underpinning physics, allowing a future framing of the debate over the validity of a particular constraint around underlying physical arguments, rather than statistical assumptions. We also find that when physical arguments lead to confidence in the linear relationships underpinning emergent constraints, prediction intervals are only slightly widened by including additional uncertainties, and they show this across a range of emergent constraints for ECS.