Abstract Let ( x i , y i ) i = 1 , … , n denote independent samples from a general mixture distribution ∑ c ∈ C ρ c P c x , and consider the hypothesis class of generalized linear models y ^ = F ( Θ ⊤ x ) . In this study, we investigate the asymptotic joint statistics of a family of generalized linear estimators ( Θ 1 , … , Θ M ) obtained either from (a) minimizing an empirical risk R ^ n ( Θ ; X , y ) or (b) sampling from the associated Gibbs measure exp ( − β n R ^ n ( Θ ; X , y ) ) . Our main contribution is to characterize under which conditions the asymptotic joint statistics of this family depends (in a weak sense) only on the means and covariances of the class conditional features distribution P c x . In particular, this allows us to prove the universality of different quantities of interest, such as the training and generalization errors, redeeming a recent line of work in high-dimensional statistics working under the Gaussian mixture hypothesis. Finally, we discuss the applications of our results in different machine learning tasks of interest, such as ensembling and uncertainty quantification.