AbstractLatent diffusion models (LDMs) have demonstrated remarkable success in generative modeling. It is promising to leverage the potential of diffusion priors to enhance performance in image and video tasks. However, applying LDMs to video super‐resolution (VSR) presents significant challenges due to the high demands for realistic details and temporal consistency in generated videos, exacerbated by the inherent stochasticity in the diffusion process. In this work, we propose a novel diffusion‐based framework, Temporal‐awareness Latent Diffusion Model (TempDiff), specifically designed for real‐world video super‐resolution, where degradations are diverse and complex. TempDiff harnesses the powerful generative prior of a pre‐trained diffusion model and enhances temporal awareness through the following mechanisms: 1) Incorporating temporal layers into the denoising U‐Net and VAE‐Decoder, and fine‐tuning these added modules to maintain temporal coherency; 2) Estimating optical flow guidance using a pre‐trained flow net for latent optimization and propagation across video sequences, ensuring overall stability in the generated high‐quality video. Extensive experiments demonstrate that TempDiff achieves compelling results, outperforming state‐of‐the‐art methods on both synthetic and real‐world VSR benchmark datasets. Code will be available at https://github.com/jiangqin567/TempDiff