Abstract

Generative models have emerged as potential tools for creating high-quality images, videos, and text. This paper explores the application of generative models in automating vlog content creation. It addresses both static and dynamic visual elements, eliminating the need for human intervention. Traditional vlogs often require specific environmental conditions and proper lighting for the vlog creation. To streamline this process, an automated system utilizing the generative models is proposed here. Generative models excel at generating realistic content that seamlessly integrates with real-world content. They enhance overall video quality and introduce creative elements by generating new scenes and backgrounds. This paper categorizes various generative modeling techniques based on frame elements and foreground-background conditions. It offers a comparative analysis of different generative model variants tailored for specific objectives. Furthermore, the paper reviews existing research on generative models for video and image content generation, visual quality enhancement, diversity, and coherence outcomes. Additionally, the paper highlights practical uses of the generative model for content creation in various contexts, such as face swapping, scene translation, and virtual content insertion. The paper also examines the public datasets used to train generative models. These datasets contain diverse visual content such as celebrity images, urban landscapes, and everyday scenes.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call