Three-dimensional (3D) light-field display technology can reconstruct the spatial characteristics of 3D scenes and provide users with an immersive visual experience without the need for any additional external devices. Here, an arbitrary stylized light-field generation method for 3D light-field displays is presented, and the tripartite constraints are addressed by conducting style transformation in a refined feature space. A multi-dimensional feature refinement module is designed to learn which aspects and regions should be prioritized within the high-level feature grid of the scene, which allows content and style patterns to be better matched and blended. To preserve more structural details and reduce artifacts, a novel global quality and local structure joint loss function is used for optimization. A mask-guided light-field coding method based on ray-casting is employed to accelerate the synthesis of stylized light-field images. Experimental results demonstrate that the proposed method can generate higher-quality stylized 3D light-field content with any given style in a zero-shot manner. Additionally, the method provides two user control extensions to further enrich the ways for the personalized editing of 3D content displayed on 3D light-field displays.
Read full abstract