Abstract

Optical flows and video frame interpolation are considered as a chicken-egg problem such that one problem affects the other and vice versa. This paper presents a stack of deep networks to estimate intermediate optical flows from the very first intermediate synthesized frame and later generate the very end interpolated frame by combining the very first one and two learned intermediate optical flows based warped frames. The primary benefit is that it glues two problems into a single comprehensive framework that learns altogether by using both an analysis-by-synthesis technique for optical flow estimation and Convolutional Neural Networks (CNN) kernels-based frame synthesis. The proposed network is the first attempt to merge two previous branches of previous approaches, optical flow-based synthesis and CNN kernels-based synthesis into a comprehensive network. Experiments are carried out with various challenging datasets, all showing that the proposed network outperforms the state-of-the-art methods with significant margins for video frame interpolation and the estimated optical flows are more accurate for challenging movements. Furthermore, the proposed Motion Estimation Motion Compensation (MEMC) network shows its outstanding enhancement of the quality of compressed videos.

Highlights

  • Video frame interpolation is widely used in various applications from computer vision to visual display applications such as frame rate up conversion (FRUC), slow motion display and animation

  • This paper presents a comprehensive framework that glues two of the above previous approaches into a single stacked network such that an analysis-by-synthesis technique is used to estimate bidirectional intermediate optical flows and later a synthesis network glues intermediate results generated by component branches to synthesize the very end intermediate frame

  • ANALYSIS BASED SYNTHESIS INTERMEDIATE OPTICAL FLOW ESTIMATIONS In the first layer of the stack, the motion derivation module is the glue between two branches of approaches, the optical flow-based frame interpolation and the Convolutional Neural Networks (CNN) kernels-based frame synthesis

Read more

Summary

Introduction

Video frame interpolation is widely used in various applications from computer vision to visual display applications such as frame rate up conversion (FRUC), slow motion display and animation. The proposed network is a combination of two branches of approaches: optical flow-based frame interpolation and CNN kernels-based frame synthesis.

Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call