Extracting motion and appearance via inter-frame attention for efficient video frame interpolation

G Zhang, Y Zhu, H Wang, Y Chen… - Proceedings of the …, 2023 - openaccess.thecvf.com
Effectively extracting inter-frame motion and appearance information is important for video
frame interpolation (VFI). Previous works either extract both types of information in a mixed …

Generative image dynamics

Z Li, R Tucker, N Snavely… - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
We present an approach to modeling an image-space prior on scene motion. Our prior is
learned from a collection of motion trajectories extracted from real video sequences …

Videoflow: Exploiting temporal cues for multi-frame optical flow estimation

X Shi, Z Huang, W Bian, D Li… - Proceedings of the …, 2023 - openaccess.thecvf.com
We introduce VideoFlow, a novel optical flow estimation framework for videos. In contrast to
previous methods that learn to estimate optical flow from two frames, VideoFlow concurrently …

Amt: All-pairs multi-field transforms for efficient frame interpolation

Z Li, ZL Zhu, LH Han, Q Hou… - Proceedings of the …, 2023 - openaccess.thecvf.com
Abstract We present All-Pairs Multi-Field Transforms (AMT), a new network architecture for
video frame interpolation. It is based on two essential designs. First, we build bidirectional …

Openstl: A comprehensive benchmark of spatio-temporal predictive learning

C Tan, S Li, Z Gao, W Guan, Z Wang… - Advances in …, 2023 - proceedings.neurips.cc
Spatio-temporal predictive learning is a learning paradigm that enables models to learn
spatial and temporal patterns by predicting future frames from given past frames in an …

Dynamicrafter: Animating open-domain images with video diffusion priors

J Xing, M Xia, Y Zhang, H Chen, X Wang… - arXiv preprint arXiv …, 2023 - arxiv.org
Animating a still image offers an engaging visual experience. Traditional image animation
techniques mainly focus on animating natural scenes with stochastic dynamics (eg clouds …

Mmvp: Motion-matrix-based video prediction

Y Zhong, L Liang, I Zharkov… - Proceedings of the …, 2023 - openaccess.thecvf.com
A central challenge of video prediction lies where the system has to reason the object's
future motion from image frames while simultaneously maintaining the consistency of its …

Lamp: Learn a motion pattern for few-shot-based video generation

R Wu, L Chen, T Yang, C Guo, C Li, X Zhang - arXiv preprint arXiv …, 2023 - arxiv.org
With the impressive progress in diffusion-based text-to-image generation, extending such
powerful generative ability to text-to-video raises enormous attention. Existing methods …

Vmrnn: Integrating vision mamba and lstm for efficient and accurate spatiotemporal forecasting

Y Tang, P Dong, Z Tang, X Chu… - Proceedings of the …, 2024 - openaccess.thecvf.com
Abstract Combining Convolutional Neural Networks (CNNs) or Vision Transformers (ViTs)
with Recurrent Neural Networks (RNNs) for spatiotemporal forecasting has yielded …

LAMP: Learn A Motion Pattern for Few-Shot Video Generation

R Wu, L Chen, T Yang, C Guo, C Li… - Proceedings of the …, 2024 - openaccess.thecvf.com
In this paper we present a few-shot text-to-video framework LAMP which enables a text-to-
image diffusion model to Learn A specific Motion Pattern with 8 16 videos on a single GPU …