GY Park, H Jeong, SW Lee, JC Ye - arXiv preprint arXiv:2403.15249, 2024 - arxiv.org
The evolution of diffusion models has greatly impacted video generation and understanding. Particularly, text-to-video diffusion models (VDMs) have significantly facilitated the …
We introduce Emu Video Edit (EVE), a model that establishes a new state-of-the art in video editing without relying on any supervised video editing data. To develop EVE we separately …
X Fan, A Bhattad, R Krishna - arXiv preprint arXiv:2403.14617, 2024 - arxiv.org
We introduce Videoshop, a training-free video editing algorithm for localized semantic edits. Videoshop allows users to use any editing software, including Photoshop and generative …
Z Xiao, Y Zhou, S Yang, X Pan - arXiv preprint arXiv:2405.14864, 2024 - arxiv.org
Video generation primarily aims to model authentic and customized motion across frames, making understanding and controlling the motion a crucial topic. Most diffusion-based …
The remarkable generative capabilities of diffusion models have motivated extensive research in both image and video editing. Compared to video editing which faces additional …
Z Xue, M Luo, C Chen, K Grauman - arXiv preprint arXiv:2406.07754, 2024 - arxiv.org
We study the problem of precisely swapping objects in videos, with a focus on those interacted with by hands, given one user-provided reference object image. Despite the great …
Diffusion Models are popular generative modeling methods in various vision tasks, attracting significant attention. They can be considered a unique instance of self-supervised learning …
Despite significant advancements in video generation and editing using diffusion models, achieving accurate and localized video editing remains a substantial challenge …
Existing diffusion-based video editing methods have achieved impressive results in motion editing. Most of the existing methods focus on the motion alignment between the edited …