Tracking everything everywhere all at once

Q Wang, YY Chang, R Cai, Z Li… - Proceedings of the …, 2023 - openaccess.thecvf.com
We present a new test-time optimization method for estimating dense and long-range motion
from a video sequence. Prior optical flow or particle video tracking algorithms typically …

Dynamic 3d gaussians: Tracking by persistent dynamic view synthesis

J Luiten, G Kopanas, B Leibe, D Ramanan - arXiv preprint arXiv …, 2023 - arxiv.org
We present a method that simultaneously addresses the tasks of dynamic scene novel-view
synthesis and six degree-of-freedom (6-DOF) tracking of all dense scene elements. We …

Dynamic 3d gaussians: Tracking by persistent dynamic view synthesis

J Luiten, G Kopanas, B Leibe… - … Conference on 3D …, 2024 - ieeexplore.ieee.org
We present a method that simultaneously addresses the tasks of dynamic scene novel-view
synthesis and six degree-of-freedom (6-DOF) tracking of all dense scene elements. We …

[HTML][HTML] Tracking and mapping in medical computer vision: A review

A Schmidt, O Mohareri, S DiMaio, MC Yip… - Medical Image …, 2024 - Elsevier
As computer vision algorithms increase in capability, their applications in clinical systems
will become more pervasive. These applications include: diagnostics, such as colonoscopy …

Space-time diffusion features for zero-shot text-driven motion transfer

D Yatim, R Fridman, O Bar-Tal… - Proceedings of the …, 2024 - openaccess.thecvf.com
We present a new method for text-driven motion transfer-synthesizing a video that complies
with an input text prompt describing the target objects and scene while maintaining an input …

Readout guidance: Learning control from diffusion features

G Luo, T Darrell, O Wang… - Proceedings of the …, 2024 - openaccess.thecvf.com
Abstract We present Readout Guidance a method for controlling text-to-image diffusion
models with learned signals. Readout Guidance uses readout heads lightweight networks …

Videoswap: Customized video subject swapping with interactive semantic point correspondence

Y Gu, Y Zhou, B Wu, L Yu, JW Liu… - Proceedings of the …, 2024 - openaccess.thecvf.com
Current diffusion-based video editing primarily focuses on structure-preserved editing by
utilizing various dense correspondences to ensure temporal consistency and motion …

Synctalk: The devil is in the synchronization for talking head synthesis

Z Peng, W Hu, Y Shi, X Zhu, X Zhang… - Proceedings of the …, 2024 - openaccess.thecvf.com
Achieving high synchronization in the synthesis of realistic speech-driven talking head
videos presents a significant challenge. Traditional Generative Adversarial Networks (GAN) …

Motion-i2v: Consistent and controllable image-to-video generation with explicit motion modeling

X Shi, Z Huang, FY Wang, W Bian, D Li… - ACM SIGGRAPH 2024 …, 2024 - dl.acm.org
We introduce Motion-I2V, a novel framework for consistent and controllable text-guided
image-to-video generation (I2V). In contrast to previous methods that directly learn the …

Mft: Long-term tracking of every pixel

M Neoral, J Šerých, J Matas - Proceedings of the IEEE/CVF …, 2024 - openaccess.thecvf.com
Abstract We propose MFT--Multi-Flow dense Tracker--a novel method for dense, pixel-level,
long-term tracking. The approach exploits optical flows estimated not only between …