Drag your noise: Interactive point-based editing via diffusion semantic propagation

H Liu, C Xu, Y Yang, L Zeng… - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
Point-based interactive editing serves as an essential tool to complement the controllability
of existing generative models. A concurrent work DragDiffusion updates the diffusion latent …

Where is my spot? few-shot image generation via latent subspace optimization

C Zheng, B Liu, H Zhang, X Xu… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Image generation relies on massive training data that can hardly produce diverse images of
an unseen category according to a few examples. In this paper, we address this dilemma by …

Intra-& extra-source exemplar-based style synthesis for improved domain generalization

Y Li, D Zhang, M Keuper, A Khoreva - International Journal of Computer …, 2024 - Springer
The generalization with respect to domain shifts, as they frequently appear in applications
such as autonomous driving, is one of the remaining big challenges for deep learning …

Rigid: Recurrent gan inversion and editing of real face videos

Y Xu, S He, KYK Wong, P Luo - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
GAN inversion is indispensable for applying the powerful editability of GAN to real images.
However, existing methods invert video frames individually often leading to undesired …

Learning an interpretable stylized subspace for 3d-aware animatable artforms

C Zheng, B Liu, X Xu, H Zhang… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Throughout history, static paintings have captivated viewers within display frames, yet the
possibility of making these masterpieces vividly interactive remains intriguing. This research …

ZIGNeRF: Zero-shot 3D Scene Representation with Invertible Generative Neural Radiance Fields

K Ko, M Lee - Proceedings of the IEEE/CVF Winter …, 2024 - openaccess.thecvf.com
Abstract Generative Neural Radiance Fields (NeRFs) have demonstrated remarkable
proficiency in synthesizing multi-view images by learning the distribution of a set of unposed …

Make it so: Steering stylegan for any image inversion and editing

A Bhattad, V Shah, D Hoiem, DA Forsyth - arXiv preprint arXiv:2304.14403, 2023 - arxiv.org
StyleGAN's disentangled style representation enables powerful image editing by
manipulating the latent variables, but accurately mapping real-world images to their latent …

Expanding Expressiveness of Diffusion Models with Limited Data via Self-Distillation based Fine-Tuning

J Hur, J Choi, G Han, DJ Lee… - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
Training diffusion models on limited datasets poses challenges in terms of limited
generation capacity and expressiveness, leading to unsatisfactory results in various …

Out-of-domain GAN inversion via Invertibility Decomposition for Photo-Realistic Human Face Manipulation

X Yang, X Xu, Y Chen - Proceedings of the IEEE/CVF …, 2023 - openaccess.thecvf.com
Abstract The fidelity of Generative Adversarial Networks (GAN) inversion is impeded by Out-
Of-Domain (OOD) areas (eg, background, accessories) in the image. Detecting the OOD …

-VTON: Dynamic Semantics Disentangling for Differential Diffusion Based Virtual Try-On

Z Yang, Z Jiang, X Li, H Zhou, J Dong, H Zhang… - … on Computer Vision, 2025 - Springer
In this paper, we introduce\(\textrm {D}^ 4\)-VTON, an innovative solution for image-based
virtual try-on. We address challenges from previous studies, such as semantic …