Paint by example: Exemplar-based image editing with diffusion models

B Yang, S Gu, B Zhang, T Zhang… - Proceedings of the …, 2023 - openaccess.thecvf.com
Abstract Language-guided image editing has achieved great success recently. In this paper,
we investigate exemplar-guided image editing for more precise control. We achieve this …

Vqgan-clip: Open domain image generation and editing with natural language guidance

K Crowson, S Biderman, D Kornis, D Stander… - … on Computer Vision, 2022 - Springer
Generating and editing images from open domain text prompts is a challenging task that
heretofore has required expensive and specially trained models. We demonstrate a novel …

Ilvr: Conditioning method for denoising diffusion probabilistic models

J Choi, S Kim, Y Jeong, Y Gwon, S Yoon - arXiv preprint arXiv:2108.02938, 2021 - arxiv.org
Denoising diffusion probabilistic models (DDPM) have shown remarkable performance in
unconditional image generation. However, due to the stochasticity of the generative process …

[PDF][PDF] Deep vit features as dense visual descriptors

S Amir, Y Gandelsman, S Bagon… - arXiv preprint arXiv …, 2021 - dino-vit-features.github.io
We study the use of deep features extracted from a pretrained Vision Transformer (ViT) as
dense visual descriptors. We observe and empirically demonstrate that such features, when …

Image-to-image translation: Methods and applications

Y Pang, J Lin, T Qin, Z Chen - IEEE Transactions on Multimedia, 2021 - ieeexplore.ieee.org
Image-to-image translation (I2I) aims to transfer images from a source domain to a target
domain while preserving the content representations. I2I has drawn increasing attention and …

Exploiting spatial dimensions of latent in gan for real-time image editing

H Kim, Y Choi, J Kim, S Yoo… - Proceedings of the IEEE …, 2021 - openaccess.thecvf.com
Generative adversarial networks (GANs) synthesize realistic images from random latent
vectors. Although manipulating the latent vectors controls the synthesized outputs, editing …

Explaining in style: Training a gan to explain a classifier in stylespace

O Lang, Y Gandelsman, M Yarom… - Proceedings of the …, 2021 - openaccess.thecvf.com
Image classification models can depend on multiple different semantic attributes of the
image. An explanation of the decision of the classifier needs to both discover and visualize …

Pose with style: Detail-preserving pose-guided image synthesis with conditional stylegan

B Albahar, J Lu, J Yang, Z Shu, E Shechtman… - ACM Transactions on …, 2021 - dl.acm.org
We present an algorithm for re-rendering a person from a single image under arbitrary
poses. Existing methods often have difficulties in hallucinating occluded contents photo …

Facial-sketch synthesis: A new challenge

DP Fan, Z Huang, P Zheng, H Liu, X Qin… - Machine Intelligence …, 2022 - Springer
This paper aims to conduct a comprehensive study on facial-sketch synthesis (FSS).
However, due to the high cost of obtaining hand-drawn sketch datasets, there is a lack of a …

Ensembling off-the-shelf models for gan training

N Kumari, R Zhang, E Shechtman… - Proceedings of the …, 2022 - openaccess.thecvf.com
The advent of large-scale training has produced a cornucopia of powerful visual recognition
models. However, generative models, such as GANs, have traditionally been trained from …