Instructpix2pix: Learning to follow image editing instructions

T Brooks, A Holynski, AA Efros - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
We propose a method for editing images from human instructions: given an input image and
a written instruction that tells the model what to do, our model follows these instructions to …

Emergent correspondence from image diffusion

L Tang, M Jia, Q Wang, CP Phoo… - Advances in Neural …, 2023 - proceedings.neurips.cc
Finding correspondences between images is a fundamental problem in computer vision. In
this paper, we show that correspondence emerges in image diffusion models without any …

Synthetic data from diffusion models improves imagenet classification

S Azizi, S Kornblith, C Saharia, M Norouzi… - arXiv preprint arXiv …, 2023 - arxiv.org
Deep generative models are becoming increasingly powerful, now generating diverse high
fidelity photo-realistic samples given text prompts. Have they reached the point where …

Fake it till you make it: Learning transferable representations from synthetic imagenet clones

MB Sarıyıldız, K Alahari, D Larlus… - Proceedings of the …, 2023 - openaccess.thecvf.com
Recent image generation models such as Stable Diffusion have exhibited an impressive
ability to generate fairly realistic images starting from a simple text prompt. Could such …

Survey on leveraging pre-trained generative adversarial networks for image editing and restoration

M Liu, Y Wei, X Wu, W Zuo, L Zhang - Science China Information Sciences, 2023 - Springer
Generative adversarial networks (GANs) have drawn enormous attention due to their simple
yet effective training mechanism and superior image generation quality. With the ability to …

A tale of two features: Stable diffusion complements dino for zero-shot semantic correspondence

J Zhang, C Herrmann, J Hur… - Advances in …, 2024 - proceedings.neurips.cc
Text-to-image diffusion models have made significant advances in generating and editing
high-quality images. As a result, numerous approaches have explored the ability of diffusion …

Diffusion hyperfeatures: Searching through time and space for semantic correspondence

G Luo, L Dunlap, DH Park… - Advances in Neural …, 2024 - proceedings.neurips.cc
Diffusion models have been shown to be capable of generating high-quality images,
suggesting that they could contain meaningful internal representations. Unfortunately, the …

Diversify your vision datasets with automatic diffusion-based augmentation

L Dunlap, A Umino, H Zhang, J Yang… - Advances in neural …, 2023 - proceedings.neurips.cc
Many fine-grained classification tasks, like rare animal identification, have limited training
data and consequently classifiers trained on these datasets often fail to generalize to …

Dreamsim: Learning new dimensions of human visual similarity using synthetic data

S Fu, N Tamir, S Sundaram, L Chai, R Zhang… - arXiv preprint arXiv …, 2023 - arxiv.org
Current perceptual similarity metrics operate at the level of pixels and patches. These
metrics compare images in terms of their low-level colors and textures, but fail to capture mid …

State‐of‐the‐Art in the Architecture, Methods and Applications of StyleGAN

AH Bermano, R Gal, Y Alaluf, R Mokady… - Computer Graphics …, 2022 - Wiley Online Library
Abstract Generative Adversarial Networks (GANs) have established themselves as a
prevalent approach to image synthesis. Of these, StyleGAN offers a fascinating case study …