Controllable Data Generation by Deep Learning: A Review

S Wang, Y Du, X Guo, B Pan, Z Qin, L Zhao - ACM Computing Surveys, 2024 - dl.acm.org
Designing and generating new data under targeted properties has been attracting various
critical applications such as molecule design, image editing and speech synthesis …

Stylegan-t: Unlocking the power of gans for fast large-scale text-to-image synthesis

A Sauer, T Karras, S Laine… - … on machine learning, 2023 - proceedings.mlr.press
Text-to-image synthesis has recently seen significant progress thanks to large pretrained
language models, large-scale training data, and the introduction of scalable model families …

Dragdiffusion: Harnessing diffusion models for interactive point-based image editing

Y Shi, C Xue, JH Liew, J Pan, H Yan… - Proceedings of the …, 2024 - openaccess.thecvf.com
Accurate and controllable image editing is a challenging task that has attracted significant
attention recently. Notably DragGAN developed by Pan et al.(2023) is an interactive point …

A survey on deep generative 3d-aware image synthesis

W Xia, JH Xue - ACM Computing Surveys, 2023 - dl.acm.org
Recent years have seen remarkable progress in deep learning powered visual content
creation. This includes deep generative 3D-aware image synthesis, which produces high …

Drag your gan: Interactive point-based manipulation on the generative image manifold

X Pan, A Tewari, T Leimkühler, L Liu, A Meka… - ACM SIGGRAPH 2023 …, 2023 - dl.acm.org
Synthesizing visual content that meets users' needs often requires flexible and precise
controllability of the pose, shape, expression, and layout of the generated objects. Existing …

Stylesdf: High-resolution 3d-consistent image and geometry generation

R Or-El, X Luo, M Shan, E Shechtman… - Proceedings of the …, 2022 - openaccess.thecvf.com
We introduce a high resolution, 3D-consistent image and shape generation technique which
we call StyleSDF. Our method is trained on single view RGB data only, and stands on the …

Diffusion autoencoders: Toward a meaningful and decodable representation

K Preechakul, N Chatthee… - Proceedings of the …, 2022 - openaccess.thecvf.com
Diffusion probabilistic models (DPMs) have achieved remarkable quality in image
generation that rivals GANs'. But unlike GANs, DPMs use a set of latent variables that lack …

Stylegan-nada: Clip-guided domain adaptation of image generators

R Gal, O Patashnik, H Maron, AH Bermano… - ACM Transactions on …, 2022 - dl.acm.org
Can a generative model be trained to produce images from a specific domain, guided only
by a text prompt, without seeing any image? In other words: can an image generator be …

Alias-free generative adversarial networks

T Karras, M Aittala, S Laine… - Advances in neural …, 2021 - proceedings.neurips.cc
We observe that despite their hierarchical convolutional nature, the synthesis process of
typical generative adversarial networks depends on absolute pixel coordinates in an …

Hyperstyle: Stylegan inversion with hypernetworks for real image editing

Y Alaluf, O Tov, R Mokady, R Gal… - Proceedings of the …, 2022 - openaccess.thecvf.com
The inversion of real images into StyleGAN's latent space is a well-studied problem.
Nevertheless, applying existing approaches to real-world scenarios remains an open …