Accelerating sparse deep neural networks

A Mishra, JA Latorre, J Pool, D Stosic, D Stosic… - arXiv preprint arXiv …, 2021 - arxiv.org
As neural network model sizes have dramatically increased, so has the interest in various
techniques to reduce their parameter counts and accelerate their execution. An active area …

Boost vision transformer with gpu-friendly sparsity and quantization

C Yu, T Chen, Z Gan, J Fan - Proceedings of the IEEE/CVF …, 2023 - openaccess.thecvf.com
The transformer extends its success from the language to the vision domain. Because of the
numerous stacked self-attention and cross-attention blocks in the transformer, which involve …

Re-GAN: Data-efficient GANs training via architectural reconfiguration

D Saxena, J Cao, J Xu… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Abstract Training Generative Adversarial Networks (GANs) on high-fidelity images usually
requires a vast number of training images. Recent research on GAN tickets reveals that …

Revisiting discriminator in gan compression: A generator-discriminator cooperative compression scheme

S Li, J Wu, X Xiao, F Chao… - Advances in Neural …, 2021 - proceedings.neurips.cc
Recently, a series of algorithms have been explored for GAN compression, which aims to
reduce tremendous computational overhead and memory usages when deploying GANs on …

Balanced training for sparse gans

Y Wang, J Wu, N Hovakimyan… - Advances in Neural …, 2024 - proceedings.neurips.cc
Over the past few years, there has been growing interest in developing larger and deeper
neural networks, including deep generative models like generative adversarial networks …

Compressing image-to-image translation gans using local density structures on their learned manifold

A Ganjdanesh, S Gao, H Alipanah… - Proceedings of the AAAI …, 2024 - ojs.aaai.org
Generative Adversarial Networks (GANs) have shown remarkable success in modeling
complex data distributions for image-to-image translation. Still, their high computational …

Transferring Core Knowledge via Learngenes

F Feng, J Wang, X Geng - arXiv preprint arXiv:2401.08139, 2024 - arxiv.org
The pre-training paradigm fine-tunes the models trained on large-scale datasets to
downstream tasks with enhanced performance. It transfers all knowledge to downstream …

Cut inner layers: A structured pruning strategy for efficient u-net gans

BK Kim, S Choi, H Park - arXiv preprint arXiv:2206.14658, 2022 - arxiv.org
Pruning effectively compresses overparameterized models. Despite the success of pruning
methods for discriminative models, applying them for generative models has been relatively …

Towards Device Efficient Conditional Image Generation

NA Shah, G Bharaj - arXiv preprint arXiv:2203.10363, 2022 - arxiv.org
We present a novel algorithm to reduce tensor compute required by a conditional image
generation autoencoder without sacrificing quality of photo-realistic image generation. Our …

Lightweight Model for Occlusion Removal from Face Images

S John, A Danti - Annals of Emerging Technologies in Computing …, 2024 - aetic.theiaer.org
In the realm of deep learning, the prevalence of models with large number of parameters
poses a significant challenge for low computation device. Critical influence of model size …