Spvit: Enabling faster vision transformers via latency-aware soft token pruning

Z Kong, P Dong, X Ma, X Meng, W Niu, M Sun… - European conference on …, 2022 - Springer
Abstract Recently, Vision Transformer (ViT) has continuously established new milestones in
the computer vision field, while the high computation and memory cost makes its …

Sparcl: Sparse continual learning on the edge

Z Wang, Z Zhan, Y Gong, G Yuan… - Advances in …, 2022 - proceedings.neurips.cc
Existing work in continual learning (CL) focuses on mitigating catastrophic forgetting, ie,
model performance deterioration on past tasks when learning a new task. However, the …

Chex: Channel exploration for cnn model compression

Z Hou, M Qin, F Sun, X Ma, K Yuan… - Proceedings of the …, 2022 - openaccess.thecvf.com
Channel pruning has been broadly recognized as an effective technique to reduce the
computation and memory cost of deep convolutional neural networks. However …

Compute-efficient deep learning: Algorithmic trends and opportunities

BR Bartoldson, B Kailkhura, D Blalock - Journal of Machine Learning …, 2023 - jmlr.org
Although deep learning has made great progress in recent years, the exploding economic
and environmental costs of training neural networks are becoming unsustainable. To …

Fedcv: a federated learning framework for diverse computer vision tasks

C He, AD Shah, Z Tang, DFAN Sivashunmugam… - arXiv preprint arXiv …, 2021 - arxiv.org
Federated Learning (FL) is a distributed learning paradigm that can learn a global or
personalized model from decentralized datasets on edge devices. However, in the computer …

Dynamic sparse no training: Training-free fine-tuning for sparse llms

Y Zhang, L Zhao, M Lin, Y Sun, Y Yao, X Han… - arXiv preprint arXiv …, 2023 - arxiv.org
The ever-increasing large language models (LLMs), though opening a potential path for the
upcoming artificial general intelligence, sadly drops a daunting obstacle on the way towards …

Dynamic sparse network for time series classification: Learning what to “see”

Q Xiao, B Wu, Y Zhang, S Liu… - Advances in …, 2022 - proceedings.neurips.cc
The receptive field (RF), which determines the region of time series to be “seen” and used, is
critical to improve the performance for time series classification (TSC). However, the …

Why random pruning is all we need to start sparse

AH Gadhikar, S Mukherjee… - … Conference on Machine …, 2023 - proceedings.mlr.press
Random masks define surprisingly effective sparse neural network models, as has been
shown empirically. The resulting sparse networks can often compete with dense …

Towards sparsification of graph neural networks

H Peng, D Gurevin, S Huang, T Geng… - 2022 IEEE 40th …, 2022 - ieeexplore.ieee.org
As real-world graphs expand in size, larger GNN models with billions of parameters are
deployed. High parameter count in such models makes training and inference on graphs …

Fededge: Accelerating edge-assisted federated learning

K Wang, Q He, F Chen, H Jin, Y Yang - Proceedings of the ACM Web …, 2023 - dl.acm.org
Federated learning (FL) has been widely acknowledged as a promising solution to training
machine learning (ML) model training with privacy preservation. To reduce the traffic …