Spvit: Enabling faster vision transformers via latency-aware soft token pruning

Z Kong, P Dong, X Ma, X Meng, W Niu, M Sun… - European conference on …, 2022 - Springer
Abstract Recently, Vision Transformer (ViT) has continuously established new milestones in
the computer vision field, while the high computation and memory cost makes its …

Quantized transformer language model implementations on edge devices

MWU Rahman, MM Abrar, HG Copening… - 2023 International …, 2023 - ieeexplore.ieee.org
Large-scale transformer-based models like the Bidi-rectional Encoder Representations from
Transformers (BERT) are widely used for Natural Language Processing (NLP) applications …

Optimizing Large Language Models for Edge Devices: A Comparative Study on Reputation Analysis

MWU Rahman - 2023 - search.proquest.com
The widespread adoption of social media platforms has led to an exponential surge in user-
generated data, shaping the reputations of companies and public figures on a global scale …