A survey on efficient convolutional neural networks and hardware acceleration

D Ghimire, D Kil, S Kim - Electronics, 2022 - mdpi.com
Over the past decade, deep-learning-based representations have demonstrated remarkable
performance in academia and industry. The learning capability of convolutional neural …

Model compression and hardware acceleration for neural networks: A comprehensive survey

L Deng, G Li, S Han, L Shi, Y Xie - Proceedings of the IEEE, 2020 - ieeexplore.ieee.org
Domain-specific hardware is becoming a promising topic in the backdrop of improvement
slow down for general-purpose processors due to the foreseeable end of Moore's Law …

Chip: Channel independence-based pruning for compact neural networks

Y Sui, M Yin, Y Xie, H Phan… - Advances in Neural …, 2021 - proceedings.neurips.cc
Filter pruning has been widely used for neural network compression because of its enabled
practical acceleration. To date, most of the existing filter pruning works explore the …

Fact: Factor-tuning for lightweight adaptation on vision transformer

S Jie, ZH Deng - Proceedings of the AAAI conference on artificial …, 2023 - ojs.aaai.org
Recent work has explored the potential to adapt a pre-trained vision transformer (ViT) by
updating only a few parameters so as to improve storage efficiency, called parameter …

Towards efficient tensor decomposition-based dnn model compression with optimization framework

M Yin, Y Sui, S Liao, B Yuan - Proceedings of the IEEE/CVF …, 2021 - openaccess.thecvf.com
Advanced tensor decomposition, such as Tensor train (TT) and Tensor ring (TR), has been
widely studied for deep neural network (DNN) model compression, especially for recurrent …

A tensorized transformer for language modeling

X Ma, P Zhang, S Zhang, N Duan… - Advances in neural …, 2019 - proceedings.neurips.cc
Latest development of neural models has connected the encoder and decoder through a
self-attention mechanism. In particular, Transformer, which is solely based on self-attention …

Convolutional tensor-train LSTM for spatio-temporal learning

J Su, W Byeon, J Kossaifi, F Huang… - Advances in …, 2020 - proceedings.neurips.cc
Learning from spatio-temporal data has numerous applications such as human-behavior
analysis, object tracking, video compression, and physics simulation. However, existing …

Wide compression: Tensor ring nets

W Wang, Y Sun, B Eriksson… - Proceedings of the …, 2018 - openaccess.thecvf.com
Deep neural networks have demonstrated state-of-the-art performance in a variety of real-
world applications. In order to obtain performance gains, these networks have grown larger …

Next point-of-interest recommendation on resource-constrained mobile devices

Q Wang, H Yin, T Chen, Z Huang, H Wang… - Proceedings of the Web …, 2020 - dl.acm.org
In the modern tourism industry, next point-of-interest (POI) recommendation is an important
mobile service as it effectively aids hesitating travelers to decide the next POI to visit …

Qtt-dlstm: A cloud-edge-aided distributed lstm for cyber–physical–social big data

X Wang, L Ren, R Yuan, LT Yang… - IEEE Transactions on …, 2022 - ieeexplore.ieee.org
Cyber–physical–social systems (CPSS), an emerging cross-disciplinary research area,
combines cyber–physical systems (CPS) with social networking for the purpose of providing …