A review of convolutional neural network architectures and their optimizations

S Cong, Y Zhou - Artificial Intelligence Review, 2023 - Springer
The research advances concerning the typical architectures of convolutional neural
networks (CNNs) as well as their optimizations are analyzed and elaborated in detail in this …

Mest: Accurate and fast memory-economic sparse training framework on the edge

G Yuan, X Ma, W Niu, Z Li, Z Kong… - Advances in …, 2021 - proceedings.neurips.cc
Recently, a new trend of exploring sparsity for accelerating neural network training has
emerged, embracing the paradigm of training on the edge. This paper proposes a novel …

Drew: Efficient winograd cnn inference with deep reuse

R Wu, F Zhang, J Guan, Z Zheng, X Du… - Proceedings of the ACM …, 2022 - dl.acm.org
Deep learning has been used in various domains, including Web services. Convolutional
neural networks (CNNs), which are deep learning representatives, are among the most …

TinyEmergencyNet: a hardware-friendly ultra-lightweight deep learning model for aerial scene image classification

OM Mogaka, R Zewail, K Inoue, MS Sayed - Journal of Real-Time Image …, 2024 - Springer
In the context of emergency response applications, real-time situational awareness is vital.
Unmanned aerial vehicles (UAVs) with imagers have emerged as crucial tools for providing …

η-lstm: Co-designing highly-efficient large lstm training via exploiting memory-saving and architectural design opportunities

X Zhang, H Xia, D Zhuang, H Sun, X Fu… - 2021 ACM/IEEE 48th …, 2021 - ieeexplore.ieee.org
Recently, the recurrent neural network, or its most popular type—the Long Short Term
Memory (LSTM) network—has achieved great success in a broad spectrum of real-world …

Weight update skipping: Reducing training time for artificial neural networks

P Safayenikoo, I Akturk - … on Emerging and Selected Topics in …, 2021 - ieeexplore.ieee.org
Artificial Neural Networks (ANNs) are known as state-of-the-art techniques in Machine
Learning (ML) and have achieved outstanding results in data-intensive applications, such as …

Layer freezing & data sieving: missing pieces of a generic framework for sparse training

G Yuan, Y Li, S Li, Z Kong, S Tulyakov… - Advances in …, 2022 - proceedings.neurips.cc
Recently, sparse training has emerged as a promising paradigm for efficient deep learning
on edge devices. The current research mainly devotes the efforts to reducing training costs …

Tdc: Towards extremely efficient cnns on gpus via hardware-aware tucker decomposition

L Xiang, M Yin, C Zhang, A Sukumaran-Rajam… - Proceedings of the 28th …, 2023 - dl.acm.org
Tucker decomposition is one of the SOTA CNN model compression techniques. However,
unlike the FLOPs reduction, we observe very limited inference time reduction with Tucker …

[HTML][HTML] Predicting wind farm wake losses with deep convolutional hierarchical encoder–decoder neural networks

DA Romero, S Hasanpoor, EGA Antonini… - APL Machine …, 2024 - pubs.aip.org
Wind turbine wakes are the most significant factor affecting wind farm performance,
decreasing energy production and increasing fatigue loads in downstream turbines. Wind …

NAS-SE: Designing A Highly-Efficient In-Situ Neural Architecture Search Engine for Large-Scale Deployment

Q Wan, L Wang, J Wang, SL Song, X Fu - … of the 56th Annual IEEE/ACM …, 2023 - dl.acm.org
The emergence of Neural Architecture Search (NAS) enables an automated neural network
development process that potentially replaces manually-enabled machine learning …