An overview of efficient interconnection networks for deep neural network accelerators

SM Nabavinejad, M Baharloo, KC Chen… - IEEE Journal on …, 2020 - ieeexplore.ieee.org
Deep Neural Networks (DNNs) have shown significant advantages in many domains, such
as pattern recognition, prediction, and control optimization. The edge computing demand in …

Coordinated batching and DVFS for DNN inference on GPU accelerators

SM Nabavinejad, S Reda… - IEEE transactions on …, 2022 - ieeexplore.ieee.org
Employing hardware accelerators to improve the performance and energy-efficiency of DNN
applications is on the rise. One challenge of using hardware accelerators, including the GPU …

EAIS: Energy-aware adaptive scheduling for CNN inference on high-performance GPUs

C Yao, W Liu, W Tang, S Hu - Future Generation Computer Systems, 2022 - Elsevier
Recently, a large number of convolutional neural network (CNN) inference services have
emerged on high-performance Graphic Processing Units (GPUs). However, GPUs are high …

When the metaverse meets carbon neutrality: Ongoing efforts and directions

F Liu, Q Pei, S Chen, Y Yuan, L Wang… - arXiv preprint arXiv …, 2023 - arxiv.org
The metaverse has recently gained increasing attention from the public. It builds up a virtual
world where we can live as a new role regardless of the role we play in the physical world …

Dynamic voltage and frequency scaling to improve energy-efficiency of hardware accelerators

S Liu, A Karanth - 2021 IEEE 28th International Conference on …, 2021 - ieeexplore.ieee.org
Neural networks (NNs) have been used in a wide variety of artificial intelligence (AI)
applications, including speech recognition, image recognition, automatic robotics, and …

Know Your Enemy To Save Cloud Energy: Energy-Performance Characterization of Machine Learning Serving

J Yu, J Kim, E Seo - 2023 IEEE International Symposium on …, 2023 - ieeexplore.ieee.org
The proportion of machine learning (ML) inference in modern cloud workloads is rapidly
increasing, and graphic processing units (GPUs) are the most preferred computational …

BatchSizer: Power-performance trade-off for DNN inference

SM Nabavinejad, S Reda, M Ebrahimi - Proceedings of the 26th Asia and …, 2021 - dl.acm.org
GPU accelerators can deliver significant improvement for DNN processing; however, their
performance is limited by internal and external parameters. A well-known parameter that …

EALI: Energy-aware layer-level scheduling for convolutional neural network inference services on GPUs

C Yao, W Liu, Z Liu, L Yan, S Hu, W Tang - Neurocomputing, 2022 - Elsevier
The success of convolutional neural networks (CNNs) has made low-latency inference
services on Graphic Processing Units (GPUs) a hot research topic. However, GPUs are …

Improving GPU Energy Efficiency through an Application-transparent Frequency Scaling Policy with Performance Assurance

Y Zhang, Q Wang, Z Lin, P Xu, B Wang - Proceedings of the Nineteenth …, 2024 - dl.acm.org
Power consumption is one of the top limiting factors in high-performance computing systems
and data centers, and dynamic voltage and frequency scaling (DVFS) is an important …

Decoupling GPGPU voltage-frequency scaling for deep-learning applications

F Mendes, P Tomás, N Roma - Journal of Parallel and Distributed …, 2022 - Elsevier
The use of GPUs to accelerate DNN training and inference is already widely adopted,
allowing for a significant performance increase. However, this performance is usually …