DNN inference acceleration with partitioning and early exiting in edge computing

C Li, H Xu, Y Xu, Z Wang, L Huang - … WASA 2021, Nanjing, China, June 25 …, 2021 - Springer
Recently, deep neural networks (DNNs) have been applied to most intelligent applications
and deployed on different kinds of devices. However, DNN inference is resource-intensive …

EdgeLD: Locally distributed deep learning inference on edge device clusters

F Xue, W Fang, W Xu, Q Wang, X Ma… - 2020 IEEE 22nd …, 2020 - ieeexplore.ieee.org
Deep Neural Networks (DNN) have been widely used in a large number of application
scenarios. However, DNN models are generally both computation-intensive and memory …

[HTML][HTML] EDITORS: Energy-aware DynamIc Task Offloading using Deep Reinforcement transfer learning in SDN-enabled edge nodes

T Baker, Z Al Aghbari, AM Khedr, N Ahmed, S Girija - Internet of Things, 2024 - Elsevier
In mobile edge computing systems, a task offloading approach should balance efficiency,
adaptability, trust management, and reliability. This approach aims to maximize resource …

Context-Aware Layer Scheduling for Seamless Neural Network Inference in Cloud-Edge Systems

M Stammler, V Sidorenko, F Kreß… - 2023 IEEE 16th …, 2023 - ieeexplore.ieee.org
With deep neural networks (DNNs) gaining popularity for tasks like object detection and
image segmentation in domains like autonomous driving and smart agriculture, DNN …

Dystri: A Dynamic Inference based Distributed DNN Service Framework on Edge

X Hou, Y Guan, T Han - … of the 52nd International Conference on Parallel …, 2023 - dl.acm.org
Deep neural network (DNN) inference poses unique challenges in serving computational
requests due to high request intensity, concurrent multi-user scenarios, and diverse …

Enabling low latency edge intelligence based on multi-exit dnns in the wild

Z Huang, F Dong, D Shen, J Zhang… - 2021 IEEE 41st …, 2021 - ieeexplore.ieee.org
In recent years, deep neural networks (DNNs) have witnessed a booming of artificial
intelligence Internet of Things applications with stringent demands across high accuracy and …

DECC: Delay-Aware Edge-Cloud Collaboration for Accelerating DNN Inference

Z Zhuang, J Chen, W Xu, Q Qi, S Guo… - … on Emerging Topics …, 2024 - ieeexplore.ieee.org
Deep neural network (DNN)-enabled edge intelligence has been widely adopted to support
a variety of smart applications because of its ability to preserve privacy and conserve …

AdaMEC: Towards a Context-adaptive and Dynamically Combinable DNN Deployment Framework for Mobile Edge Computing

B Pang, S Liu, H Wang, B Guo, Y Wang… - ACM Transactions on …, 2023 - dl.acm.org
With the rapid development of deep learning, recent research on intelligent and interactive
mobile applications (eg, health monitoring, speech recognition) has attracted extensive …

AdaInNet: an adaptive inference engine for distributed deep neural networks offloading in IoT-FOG applications based on reinforcement learning

A Etefaghi, S Sharifian - The Journal of Supercomputing, 2023 - Springer
The increasing expansion of Internet-of-Things (IoT) in the world requires Big Data analytic
infrastructures to produce valuable knowledge in IoT applications. IoT includes devices with …

Distributed inference acceleration with adaptive DNN partitioning and offloading

T Mohammed, C Joe-Wong, R Babbar… - … -IEEE Conference on …, 2020 - ieeexplore.ieee.org
Deep neural networks (DNN) are the de-facto solution behind many intelligent applications
of today, ranging from machine translation to autonomous driving. DNNs are accurate but …