PArtNNer: Platform-agnostic adaptive edge-cloud DNN partitioning for minimizing end-to-end latency

SK Ghosh, A Raha, V Raghunathan… - ACM Transactions on …, 2024 - dl.acm.org
The last decade has seen the emergence of Deep Neural Networks (DNNs) as the de facto
algorithm for various computer vision applications. In intelligent edge devices, sensor data …

EdgeSP: Scalable multi-device parallel dnn inference on heterogeneous edge clusters

Z Gao, S Sun, Y Zhang, Z Mo, C Zhao - International Conference on …, 2021 - Springer
Edge computing has emerged as a promising line of research for processing large-scale
data and providing low-latency services. Unfortunately, deploying deep neural networks …

Performance analysis of local exit for distributed deep neural networks over cloud and edge computing

C Lee, S Hong, S Hong, T Kim - Etri Journal, 2020 - Wiley Online Library
In edge computing, most procedures, including data collection, data processing, and service
provision, are handled at edge nodes and not in the central cloud. This decreases the …

An adaptive DNN inference acceleration framework with end–edge–cloud collaborative computing

G Liu, F Dai, X Xu, X Fu, W Dou, N Kumar… - Future Generation …, 2023 - Elsevier
Abstract Deep Neural Networks (DNNs) based on intelligent applications have been
intensively deployed on mobile devices. Unfortunately, resource-constrained mobile devices …

EdgeLD: Locally distributed deep learning inference on edge device clusters

F Xue, W Fang, W Xu, Q Wang, X Ma… - 2020 IEEE 22nd …, 2020 - ieeexplore.ieee.org
Deep Neural Networks (DNN) have been widely used in a large number of application
scenarios. However, DNN models are generally both computation-intensive and memory …

Multi-compression scale DNN inference acceleration based on cloud-edge-end collaboration

H Qi, F Ren, L Wang, P Jiang, S Wan… - ACM Transactions on …, 2024 - dl.acm.org
Edge intelligence has emerged as a promising paradigm to accelerate DNN inference by
model partitioning, which is particularly useful for intelligent scenarios that demand high …

A Converting Autoencoder Toward Low-latency and Energy-efficient DNN Inference at the Edge

H Mahmud, P Kang, K Desai, P Lama… - arXiv preprint arXiv …, 2024 - arxiv.org
Reducing inference time and energy usage while maintaining prediction accuracy has
become a significant concern for deep neural networks (DNN) inference on resource …

Automated exploration and implementation of distributed cnn inference at the edge

X Guo, AD Pimentel, T Stefanov - IEEE Internet of Things …, 2023 - ieeexplore.ieee.org
For model inference of convolutional neural networks (CNNs), we nowadays witness a shift
from the Cloud to the Edge. Unfortunately, deploying and inferring large, compute-and …

Cost-effective machine learning inference offload for edge computing

C Makaya, A Iyer, J Salfity, M Athreya… - arXiv preprint arXiv …, 2020 - arxiv.org
Computing at the edge is increasingly important since a massive amount of data is
generated. This poses challenges in transporting all that data to the remote data centers and …

[PDF][PDF] Scission: Context-aware and performance-driven edge-based distributed deep neural networks

L Lockhart, P Harvey, P Imai, P Willis… - arXiv e-prints, pp. arXiv …, 2020 - academia.edu
Partitioning and distributing deep neural networks (DNNs) across end-devices, edge
resources and the cloud has a potential twofold advantage: preserving privacy of the input …