[PDF][PDF] Optimizing DNNs Model Partitioning for Enhanced Performance on Edge Devices.

M Al Maruf, A Azim - Canadian AI, 2023 - assets.pubpub.org
Abstract Deep Neural Networks (DNNs) have proven effective in various applications due to
their dominant performance. However, integrating DNNs into edge devices remains …

Pipeline Dnn Model Parallelism for Improving Performance of Embedded Applications

MA Maruf, A Azim, N Auluck, M Sahi - Available at SSRN 4382942 - papers.ssrn.com
Abstract Deep Neural Networks (DNNs) have gained widespread popularity in different
domain applications due to their dominant performance. Despite the prevalence of …

[HTML][HTML] Optimizing DNN training with pipeline model parallelism for enhanced performance in embedded systems

M Al Maruf, A Azim, N Auluck, M Sahi - Journal of Parallel and Distributed …, 2024 - Elsevier
Abstract Deep Neural Networks (DNNs) have gained widespread popularity in different
domain applications due to their dominant performance. Despite the prevalence of …

Model Parallelism Optimization for Distributed DNN Inference on Edge Devices

M Wang, L Qian, N Meng, Y Cheng… - 2023 IEEE 14th …, 2023 - ieeexplore.ieee.org
Deep neural networks (DNNs) have recently gained widespread application in various
domains. However, the computational and memory requirements of DNN models pose …

AccEPT: An Acceleration Scheme for Speeding Up Edge Pipeline-parallel Training

Y Chen, Y Yan, Q Yang, Y Shu, S He… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
It is usually infeasible to fit and train an entire large deep neural network (DNN) model using
a single edge device due to the limited resources. To facilitate intelligent applications across …

RoaD-RuNNer: Collaborative DNN partitioning and offloading on heterogeneous edge systems

AK Kakolyris, M Katsaragakis… - … , Automation & Test …, 2023 - ieeexplore.ieee.org
Deep Neural Networks (DNNs) are becoming extremely popular for many modern
applications deployed at the edge of the computing continuum. Despite their effectiveness …

EdgeSP: Scalable Multi-Device Parallel DNN Inference on Heterogeneous Edge Clusters

Z Gao, S Sun, Y Zhang, Z Mo, C Zhao - International Conference on …, 2021 - Springer
Edge computing has emerged as a promising line of research for processing large-scale
data and providing low-latency services. Unfortunately, deploying deep neural networks …

PORTEND: A Joint Performance Model for Partitioned Early-Exiting DNNs

M Ebrahimi, A da Silva Veith, M Gabel… - 2023 IEEE 29th …, 2023 - ieeexplore.ieee.org
The computation and storage requirements of Deep Neural Networks (DNNs) make them
challenging to deploy on edge devices, which often have limited resources. Conversely …

DNNSplit: Latency and Cost-efficient Split Point Identification for Multi-tier DNN Partitioning

P Kayal, A Leon-Garcia - IEEE Access, 2024 - ieeexplore.ieee.org
Due to the high computational demands inherent in Deep Neural Network (DNN)
executions, multi-tier environments have emerged as preferred platforms for DNN inference …

Pase: Parallelization strategies for efficient dnn training

V Elango - 2021 IEEE International Parallel and Distributed …, 2021 - ieeexplore.ieee.org
Training a deep neural network (DNN) requires substantial computational and memory
requirements. It is common to use multiple devices to train a DNN to reduce the overall …