Partitioning dnns for optimizing distributed inference performance on cooperative edge devices: A genetic algorithm approach

J Na, H Zhang, J Lian, B Zhang - Applied Sciences, 2022 - mdpi.com
To fully unleash the potential of edge devices, it is popular to cut a neural network into
multiple pieces and distribute them among available edge devices to perform inference …

Automatic distributed deep learning using resource-constrained edge devices

A Gutierrez-Torre, K Bahadori, W Iqbal… - IEEE Internet of …, 2021 - ieeexplore.ieee.org
Processing data generated at high volume and speed from the Internet of Things, smart
cities, domotic, intelligent surveillance, and e-healthcare systems require efficient data …

An in-depth analysis of distributed training of deep neural networks

Y Ko, K Choi, J Seo, SW Kim - 2021 IEEE International Parallel …, 2021 - ieeexplore.ieee.org
As the popularity of deep learning in industry rapidly grows, efficient training of deep neural
networks (DNNs) becomes important. To train a DNN with a large amount of data, distributed …

Distredge: Speeding up convolutional neural network inference on distributed edge devices

X Hou, Y Guan, T Han, N Zhang - 2022 IEEE International …, 2022 - ieeexplore.ieee.org
As the number of edge devices with computing resources (eg, embedded GPUs, mobile
phones, and laptops) in-creases, recent studies demonstrate that it can be beneficial to col …

Fully distributed deep learning inference on resource-constrained edge devices

R Stahl, Z Zhao, D Mueller-Gritschneder… - … , and Simulation: 19th …, 2019 - Springer
Performing inference tasks of deep learning applications on IoT edge devices ensures
privacy of input data and can result in shorter latency when compared to a cloud solution. As …

A flexible research-oriented framework for distributed training of deep neural networks

S Barrachina, A Castelló, M Catalán… - 2021 IEEE …, 2021 - ieeexplore.ieee.org
We present PyDTNN, a framework for training deep neural networks (DNNs) on clusters of
computers that has been designed as a research-oriented tool with a low learning curve …

CoopFL: Accelerating federated learning with DNN partitioning and offloading in heterogeneous edge computing

Z Wang, H Xu, Y Xu, Z Jiang, J Liu - Computer Networks, 2023 - Elsevier
Federated learning (FL), a novel distributed machine learning (DML) approach, has been
widely adopted to train deep neural networks (DNNs), over massive data in edge computing …

Cooperative distributed deep neural network deployment with edge computing

CY Yang, JJ Kuo, JP Sheu… - ICC 2021-IEEE …, 2021 - ieeexplore.ieee.org
Deep Neural Networks (DNNs) are widely used to analyze the abundance of data collected
by massive Internet-of-Thing (IoT) devices. The traditional approaches usually send the data …

A Data and Model Parallelism based Distributed Deep Learning System in a Network of Edge Devices

T Sen, H Shen - 2023 32nd International Conference on …, 2023 - ieeexplore.ieee.org
With the emergence of edge computing along with its local computation advantage over the
cloud, methods for distributed deep learning (DL) training on edge nodes have been …

Efficient-grad: Efficient training deep convolutional neural networks on edge devices with grad ient optimizations

Z Hong, CP Yue - ACM Transactions on Embedded Computing Systems …, 2022 - dl.acm.org
With the prospering of mobile devices, the distributed learning approach, enabling model
training with decentralized data, has attracted great interest from researchers. However, the …