Collaborative Deep Neural Network Inference via Mobile Edge Computing

W Wu, Y Tang, P Yang, W Zhang, N Zhang - … , Computing, and Control for …, 2022 - Springer
Deep neural network (DNN) inference with low delay and high accuracy requirements is
usually computation intensive. The collaboration among mobile devices and the network …

Ftpipehd: A fault-tolerant pipeline-parallel distributed training approach for heterogeneous edge devices

Y Chen, Q Yang, S He, Z Shi, J Chen… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
With the increasing proliferation of Internet-of-Things (IoT) devices, there is a growing trend
towards distributing the power of deep learning (DL) among edge devices rather than …

Throughput maximization of delay-aware DNN inference in edge computing by exploring DNN model partitioning and inference parallelism

J Li, W Liang, Y Li, Z Xu, X Jia… - IEEE Transactions on …, 2021 - ieeexplore.ieee.org
Mobile Edge Computing (MEC) has emerged as a promising paradigm catering to
overwhelming explosions of mobile applications, by offloading compute-intensive tasks to …

Accelerating Deep Neural Network Tasks Through Edge-Device Adaptive Inference

X Zhang, Y Teng, N Wang, B Sun… - 2023 IEEE 34th Annual …, 2023 - ieeexplore.ieee.org
As the key technology of artificial intelligence (AI), Deep Neural Networks (DNNs) have been
widely used in mobile applications, such as video analytics in autonomous driving …

ADDA: Adaptive distributed DNN inference acceleration in edge computing environment

H Wang, G Cai, Z Huang, F Dong - 2019 IEEE 25th …, 2019 - ieeexplore.ieee.org
Implementing intelligent mobile applications on IoT devices with DNN technology has
become an inevitable trend. Due to the limitations of the size of DNN model deployed onto …

Dynamic DNN decomposition for lossless synergistic inference

B Zhang, T Xiang, H Zhang, T Li… - 2021 IEEE 41st …, 2021 - ieeexplore.ieee.org
Deep neural networks (DNNs) sustain high performance in today's data processing
applications. DNN inference is resource-intensive thus is difficult to fit into a mobile device …

Distributed inference with deep learning models across heterogeneous edge devices

C Hu, B Li - IEEE INFOCOM 2022-IEEE Conference on …, 2022 - ieeexplore.ieee.org
Recent years witnessed an increasing research attention in deploying deep learning models
on edge devices for inference. Due to limited capabilities and power constraints, it may be …

Pico: Pipeline inference framework for versatile cnns on diverse mobile devices

X Yang, Z Xu, Q Qi, J Wang, H Sun… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Distributing the inference of convolutional neural network (CNN) to multiple mobile devices
has been studied in recent years to achieve real-time inference without losing accuracy …

OfpCNN: On-Demand Fine-Grained Partitioning for CNN Inference Acceleration in Heterogeneous Devices

L Yang, C Zheng, X Shen, G Xie - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Collaborative inference is a promising method for balancing the limited computational power
of Internet of Things (IoT) devices with the huge computational demands of convolutional …

Optimizing job offloading schedule for collaborative DNN inference

Y Duan, J Wu - IEEE Transactions on Mobile Computing, 2023 - ieeexplore.ieee.org
Deep Neural Networks (DNNs) have been widely deployed in mobile applications. DNN
inference latency is a critical metric to measure the service quality of those applications …