Edge intelligence: On-demand deep learning model co-inference with device-edge synergy

E Li, Z Zhou, X Chen - Proceedings of the 2018 workshop on mobile …, 2018 - dl.acm.org
As the backbone technology of machine learning, deep neural networks (DNNs) have have
quickly ascended to the spotlight. Running DNNs on resource-constrained mobile devices …

Edge AI: On-demand accelerating deep neural network inference via edge computing

E Li, L Zeng, Z Zhou, X Chen - IEEE Transactions on Wireless …, 2019 - ieeexplore.ieee.org
As a key technology of enabling Artificial Intelligence (AI) applications in 5G era, Deep
Neural Networks (DNNs) have quickly attracted widespread attention. However, it is …

Communication-computation trade-off in resource-constrained edge inference

J Shao, J Zhang - IEEE Communications Magazine, 2020 - ieeexplore.ieee.org
The recent breakthrough in artificial intelligence (AI), especially deep neural networks
(DNNs), has affected every branch of science and technology. Particularly, edge AI has …

An adaptive DNN inference acceleration framework with end–edge–cloud collaborative computing

G Liu, F Dai, X Xu, X Fu, W Dou, N Kumar… - Future Generation …, 2023 - Elsevier
Abstract Deep Neural Networks (DNNs) based on intelligent applications have been
intensively deployed on mobile devices. Unfortunately, resource-constrained mobile devices …

Convergence of edge computing and deep learning: A comprehensive survey

X Wang, Y Han, VCM Leung, D Niyato… - … Surveys & Tutorials, 2020 - ieeexplore.ieee.org
Ubiquitous sensors and smart devices from factories and communities are generating
massive amounts of data, and ever-increasing computing power is driving the core of …

Improving device-edge cooperative inference of deep learning via 2-step pruning

W Shi, Y Hou, S Zhou, Z Niu, Y Zhang… - IEEE INFOCOM 2019 …, 2019 - ieeexplore.ieee.org
Deep neural networks (DNNs) are state-of-the-art solutions for many machine learning
applications, and have been widely used on mobile devices. Running DNNs on …

Multi-exit DNN inference acceleration based on multi-dimensional optimization for edge intelligence

F Dong, H Wang, D Shen, Z Huang… - IEEE Transactions …, 2022 - ieeexplore.ieee.org
Edge intelligence, as a prospective paradigm for accelerating DNN inference, is mostly
implemented by model partitioning which inevitably incurs the large transmission overhead …

Throughput maximization of delay-aware DNN inference in edge computing by exploring DNN model partitioning and inference parallelism

J Li, W Liang, Y Li, Z Xu, X Jia… - IEEE Transactions on …, 2021 - ieeexplore.ieee.org
Mobile Edge Computing (MEC) has emerged as a promising paradigm catering to
overwhelming explosions of mobile applications, by offloading compute-intensive tasks to …

Mistify: Automating {DNN} Model Porting for {On-Device} Inference at the Edge

P Guo, B Hu, W Hu - 18th USENIX Symposium on Networked Systems …, 2021 - usenix.org
AI applications powered by deep learning inference are increasingly run natively on edge
devices to provide better interactive user experience. This often necessitates fitting a model …

Head network distillation: Splitting distilled deep neural networks for resource-constrained edge computing systems

Y Matsubara, D Callegaro, S Baidya, M Levorato… - IEEE …, 2020 - ieeexplore.ieee.org
As the complexity of Deep Neural Network (DNN) models increases, their deployment on
mobile devices becomes increasingly challenging, especially in complex vision tasks such …