Graft: Efficient inference serving for hybrid deep learning with SLO guarantees via DNN re-alignment

J Wu, L Wang, Q Jin, F Liu - IEEE Transactions on Parallel and …, 2023 - ieeexplore.ieee.org
Deep neural networks (DNNs) have been widely adopted for various mobile inference tasks,
yet their ever-increasing computational demands are hindering their deployment on …

HiTDL: High-throughput deep learning inference at the hybrid mobile edge

J Wu, L Wang, Q Pei, X Cui, F Liu… - IEEE Transactions on …, 2022 - ieeexplore.ieee.org
Deep neural networks (DNNs) have become a critical component for inference in modern
mobile applications, but the efficient provisioning of DNNs is non-trivial. Existing mobile-and …

Edge AI: On-demand accelerating deep neural network inference via edge computing

E Li, L Zeng, Z Zhou, X Chen - IEEE Transactions on Wireless …, 2019 - ieeexplore.ieee.org
As a key technology of enabling Artificial Intelligence (AI) applications in 5G era, Deep
Neural Networks (DNNs) have quickly attracted widespread attention. However, it is …

Edge intelligence: On-demand deep learning model co-inference with device-edge synergy

E Li, Z Zhou, X Chen - Proceedings of the 2018 workshop on mobile …, 2018 - dl.acm.org
As the backbone technology of machine learning, deep neural networks (DNNs) have have
quickly ascended to the spotlight. Running DNNs on resource-constrained mobile devices …

MNN: A universal and efficient inference engine

X Jiang, H Wang, Y Chen, Z Wu… - Proceedings of …, 2020 - proceedings.mlsys.org
Deploying deep learning (DL) models on mobile devices draws more and more attention
recently. However, designing an efficient inference engine on devices is under the great …

An adaptive DNN inference acceleration framework with end–edge–cloud collaborative computing

G Liu, F Dai, X Xu, X Fu, W Dou, N Kumar… - Future Generation …, 2023 - Elsevier
Abstract Deep Neural Networks (DNNs) based on intelligent applications have been
intensively deployed on mobile devices. Unfortunately, resource-constrained mobile devices …

Band: coordinated multi-dnn inference on heterogeneous mobile processors

JS Jeong, J Lee, D Kim, C Jeon, C Jeong… - Proceedings of the 20th …, 2022 - dl.acm.org
The rapid development of deep learning algorithms, as well as innovative hardware
advancements, encourages multi-DNN workloads such as augmented reality applications …

Enable pipeline processing of DNN co-inference tasks in the mobile-edge cloud

S Hu, C Dong, W Wen - 2021 IEEE 6th International …, 2021 - ieeexplore.ieee.org
Deep Neural Network (DNN) based artificial intelligence help driving the great development
of mobile Internet. However, the hardware of a mobile device may not be sufficiently to meet …

Mdinference: Balancing inference accuracy and latency for mobile applications

SS Ogden, T Guo - 2020 IEEE International Conference on …, 2020 - ieeexplore.ieee.org
Deep Neural Networks are allowing mobile devices to incorporate a wide range of features
into user applications. However, the computational complexity of these models makes it …

Computation offloading scheduling for deep neural network inference in mobile computing

Y Duan, J Wu - 2021 IEEE/ACM 29th International Symposium …, 2021 - ieeexplore.ieee.org
The quality of service (QoS) of intelligent applications on mobile devices heavily depends on
the inference speed of Deep Neural Network (DNN) models. Cooperative DNN inference …