Decentralized low-latency collaborative inference via ensembles on the edge

M Malka, E Farhan, H Morgenstern… - arXiv preprint arXiv …, 2022 - arxiv.org
The success of deep neural networks (DNNs) is heavily dependent on computational
resources. While DNNs are often employed on cloud servers, there is a growing need to …

Collaborative inference via ensembles on the edge

N Shlezinger, E Farhan… - ICASSP 2021-2021 …, 2021 - ieeexplore.ieee.org
The success of deep neural networks (DNNs) as an enabler of artificial intelligence (AI) is
heavily dependent on high computational resources. The increasing demands for …

DNN real-time collaborative inference acceleration with mobile edge computing

R Yang, Y Li, H He, W Zhang - 2022 International Joint …, 2022 - ieeexplore.ieee.org
The collaborative inference approach splits the Deep Neural Networks (DNNs) model into
two parts. It runs collaboratively on the end device and cloud server to minimize inference …

Auto-tuning neural network quantization framework for collaborative inference between the cloud and edge

G Li, L Liu, X Wang, X Dong, P Zhao, X Feng - Artificial Neural Networks …, 2018 - Springer
Recently, deep neural networks (DNNs) have been widely applied in mobile intelligent
applications. The inference for the DNNs is usually performed in the cloud. However, it leads …

Towards real-time cooperative deep inference over the cloud and edge end devices

S Zhang, Y Li, X Liu, S Guo, W Wang, J Wang… - Proceedings of the …, 2020 - dl.acm.org
Deep neural networks (DNNs) have been widely used in many intelligent applications such
as object recognition and automatic driving due to their superior performance in conducting …

SPINN: synergistic progressive inference of neural networks over device and cloud

S Laskaridis, SI Venieris, M Almeida… - Proceedings of the 26th …, 2020 - dl.acm.org
Despite the soaring use of convolutional neural networks (CNNs) in mobile applications,
uniformly sustaining high-performance inference on mobile has been elusive due to the …

DISCO: Distributed Inference with Sparse Communications

M Qin, C Sun, J Hofmann… - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
Deep neural networks (DNNs) have great potential to solve many real-world problems, but
they usually require an extensive amount of computation and memory. It is of great difficulty …

Canoe: A system for collaborative learning for neural nets

H Daga, Y Chen, A Agrawal, A Gavrilovska - arXiv preprint arXiv …, 2021 - arxiv.org
For highly distributed environments such as edge computing, collaborative learning
approaches eschew the dependence on a global, shared model, in favor of models tailored …

DECC: Delay-Aware Edge-Cloud Collaboration for Accelerating DNN Inference

Z Zhuang, J Chen, W Xu, Q Qi, S Guo… - … on Emerging Topics …, 2024 - ieeexplore.ieee.org
Deep neural network (DNN)-enabled edge intelligence has been widely adopted to support
a variety of smart applications because of its ability to preserve privacy and conserve …

OfpCNN: On-Demand Fine-Grained Partitioning for CNN Inference Acceleration in Heterogeneous Devices

L Yang, C Zheng, X Shen, G Xie - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Collaborative inference is a promising method for balancing the limited computational power
of Internet of Things (IoT) devices with the huge computational demands of convolutional …