Efficient acceleration of deep learning inference on resource-constrained edge devices: A review

MMH Shuvo, SK Islam, J Cheng… - Proceedings of the …, 2022 - ieeexplore.ieee.org
Successful integration of deep neural networks (DNNs) or deep learning (DL) has resulted
in breakthroughs in many areas. However, deploying these highly accurate models for data …

Edge intelligence: Architectures, challenges, and applications

D Xu, T Li, Y Li, X Su, S Tarkoma, T Jiang… - arXiv preprint arXiv …, 2020 - arxiv.org
Edge intelligence refers to a set of connected systems and devices for data collection,
caching, processing, and analysis in locations close to where data is captured based on …

Knowledge distillation: A survey

J Gou, B Yu, SJ Maybank, D Tao - International Journal of Computer Vision, 2021 - Springer
In recent years, deep neural networks have been successful in both industry and academia,
especially for computer vision tasks. The great success of deep learning is mainly due to its …

Edge intelligence: Empowering intelligence to the edge of network

D Xu, T Li, Y Li, X Su, S Tarkoma, T Jiang… - Proceedings of the …, 2021 - ieeexplore.ieee.org
Edge intelligence refers to a set of connected systems and devices for data collection,
caching, processing, and analysis proximity to where data are captured based on artificial …

Deep interest evolution network for click-through rate prediction

G Zhou, N Mou, Y Fan, Q Pi, W Bian, C Zhou… - Proceedings of the AAAI …, 2019 - aaai.org
Click-through rate (CTR) prediction, whose goal is to estimate the probability of a user
clicking on the item, has become one of the core tasks in the advertising system. For CTR …

Knowledge distillation via instance relationship graph

Y Liu, J Cao, B Li, C Yuan, W Hu… - Proceedings of the …, 2019 - openaccess.thecvf.com
The key challenge of knowledge distillation is to extract general, moderate and sufficient
knowledge from a teacher network to guide a student network. In this paper, a novel …

Knowledge distillation via route constrained optimization

X Jin, B Peng, Y Wu, Y Liu, J Liu… - Proceedings of the …, 2019 - openaccess.thecvf.com
Distillation-based learning boosts the performance of the miniaturized neural network based
on the hypothesis that the representation of a teacher model can be used as structured and …

Low-resolution face recognition in the wild via selective knowledge distillation

S Ge, S Zhao, C Li, J Li - IEEE Transactions on Image …, 2018 - ieeexplore.ieee.org
Typically, the deployment of face recognition models in the wild needs to identify low-
resolution faces with extremely low computational cost. To address this problem, a feasible …

2D Human pose estimation: A survey

H Chen, R Feng, S Wu, H Xu, F Zhou, Z Liu - Multimedia systems, 2023 - Springer
Human pose estimation aims at localizing human anatomical keypoints or body parts in the
input data (eg, images, videos, or signals). It forms a crucial component in enabling …

Transforming large-size to lightweight deep neural networks for IoT applications

R Mishra, H Gupta - ACM Computing Surveys, 2023 - dl.acm.org
Deep Neural Networks (DNNs) have gained unprecedented popularity due to their high-
order performance and automated feature extraction capability. This has encouraged …