A survey on approximate edge AI for energy efficient autonomous driving services

D Katare, D Perino, J Nurmi, M Warnier… - … Surveys & Tutorials, 2023 - ieeexplore.ieee.org
Autonomous driving services depends on active sensing from modules such as camera,
LiDAR, radar, and communication units. Traditionally, these modules process the sensed …

[HTML][HTML] Adaptive approximate computing in edge AI and IoT applications: A review

HJ Damsgaard, A Grenier, D Katare, Z Taufique… - Journal of Systems …, 2024 - Elsevier
Recent advancements in hardware and software systems have been driven by the
deployment of emerging smart health and mobility applications. These developments have …

Towards mixed-precision quantization of neural networks via constrained optimization

W Chen, P Wang, J Cheng - Proceedings of the IEEE/CVF …, 2021 - openaccess.thecvf.com
Quantization is a widely used technique to compress and accelerate deep neural networks.
However, conventional quantization methods use the same bit-width for all (or most of) the …

Human activity recognition on microcontrollers with quantized and adaptive deep neural networks

F Daghero, A Burrello, C Xie, M Castellano… - ACM Transactions on …, 2022 - dl.acm.org
Human Activity Recognition (HAR) based on inertial data is an increasingly diffused task on
embedded devices, from smartphones to ultra low-power sensors. Due to the high …

[HTML][HTML] A privacy protection approach in edge-computing based on maximized dnn partition strategy with energy saving

G Chaopeng, L Zhengqing, S Jie - Journal of Cloud Computing, 2023 - Springer
With the development of deep neural network (DNN) techniques, applications of DNNs show
state-of-art performance. In the cloud edge collaborative mode, edge devices upload the raw …

[HTML][HTML] SensiMix: Sensitivity-Aware 8-bit index & 1-bit value mixed precision quantization for BERT compression

T Piao, I Cho, U Kang - PloS one, 2022 - journals.plos.org
Given a pre-trained BERT, how can we compress it to a fast and lightweight one while
maintaining its accuracy? Pre-training language model, such as BERT, is effective for …

Mixed-precision neural networks: A survey

M Rakka, ME Fouda, P Khargonekar… - arXiv preprint arXiv …, 2022 - arxiv.org
Mixed-precision Deep Neural Networks achieve the energy efficiency and throughput
needed for hardware deployment, particularly when the resources are limited, without …

Evolutionary quantization of neural networks with mixed-precision

Z Liu, X Zhang, S Wang, S Ma… - ICASSP 2021-2021 IEEE …, 2021 - ieeexplore.ieee.org
Quantization is an effective way for reducing the memory and computation costs of deep
neural networks. Most of existing methods exploit the fixed-precision quantization approach …

Sensitivity-Aware Mixed-Precision Quantization and Width Optimization of Deep Neural Networks Through Cluster-Based Tree-Structured Parzen Estimation

S Azizi, M Nazemi, A Fayyazi, M Pedram - arXiv preprint arXiv:2308.06422, 2023 - arxiv.org
As the complexity and computational demands of deep learning models rise, the need for
effective optimization methods for neural network designs becomes paramount. This work …

Design Space Exploration of Low-Bit Quantized Neural Networks for Visual Place Recognition

O Grainge, M Milford, I Bodala… - IEEE Robotics and …, 2024 - ieeexplore.ieee.org
Visual Place Recognition (VPR) is a critical task for performing global re-localization in
visual perception systems, requiring the ability to recognize a previously visited location …