DIANA: An end-to-end energy-efficient digital and ANAlog hybrid neural network SoC

K Ueyoshi, IA Papistas, P Houshmand… - … Solid-State Circuits …, 2022 - ieeexplore.ieee.org
Energy-efficient matrix-vector multiplications (MVMs) are key to bringing neural network
(NN) inference to edge devices. This has led to a wide range of state-of-the-art MVM …

A 22 nm, 1540 TOP/s/W, 12.1 TOP/s/mm2 in-Memory Analog Matrix-Vector-Multiplier for DNN Acceleration

IA Papistas, S Cosemans, B Rooseleer… - 2021 IEEE Custom …, 2021 - ieeexplore.ieee.org
Deep learning constitutes the state-of-the-art in machine learning, from data mining to
computer vision and natural language processing. Energy-efficient matrix-vector …

Enhancing the energy efficiency and robustness of TinyML computer vision using coarsely-quantized log-gradient input images

Q Lu, B Murmann - ACM Transactions on Embedded Computing …, 2024 - dl.acm.org
This article studies the merits of applying log-gradient input images to convolutional neural
networks (CNNs) for tinyML computer vision (CV). We show that log gradients enable:(i) …

IGZO-based compute cell for analog in-memory computing—DTCO analysis to enable ultralow-power AI at edge

D Saito, J Doevenspeck, S Cosemans… - … on Electron Devices, 2020 - ieeexplore.ieee.org
We propose, for the first time, an indium gallium zinc oxide (IGZO)-based 2T1C compute cell
(IGZO-cell) for analog in-memory computing. To assess the impact of an IGZO-cell-based …

TAB: Unified and optimized ternary, binary, and mixed-precision neural network inference on the edge

S Zhu, LHK Duong, W Liu - ACM Transactions on Embedded Computing …, 2022 - dl.acm.org
Ternary Neural Networks (TNNs) and mixed-precision Ternary Binary Networks (TBNs) have
demonstrated higher accuracy compared to Binary Neural Networks (BNNs) while providing …

Improving the energy efficiency and robustness of TinyML computer vision using log-gradient input images

Q Lu, B Murmann - arXiv preprint arXiv:2203.02571, 2022 - arxiv.org
This paper studies the merits of applying log-gradient input images to convolutional neural
networks (CNNs) for tinyML computer vision (CV). We show that log gradients enable:(i) …

OxRRAM-based analog in-memory computing for deep neural network inference: A conductance variability study

J Doevenspeck, R Degraeve, A Fantini… - … on Electron Devices, 2021 - ieeexplore.ieee.org
Analog in-memory compute (AiMC) is a promising approach to efficiently process deep
neural networks (DNNs). Due to its small size and nonvolatility, oxide resistive random …

Design-technology space exploration for energy efficient AiMC-based inference acceleration

D Bhattacharjee, N Laubeuf… - … on Circuits and …, 2021 - ieeexplore.ieee.org
Extremely energy-efficient convolutional neural network inference (CNN) is recently enabled
by analog in-memory compute (AiMC). The integration of AiMC in a primarily digital …

Precision-aware latency and energy balancing on multi-accelerator platforms for dnn inference

M Risso, A Burrello, GM Sarda, L Benini… - 2023 IEEE/ACM …, 2023 - ieeexplore.ieee.org
The need to execute Deep Neural Networks (DNNs) at low latency and low power at the
edge has spurred the development of new heterogeneous Systems-on-Chips (SoCs) …

MAFAT: Memory-Aware Fusing and Tiling of Neural Networks for Accelerated Edge Inference

J Farley, A Gerstlauer - International Embedded Systems Symposium, 2022 - Springer
A rising research challenge is running costly machine learning (ML) networks locally on
resource-constrained edge devices. ML networks with large convolutional layers can easily …