Embedded deep learning accelerators: A survey on recent advances

G Akkad, A Mansour, E Inaty - IEEE Transactions on Artificial …, 2023 - ieeexplore.ieee.org
The exponential increase in generated data as well as the advances in high-performance
computing has paved the way for the use of complex machine learning methods. Indeed, the …

In-memory computing accelerators for emerging learning paradigms

D Reis, AF Laguna, M Niemier, XS Hu - Proceedings of the 28th Asia and …, 2023 - dl.acm.org
Over the past decades, emerging, data-driven machine learning (ML) paradigms have
increased in popularity, and revolutionized many application domains. To date, a substantial …

AccelTran: A sparsity-aware accelerator for dynamic inference with transformers

S Tuli, NK Jha - IEEE Transactions on Computer-Aided Design …, 2023 - ieeexplore.ieee.org
Self-attention-based transformer models have achieved tremendous success in the domain
of natural language processing. Despite their efficacy, accelerating the transformer is …

Semantic memory–based dynamic neural network using memristive ternary CIM and CAM for 2D and 3D vision

Y Zhang, W Zhang, S Wang, N Lin, Y Yu, Y He… - Science …, 2024 - science.org
The brain is dynamic, associative, and efficient. It reconfigures by associating the inputs with
past experiences, with fused memory and processing. In contrast, AI models are static …

Designing precharge-free energy-efficient content-addressable memories

R Taco, E Garzón, R Hanhan, A Teman… - … Transactions on Very …, 2024 - ieeexplore.ieee.org
Content-addressable memory (CAM) is a specialized type of memory that facilitates
massively parallel comparison of a search pattern against its entire content. State-of-the-art …

Tic-sat: Tightly-coupled systolic accelerator for transformers

A Amirshahi, JAH Klein, G Ansaloni… - … of the 28th Asia and South …, 2023 - dl.acm.org
Transformer models have achieved impressive results in various AI scenarios, ranging from
vision to natural language processing. However, their computational complexity and their …

Advancements in Content-Addressable Memory (CAM) Circuits: State-of-the-Art, Applications, and Future Directions in the AI Domain

T Molom-Ochir, B Taylor, H Li… - IEEE Transactions on …, 2025 - ieeexplore.ieee.org
Content-Addressable Memory (CAM) circuits, distinguished by their ability to accelerate data
retrieval through a direct content-matching function, are increasingly crucial in the era of AI …

Analog In-Memory Computing Attention Mechanism for Fast and Energy-Efficient Large Language Models

N Leroux, PP Manea, C Sudarshan… - arXiv preprint arXiv …, 2024 - arxiv.org
Transformer neural networks, driven by self-attention mechanisms, are core components of
foundational and Large Language Models. In generative transformers, self-attention uses …

Dynamic neural network with memristive CIM and CAM for 2D and 3D vision

Y Zhang, W Zhang, S Wang, N Lin, Y Yu, Y He… - arXiv preprint arXiv …, 2024 - arxiv.org
The brain is dynamic, associative and efficient. It reconfigures by associating the inputs with
past experiences, with fused memory and processing. In contrast, AI models are static …

Memory Is All You Need: An Overview of Compute-in-Memory Architectures for Accelerating Large Language Model Inference

C Wolters, X Yang, U Schlichtmann… - arXiv preprint arXiv …, 2024 - arxiv.org
Large language models (LLMs) have recently transformed natural language processing,
enabling machines to generate human-like text and engage in meaningful conversations …