Memory devices and applications for in-memory computing

A Sebastian, M Le Gallo, R Khaddam-Aljameh… - Nature …, 2020 - nature.com
Traditional von Neumann computing systems involve separate processing and memory
units. However, data movement is costly in terms of time and energy and this problem is …

A full spectrum of computing-in-memory technologies

Z Sun, S Kvatinsky, X Si, A Mehonic, Y Cai… - Nature Electronics, 2023 - nature.com
Computing in memory (CIM) could be used to overcome the von Neumann bottleneck and to
provide sustainable improvements in computing throughput and energy efficiency …

Machine learning at facebook: Understanding inference at the edge

CJ Wu, D Brooks, K Chen, D Chen… - … symposium on high …, 2019 - ieeexplore.ieee.org
At Facebook, machine learning provides a wide range of capabilities that drive many
aspects of user experience including ranking posts, content understanding, object detection …

PUMA: A programmable ultra-efficient memristor-based accelerator for machine learning inference

A Ankit, IE Hajj, SR Chalamalasetti, G Ndu… - Proceedings of the …, 2019 - dl.acm.org
Memristor crossbars are circuits capable of performing analog matrix-vector multiplications,
overcoming the fundamental energy efficiency limitations of digital logic. They have been …

[HTML][HTML] Analog architectures for neural network acceleration based on non-volatile memory

TP Xiao, CH Bennett, B Feinberg, S Agarwal… - Applied Physics …, 2020 - pubs.aip.org
Analog hardware accelerators, which perform computation within a dense memory array,
have the potential to overcome the major bottlenecks faced by digital hardware for data …

Research progress on memristor: From synapses to computing systems

X Yang, B Taylor, A Wu, Y Chen… - IEEE Transactions on …, 2022 - ieeexplore.ieee.org
As the limits of transistor technology are approached, feature size in integrated circuit
transistors has been reduced very near to the minimum physically-realizable channel length …

Mixed-precision in-memory computing

M Le Gallo, A Sebastian, R Mathis, M Manica… - Nature …, 2018 - nature.com
As complementary metal–oxide–semiconductor (CMOS) scaling reaches its technological
limits, a radical departure from traditional von Neumann systems, which involve separate …

Recnmp: Accelerating personalized recommendation with near-memory processing

L Ke, U Gupta, BY Cho, D Brooks… - 2020 ACM/IEEE 47th …, 2020 - ieeexplore.ieee.org
Personalized recommendation systems leverage deep learning models and account for the
majority of data center AI cycles. Their performance is dominated by memory-bound sparse …

Large-scale optical neural networks based on photoelectric multiplication

R Hamerly, L Bernstein, A Sludds, M Soljačić… - Physical Review X, 2019 - APS
Recent success in deep neural networks has generated strong interest in hardware
accelerators to improve speed and energy consumption. This paper presents a new type of …

Sparse reram engine: Joint exploration of activation and weight sparsity in compressed neural networks

TH Yang, HY Cheng, CL Yang, IC Tseng… - Proceedings of the 46th …, 2019 - dl.acm.org
Exploiting model sparsity to reduce ineffectual computation is a commonly used approach to
achieve energy efficiency for DNN inference accelerators. However, due to the tightly …