IMCA: An efficient in-memory convolution accelerator

HE Yantır, AM Eltawil… - IEEE Transactions on Very …, 2021 - ieeexplore.ieee.org
IEEE Transactions on Very Large Scale Integration (VLSI) Systems, 2021ieeexplore.ieee.org
Traditional convolutional neural network (CNN) architectures suffer from two bottlenecks:
computational complexity and memory access cost. In this study, an efficient in-memory
convolution accelerator (IMCA) is proposed based on associative in-memory processing to
alleviate these two problems directly. In the IMCA, the convolution operations are directly
performed inside the memory as in-place operations. The proposed memory computational
structure allows for a significant improvement in computational metrics, namely, TOPS/W …
Traditional convolutional neural network (CNN) architectures suffer from two bottlenecks: computational complexity and memory access cost. In this study, an efficient in-memory convolution accelerator (IMCA) is proposed based on associative in-memory processing to alleviate these two problems directly. In the IMCA, the convolution operations are directly performed inside the memory as in-place operations. The proposed memory computational structure allows for a significant improvement in computational metrics, namely, TOPS/W. Furthermore, due to its unconventional computation style, the IMCA can take advantage of many potential opportunities, such as constant multiplication, bit-level sparsity, and dynamic approximate computing, which, while supported by traditional architectures, require extra overhead to exploit, thus reducing any potential gains. The proposed accelerator architecture exhibits a significant efficiency in terms of area and performance, achieving around 0.65 GOPS and 1.64 TOPS/W at 16-bit fixed-point precision with an area less than 0.25 mm 2 .
ieeexplore.ieee.org
以上显示的是最相近的搜索结果。 查看全部搜索结果