Quantization-error-robust deep neural network for embedded accelerators

Y Jung, H Kim, Y Choi, LS Kim - IEEE Transactions on Circuits …, 2021 - ieeexplore.ieee.org
Y Jung, H Kim, Y Choi, LS Kim
IEEE Transactions on Circuits and Systems II: Express Briefs, 2021ieeexplore.ieee.org
Quantization with low precision has become an essential technique for adopting deep
neural networks in energy-and memory-constrained devices. However, there is a limit to the
reducing precision by the inevitable loss of accuracy due to the quantization error. To
overcome this obstacle, we propose methods reforming and quantizing a network that
achieves high accuracy even at low precision without any runtime overhead in embedded
accelerators. Our proposition consists of two analytical approaches: 1) network optimization …
Quantization with low precision has become an essential technique for adopting deep neural networks in energy- and memory-constrained devices. However, there is a limit to the reducing precision by the inevitable loss of accuracy due to the quantization error. To overcome this obstacle, we propose methods reforming and quantizing a network that achieves high accuracy even at low precision without any runtime overhead in embedded accelerators. Our proposition consists of two analytical approaches: 1) network optimization to find the most error-resilient equivalent network in the precision-constrained environment and 2) quantization exploiting adaptive rounding offset control. The experimental results show accuracies of up to 98.31% and 99.96% of floating-point results in 6-bit and 8-bit quantization networks, respectively. Besides, our methods allow the lower precision accelerator design, reducing the energy consumption by 8.5%.
ieeexplore.ieee.org
以上显示的是最相近的搜索结果。 查看全部搜索结果