Y Wang, D Deng, L Liu, S Wei… - IEEE Transactions on …, 2022 - ieeexplore.ieee.org
Edge device deep neural network (DNN) training is practical to improve model adaptivity for unfamiliar datasets while avoiding privacy disclosure and huge communication cost …
A Boumendil, W Bechkit… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Providing high-quality predictions is no longer the sole goal for neural networks. As we live in an increasingly interconnected world, these models need to match the constraints of …
Recently, low bit-width quantization (eg, INT8) has been commonly used in deep neural network inference acceleration, but fewer researchers have focused on low-precision …
S Lee, J Park, D Jeon - … 2023-IEEE 49th European Solid State …, 2023 - ieeexplore.ieee.org
This paper presents an energy-efficient FP4/FP8 hybrid-precision training processor. Through hardware-software co-optimization, the design efficiently implements all general …
Y Zhong, G Nan, Y Zhang, F Chao, R Ji - arXiv preprint arXiv:2211.08544, 2022 - arxiv.org
Quantization-aware training (QAT) receives extensive popularity as it well retains the performance of quantized networks. In QAT, the contemporary experience is that all …
Continual learning is a desirable feature in many modern machine learning applications, which allows in-field adaptation and updating, ranging from accommodating distribution …
A Shen, Z Lai, L Zhang - Applied Sciences (2076-3417), 2024 - search.ebscohost.com
Abstract As Deep Neural Networks (DNNs) continue to increase in complexity, the computational demands of their training have become a significant bottleneck. Low …
Quantization-aware training (QAT) has been demonstrated to not only reduce computational cost and storage footprint, but well retain the performance of full-precision neural networks …
Foundational to the digital economy, data and its accompanying analytical models today play a critical role in the gaining of new insights from the digital world and the facilitation of …