Training dnns with hybrid block floating point

M Drumond, T Lin, M Jaggi… - Advances in Neural …, 2018 - proceedings.neurips.cc
The wide adoption of DNNs has given birth to unrelenting computing requirements, forcing
datacenter operators to adopt domain-specific accelerators to train them. These accelerators …

FlexBlock: A flexible DNN training accelerator with multi-mode block floating point support

SH Noh, J Koo, S Lee, J Park… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
When training deep neural networks (DNNs), expensive floating point arithmetic units are
used in GPUs or custom neural processing units (NPUs). To reduce the burden of floating …

Fast: Dnn training under variable precision block floating point with stochastic rounding

SQ Zhang, B McDanel, HT Kung - 2022 IEEE International …, 2022 - ieeexplore.ieee.org
Block Floating Point (BFP) can efficiently support quantization for Deep Neural Network
(DNN) training by providing a wide dynamic range via a shared exponent across a group of …

Hybrid 8-bit floating point (HFP8) training and inference for deep neural networks

X Sun, J Choi, CY Chen, N Wang… - Advances in neural …, 2019 - proceedings.neurips.cc
Reducing the numerical precision of data and computation is extremely effective in
accelerating deep learning training workloads. Towards this end, 8-bit floating point …

A block minifloat representation for training deep neural networks

S Fox, S Rasoulinezhad, J Faraone… - … Conference on Learning …, 2020 - openreview.net
Training Deep Neural Networks (DNN) with high efficiency can be difficult to achieve with
native floating-point representations and commercially available hardware. Specialized …

Training deep neural networks with 8-bit floating point numbers

N Wang, J Choi, D Brand, CY Chen… - Advances in neural …, 2018 - proceedings.neurips.cc
The state-of-the-art hardware platforms for training deep neural networks are moving from
traditional single precision (32-bit) computations towards 16 bits of precision-in large part …

DLFloat: A 16-b floating point format designed for deep learning training and inference

A Agrawal, SM Mueller, BM Fleischer… - 2019 IEEE 26th …, 2019 - ieeexplore.ieee.org
The resilience of Deep Learning (DL) training and inference workloads to low-precision
computations, coupled with the demand for power-and area-efficient hardware accelerators …

Rapidnn: In-memory deep neural network acceleration framework

M Imani, M Samragh, Y Kim, S Gupta… - arXiv preprint arXiv …, 2018 - arxiv.org
Deep neural networks (DNN) have demonstrated effectiveness for various applications such
as image processing, video segmentation, and speech recognition. Running state-of-the-art …

Training high-performance and large-scale deep neural networks with full 8-bit integers

Y Yang, L Deng, S Wu, T Yan, Y Xie, G Li - Neural Networks, 2020 - Elsevier
Deep neural network (DNN) quantization converting floating-point (FP) data in the network
to integers (INT) is an effective way to shrink the model size for memory saving and simplify …

Hardware-software codesign of accurate, multiplier-free deep neural networks

H Tann, S Hashemi, RI Bahar, S Reda - Proceedings of the 54th Annual …, 2017 - dl.acm.org
While Deep Neural Networks (DNNs) push the state-of-the-art in many machine learning
applications, they often require millions of expensive floating-point operations for each input …