FLIQS: One-Shot Mixed-Precision Floating-Point and Integer Quantization Search

J Dotzel, G Wu, A Li, M Umar, Y Ni… - arXiv preprint arXiv …, 2023 - arxiv.org
Quantization has become a mainstream compression technique for reducing model size,
computational requirements, and energy consumption for modern deep neural networks …

PL-NPU: An energy-efficient edge-device DNN training processor with posit-based logarithm-domain computing

Y Wang, D Deng, L Liu, S Wei… - IEEE Transactions on …, 2022 - ieeexplore.ieee.org
Edge device deep neural network (DNN) training is practical to improve model adaptivity for
unfamiliar datasets while avoiding privacy disclosure and huge communication cost …

On-device deep learning: survey on techniques improving energy efficiency of DNNs

A Boumendil, W Bechkit… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Providing high-quality predictions is no longer the sole goal for neural networks. As we live
in an increasingly interconnected world, these models need to match the constraints of …

Gradient distribution-aware INT8 training for neural networks

S Wang, Y Kang - Neurocomputing, 2023 - Elsevier
Recently, low bit-width quantization (eg, INT8) has been commonly used in deep neural
network inference acceleration, but fewer researchers have focused on low-precision …

A 4.27 TFLOPS/W FP4/FP8 Hybrid-Precision Neural Network Training Processor Using Shift-Add MAC and Reconfigurable PE Array

S Lee, J Park, D Jeon - … 2023-IEEE 49th European Solid State …, 2023 - ieeexplore.ieee.org
This paper presents an energy-efficient FP4/FP8 hybrid-precision training processor.
Through hardware-software co-optimization, the design efficiently implements all general …

Exploiting the Partly Scratch-off Lottery Ticket for Quantization-Aware Training

Y Zhong, G Nan, Y Zhang, F Chao, R Ji - arXiv preprint arXiv:2211.08544, 2022 - arxiv.org
Quantization-aware training (QAT) receives extensive popularity as it well retains the
performance of quantized networks. In QAT, the contemporary experience is that all …

Hadamard Domain Training with Integers for Class Incremental Quantized Learning

M Schiemer, CJS Schaefer, JP Vap, MJ Horeni… - arXiv preprint arXiv …, 2023 - arxiv.org
Continual learning is a desirable feature in many modern machine learning applications,
which allows in-field adaptation and updating, ranging from accommodating distribution …

Systematic Analysis of Low-Precision Training in Deep Neural Networks: Factors Influencing Matrix Computations.

A Shen, Z Lai, L Zhang - Applied Sciences (2076-3417), 2024 - search.ebscohost.com
Abstract As Deep Neural Networks (DNNs) continue to increase in complexity, the
computational demands of their training have become a significant bottleneck. Low …

SpQAT: A Sparse Quantization-Aware Training Method

Y Zhong, M Lin, G Nan, F Chao, R Ji - openreview.net
Quantization-aware training (QAT) has been demonstrated to not only reduce computational
cost and storage footprint, but well retain the performance of full-precision neural networks …

[引用][C] Multi-modal data modeling with awareness of efficiency, reliability, and privacy

P Zhang - 2023 - espace.library.uq.edu.au
Foundational to the digital economy, data and its accompanying analytical models today
play a critical role in the gaining of new insights from the digital world and the facilitation of …