Space-efficient optical computing with an integrated chip diffractive neural network

HH Zhu, J Zou, H Zhang, YZ Shi, SB Luo… - Nature …, 2022 - nature.com
Large-scale, highly integrated and low-power-consuming hardware is becoming
progressively more important for realizing optical neural networks (ONNs) capable of …

CirCNN: accelerating and compressing deep neural networks using block-circulant weight matrices

C Ding, S Liao, Y Wang, Z Li, N Liu, Y Zhuo… - Proceedings of the 50th …, 2017 - dl.acm.org
Large-scale deep neural networks (DNNs) are both compute and memory intensive. As the
size of DNNs continues to grow, it is critical to improve the energy efficiency and …

Ftrans: energy-efficient acceleration of transformers using fpga

B Li, S Pandey, H Fang, Y Lyv, J Li, J Chen… - Proceedings of the …, 2020 - dl.acm.org
In natural language processing (NLP), the" Transformer" architecture was proposed as the
first transduction model replying entirely on self-attention mechanisms without using …

C-LSTM: Enabling efficient LSTM using structured compression techniques on FPGAs

S Wang, Z Li, C Ding, B Yuan, Q Qiu, Y Wang… - Proceedings of the …, 2018 - dl.acm.org
Recently, significant accuracy improvement has been achieved for acoustic recognition
systems by increasing the model size of Long Short-Term Memory (LSTM) networks …

REQ-YOLO: A resource-aware, efficient quantization framework for object detection on FPGAs

C Ding, S Wang, N Liu, K Xu, Y Wang… - proceedings of the 2019 …, 2019 - dl.acm.org
Deep neural networks (DNNs), as the basis of object detection, will play a key role in the
development of future autonomous systems with full autonomy. The autonomous systems …

E-RNN: Design optimization for efficient recurrent neural networks in FPGAs

Z Li, C Ding, S Wang, W Wen, Y Zhuo… - … Symposium on High …, 2019 - ieeexplore.ieee.org
Recurrent Neural Networks (RNNs) are becoming increasingly important for time series-
related applications which require efficient and real-time implementations. The two major …

Data-dependent coresets for compressing neural networks with applications to generalization bounds

C Baykal, L Liebenwein, I Gilitschenski… - arXiv preprint arXiv …, 2018 - arxiv.org
We present an efficient coresets-based neural network compression algorithm that sparsifies
the parameters of a trained fully-connected neural network in a manner that provably …

[PDF][PDF] Adam-admm: A unified, systematic framework of structured weight pruning for dnns

T Zhang, K Zhang, S Ye, J Li, J Tang… - arXiv preprint arXiv …, 2018 - yeshaokai.github.io
Weight pruning methods of deep neural networks (DNNs) have been demonstrated to
achieve a good model pruning ratio without loss of accuracy, thereby alleviating the …

Structadmm: Achieving ultrahigh efficiency in structured pruning for dnns

T Zhang, S Ye, X Feng, X Ma, K Zhang… - IEEE transactions on …, 2021 - ieeexplore.ieee.org
Weight pruning methods of deep neural networks (DNNs) have been demonstrated to
achieve a good model pruning rate without loss of accuracy, thereby alleviating the …

A unified framework of DNN weight pruning and weight clustering/quantization using ADMM

S Ye, T Zhang, K Zhang, J Li, J Xie, Y Liang… - arXiv preprint arXiv …, 2018 - arxiv.org
Many model compression techniques of Deep Neural Networks (DNNs) have been
investigated, including weight pruning, weight clustering and quantization, etc. Weight …