Demystifying learning rate policies for high accuracy training of deep neural networks

Y Wu, L Liu, J Bae, KH Chow, A Iyengar… - … conference on big …, 2019 - ieeexplore.ieee.org
Learning Rate (LR) is an important hyper-parameter to tune for effective training of deep
neural networks (DNNs). Even for the baseline of a constant learning rate, it is non-trivial to …

An overview of the data-loader landscape: Comparative performance analysis

I Ofeidis, D Kiedanski, L Tassiulas - arXiv preprint arXiv:2209.13705, 2022 - arxiv.org
Dataloaders, in charge of moving data from storage into GPUs while training machine
learning models, might hold the key to drastically improving the performance of training jobs …

Boosting ensemble accuracy by revisiting ensemble diversity metrics

Y Wu, L Liu, Z Xie, KH Chow… - Proceedings of the IEEE …, 2021 - openaccess.thecvf.com
Neural network ensembles are gaining popularity by harnessing the complementary wisdom
of multiple base models. Ensemble teams with high diversity promote high failure …

A survey of deep learning on cpus: opportunities and co-optimizations

S Mittal, P Rajput, S Subramoney - IEEE Transactions on …, 2021 - ieeexplore.ieee.org
CPU is a powerful, pervasive, and indispensable platform for running deep learning (DL)
workloads in systems ranging from mobile to extreme-end servers. In this article, we present …

Reveal training performance mystery between TensorFlow and PyTorch in the single GPU environment

H Dai, X Peng, X Shi, L He, Q Xiong, H Jin - Science China Information …, 2022 - Springer
Deep learning has gained tremendous success in various fields while training deep neural
networks (DNNs) is very compute-intensive, which results in numerous deep learning …

Selecting and composing learning rate policies for deep neural networks

Y Wu, L Liu - ACM Transactions on Intelligent Systems and …, 2023 - dl.acm.org
The choice of learning rate (LR) functions and policies has evolved from a simple fixed LR to
the decaying LR and the cyclic LR, aiming to improve the accuracy and reduce the training …

Rethinking learning rate tuning in the era of large language models

H Jin, W Wei, X Wang, W Zhang… - 2023 IEEE 5th …, 2023 - ieeexplore.ieee.org
Large Language Models (LLMs) represent the recent success of deep learning in achieving
remarkable human-like predictive performance. It has become a mainstream strategy to …

A pipeline computing method of SpTV for three-order tensors on CPU and GPU

W Yang, K Li, K Li - ACM Transactions on Knowledge Discovery from …, 2019 - dl.acm.org
Tensors have drawn a growing attention in many applications, such as physics, engineering
science, social networks, recommended systems. Tensor decomposition is the key to …

Promoting high diversity ensemble learning with ensemblebench

Y Wu, L Liu, Z Xie, J Bae, KH Chow… - 2020 IEEE Second …, 2020 - ieeexplore.ieee.org
Ensemble learning is gaining renewed interests in recent years. This paper presents
EnsembleBench, a holistic framework for evaluating and recommending high diversity and …

Quantitative evaluation of deep learning frameworks in heterogeneous computing environment

Z Lu, C Du, Y Jiang, X Xie, T Li, F Yang - CCF Transactions on High …, 2024 - Springer
Deep learning frameworks are powerful tools to support model training. They dispatch
operators by mapping them into a series of kernel functions and launching these kernel …