NeSSA: Near-storage data selection for accelerated machine learning training

N Prakriya, Y Yang, B Mirzasoleiman… - Proceedings of the 15th …, 2023 - dl.acm.org
Large-scale machine learning (ML) models rely on extremely large datasets to learn their
exponentially growing number of parameters. While these models achieve unprecedented …

Computational Storage for an Energy-Efficient Deep Neural Network Training System

S Li, K Tang, J Lim, CH Lee, J Kim - European Conference on Parallel …, 2023 - Springer
Near-storage data processing and computational storage have recently received
considerable attention from the industry as energy-and cost-efficient ways to improve system …

Diesel: A dataset-based distributed storage and caching system for large-scale deep learning training

L Wang, S Ye, B Yang, Y Lu, H Zhang, S Yan… - Proceedings of the 49th …, 2020 - dl.acm.org
We observe three problems in existing storage and caching systems for deep-learning
training (DLT) tasks:(1) accessing a dataset containing a large number of small files takes a …

Profiling and improving the pytorch dataloader for high-latency storage: A technical report

I Svogor, C Eichenberger, M Spanring, M Neun… - arXiv preprint arXiv …, 2022 - arxiv.org
A growing number of Machine Learning Frameworks recently made Deep Learning
accessible to a wider audience of engineers, scientists, and practitioners, by allowing …

{SHADE}: Enable Fundamental Cacheability for Distributed Deep Learning Training

RIS Khan, AH Yazdani, Y Fu, AK Paul, B Ji… - … USENIX Conference on …, 2023 - usenix.org
Deep learning training (DLT) applications exhibit unique I/O workload behaviors that pose
new challenges for storage system design. DLT is I/O intensive since data samples need to …

[PDF][PDF] I/o for deep learning at scale

Q Koziol - International Conference on Massive Storage Systems …, 2019 - indico.cern.ch
I/O for Deep Learning at Scale Page 1 I/O for Deep Learning at Scale Quincey Koziol Principal
Data Architect, NERSC koziol@lbl.gov IXPUG, September 24, 2019 Page 2 Acknowledgments …

PreCog: Near-Storage Accelerator for Heterogeneous CNN Inference

J An, E Aliaj, SW Jun - 2023 IEEE 34th International …, 2023 - ieeexplore.ieee.org
Computational Storage Devices (CSD) with near-storage acceleration is gaining popularity
for data-intensive applications, by moving power-efficient hardware acceleration closer to …

Near-data processing for differentiable machine learning models

H Choe, S Lee, H Nam, S Park, S Kim… - arXiv preprint arXiv …, 2016 - arxiv.org
Near-data processing (NDP) refers to augmenting memory or storage with processing
power. Despite its potential for acceleration computing and reducing power requirements …

OptimStore: In-storage optimization of large scale DNNs with on-die processing

J Kim, M Kang, Y Han, YG Kim… - 2023 IEEE International …, 2023 - ieeexplore.ieee.org
Training deep neural network (DNN) models is a resource-intensive, iterative process. For
this reason, nowadays, complex optimizers like Adam are widely adopted as it increases the …

Accelerating machine learning i/o by overlapping data staging and mini-batch generations

K Serizawa, O Tatebe - Proceedings of the 6th IEEE/ACM International …, 2019 - dl.acm.org
The training dataset used in deep neural networks (DNNs) keeps on increasing. When a
training dataset grows larger, the reading performance of such a large training dataset …