Stannis: low-power acceleration of dnn training using computational storage devices

A HeydariGorji, M Torabzadehkashi… - 2020 57th ACM/IEEE …, 2020 - ieeexplore.ieee.org
Computational storage devices enable in-storage processing of data in place. These
devices contain 64-bit application processors and hardware accelerators that can help …

Quantifying and improving performance of distributed deep learning with cloud storage

N Krichevsky, R St Louis, T Guo - 2021 IEEE International …, 2021 - ieeexplore.ieee.org
Cloud computing provides a powerful yet low-cost environment for distributed deep learning
workloads. However, training complex deep learning models often requires accessing large …

{FlashNeuron}:{SSD-Enabled}{Large-Batch} training of very deep neural networks

J Bae, J Lee, Y Jin, S Son, S Kim, H Jang… - … USENIX Conference on …, 2021 - usenix.org
Deep neural networks (DNNs) are widely used in various AI application domains such as
computer vision, natural language processing, autonomous driving, and bioinformatics. As …

Efficient user-level storage disaggregation for deep learning

Y Zhu, W Yu, B Jiao, K Mohror, A Moody… - 2019 IEEE …, 2019 - ieeexplore.ieee.org
On large-scale high performance computing (HPC) systems, applications are provisioned
with aggregated resources to meet their peak demands for brief periods. This results in …

PRINS: Processing-in-storage acceleration of machine learning

R Kaplan, L Yavits, R Ginosar - IEEE Transactions on …, 2018 - ieeexplore.ieee.org
Machine learning algorithms have become a major tool in various applications. The high-
performance requirements on large-scale datasets pose a challenge for traditional von …

Deep partitioned training from near-storage computing to DNN accelerators

Y Jang, S Kim, D Kim, S Lee… - IEEE Computer …, 2021 - ieeexplore.ieee.org
In this letter, we present deep partitioned training to accelerate computations involved in
training DNN models. This is the first work that partitions a DNN model across storage …

The Case for Storage Optimization Decoupling in Deep Learning Frameworks

R Macedo, C Correia, M Dantas, C Brito… - 2021 IEEE …, 2021 - ieeexplore.ieee.org
Deep Learning (DL) training requires efficient access to large collections of data, leading DL
frameworks to implement individual I/O optimizations to take full advantage of storage …

Data-aware storage tiering for deep learning

C Xu, S Bhattacharya, M Foltin, S Byna… - 2021 IEEE/ACM …, 2021 - ieeexplore.ieee.org
DNN models trained with very large datasets can perform rich deep learning tasks with high
accuracy. However, feeding huge volumes of training data exerts significant pressure on IO …

Accelerating deep learning training through transparent storage tiering

M Dantas, D Leitão, P Cui, R Macedo… - 2022 22nd IEEE …, 2022 - ieeexplore.ieee.org
We present Monarch, a framework-agnostic storage middleware that transparently employs
storage tiering to accelerate Deep Learning (DL) training. It leverages existing storage tiers …

A holistic approach to data access for cloud-native analytics and machine learning

P Koutsovasilis, S Venugopal… - 2021 IEEE 14th …, 2021 - ieeexplore.ieee.org
Cloud providers offer a variety of storage solutions for hosting data, both in price and in
performance. For Analytics and machine learning applications, object storage services are …