PyDTNN: a user-friendly and extensible framework for distributed deep learning

S Barrachina, A Castelló, M Catalán, MF Dolz… - The Journal of …, 2021 - Springer
We introduce a framework for training deep neural networks on clusters of computers with
the following appealing properties:(1) It is developed in Python, exposing an amiable …

FSP: towards flexible synchronous parallel frameworks for distributed machine learning

Z Wang, Y Tu, N Wang, L Gao, J Nie… - … on Parallel and …, 2022 - ieeexplore.ieee.org
Myriad of machine learning (ML) algorithms refine model parameters iteratively. Existing
synchronous data-parallel frameworks can accelerate training with convergence …

Prototyping a GPGPU neural network for deep-learning big data analysis

A Fonseca, B Cabral - Big Data Research, 2017 - Elsevier
Big Data concerns with large-volume complex growing data. Given the fast development of
data storage and network, organizations are collecting large ever-growing datasets that can …

Parallelization of data science tasks, an experimental overview

O Castro, P Bruneau, JS Sottet… - Proceedings of the 2022 …, 2022 - dl.acm.org
The practice of data science and machine learning often involves training many kinds of
models, for inferring some target variable, or extracting structured knowledge from data …

[PDF][PDF] Parallel machine learning algorithms

SA Salman, SA Dheyab… - … Journal of Big …, 2023 - journals.mesopotamian.press
Salman et al, Mesopotamian Journal of Big Data Vol. (2023), 2023, 12–15 Page 1
Research Article Parallel Machine Learning Algorithms Saba Abdulbaqi Salman*1,, Saad …

[PDF][PDF] Black-box parallelization for machine learning

M Kamp - 2019 - core.ac.uk
The landscape of machine learning applications is changing rapidly: large centralized
datasets are replaced by high volume, high velocity data streams generated by a vast …

Tell me something new: A new framework for asynchronous parallel learning

J Alafate, Y Freund - arXiv preprint arXiv:1805.07483, 2018 - arxiv.org
We present a novel approach for parallel computation in the context of machine learning that
we call" Tell Me Something New"(TMSN). This approach involves a set of independent …

Strads: A distributed framework for scheduled model parallel machine learning

JK Kim, Q Ho, S Lee, X Zheng, W Dai… - Proceedings of the …, 2016 - dl.acm.org
Machine learning (ML) algorithms are commonly applied to big data, using distributed
systems that partition the data across machines and allow each machine to read and update …

[PDF][PDF] Parallel computing for artificial neural network training

O Gursoy, H Sharif - Periodicals of Engineering and Natural …, 2018 - pen.ius.edu.ba
Parallel Computing for Artificial Neural Network Training using Java Native Socket
Programming Page 1 Periodicals of Engineering and Natural Sciences Vol. 6, No. 1 …

Model Parallelism on Distributed Infrastructure: A Literature Review from Theory to LLM Case-Studies

F Brakel, U Odyurt, AL Varbanescu - arXiv preprint arXiv:2403.03699, 2024 - arxiv.org
Neural networks have become a cornerstone of machine learning. As the trend for these to
get more and more complex continues, so does the underlying hardware and software …