Accelerating distributed reinforcement learning with in-switch computing

Y Li, IJ Liu, Y Yuan, D Chen, A Schwing… - Proceedings of the 46th …, 2019 - dl.acm.org
Reinforcement learning (RL) has attracted much attention recently, as new and emerging AI-
based applications are demanding the capabilities to intelligently react to environment …

RLlib: Abstractions for distributed reinforcement learning

E Liang, R Liaw, R Nishihara, P Moritz… - International …, 2018 - proceedings.mlr.press
Reinforcement learning (RL) algorithms involve the deep nesting of highly irregular
computation patterns, each of which typically exhibits opportunities for distributed …

[PDF][PDF] Ray rllib: A composable and scalable reinforcement learning library

E Liang, R Liaw, R Nishihara, P Moritz, R Fox… - arXiv preprint arXiv …, 2017 - royf.org
Reinforcement learning (RL) algorithms involve the deep nesting of distinct components,
where each component typically exhibits opportunities for distributed computation. Current …

Acme: A research framework for distributed reinforcement learning

MW Hoffman, B Shahriari, J Aslanides… - arXiv preprint arXiv …, 2020 - arxiv.org
Deep reinforcement learning (RL) has led to many recent and groundbreaking advances.
However, these advances have often come at the cost of both increased scale in the …

Scaling distributed machine learning with {In-Network} aggregation

A Sapio, M Canini, CY Ho, J Nelson, P Kalnis… - … USENIX Symposium on …, 2021 - usenix.org
Training machine learning models in parallel is an increasingly important workload. We
accelerate distributed parallel training by designing a communication primitive that uses a …

GRID: Gradient routing with in-network aggregation for distributed training

J Fang, G Zhao, H Xu, C Wu… - IEEE/ACM Transactions on …, 2023 - ieeexplore.ieee.org
As the scale of distributed training increases, it brings huge communication overhead in
clusters. Some works try to reduce the communication cost through gradient compression or …

Efficient parallel methods for deep reinforcement learning

AV Clemente, HN Castejón, A Chandra - arXiv preprint arXiv:1705.04862, 2017 - arxiv.org
We propose a novel framework for efficient parallelization of deep reinforcement learning
algorithms, enabling these algorithms to learn from multiple actors on a single machine. The …

Gpu-accelerated robotic simulation for distributed reinforcement learning

J Liang, V Makoviychuk, A Handa… - … on Robot Learning, 2018 - proceedings.mlr.press
Abstract Most Deep Reinforcement Learning (Deep RL) algorithms require a prohibitively
large number of training samples for learning complex tasks. Many recent works on …

Envpool: A highly parallel reinforcement learning environment execution engine

J Weng, M Lin, S Huang, B Liu… - Advances in …, 2022 - proceedings.neurips.cc
There has been significant progress in developing reinforcement learning (RL) training
systems. Past works such as IMPALA, Apex, Seed RL, Sample Factory, and others, aim to …

Seed rl: Scalable and efficient deep-rl with accelerated central inference

L Espeholt, R Marinier, P Stanczyk, K Wang… - arXiv preprint arXiv …, 2019 - arxiv.org
We present a modern scalable reinforcement learning agent called SEED (Scalable,
Efficient Deep-RL). By effectively utilizing modern accelerators, we show that it is not only …