关注
Aviral Kumar
Aviral Kumar
Google DeepMind
在 berkeley.edu 的电子邮件经过验证 - 首页
标题
引用次数
年份
A workflow for offline model-free robotic reinforcement learning
A Kumar, A Singh, S Tian, C Finn, S Levine
arXiv preprint arXiv:2109.10813, 2021
842021
Action-quantized offline reinforcement learning for robotic skill learning
J Luo, P Dong, J Wu, A Kumar, X Geng, S Levine
Conference on Robot Learning, 1348-1361, 2023
112023
Advantage-weighted regression: Simple and scalable off-policy reinforcement learning
XB Peng, A Kumar, G Zhang, S Levine
arXiv preprint arXiv:1910.00177, 2019
4602019
ArCHer: Training Language Model Agents via Hierarchical Multi-Turn RL
Y Zhou, A Zanette, J Pan, S Levine, A Kumar
arXiv preprint arXiv:2402.19446, 2024
62024
Benchmarks for deep off-policy evaluation
J Fu, M Norouzi, O Nachum, G Tucker, Z Wang, A Novikov, M Yang, ...
arXiv preprint arXiv:2103.16596, 2021
792021
Beyond uniform sampling: Offline reinforcement learning with imbalanced datasets
ZW Hong, A Kumar, S Karnik, A Bhandwaldar, A Srivastava, J Pajarinen, ...
Advances in Neural Information Processing Systems 36, 4985-5009, 2023
62023
Cal-ql: Calibrated offline rl pre-training for efficient online fine-tuning
M Nakamoto, S Zhai, A Singh, M Sobol Mark, Y Ma, C Finn, A Kumar, ...
Advances in Neural Information Processing Systems 36, 2024
572024
Calibration of Encoder Decoder Models for Neural Machine Translation
A Kumar, S Sarawagi
https://arxiv.org/abs/1903.00802, 2019
902019
Challenges and tool implementation of hybrid rapidly-exploring random trees
S Bak, S Bogomolov, TA Henzinger, A Kumar
Numerical Software Verification: 10th International Workshop, NSV 2017 …, 2017
42017
Challenges and Tool Implementation of Hybrid Rapidly-Exploring Random Trees (RRTs)
S Bak, S Bogomolov, TA Henzinger, A Kumar
Cog: Connecting new skills to past experience with offline reinforcement learning
A Singh, A Yu, J Yang, J Zhang, A Kumar, S Levine
arXiv preprint arXiv:2010.14500, 2020
982020
Combo: Conservative offline model-based policy optimization
T Yu, A Kumar, R Rafailov, A Rajeswaran, S Levine, C Finn
Advances in neural information processing systems 34, 28954-28967, 2021
3562021
Confidence-conditioned value functions for offline reinforcement learning
J Hong, A Kumar, S Levine
arXiv preprint arXiv:2212.04607, 2022
182022
Conservative data sharing for multi-task offline reinforcement learning
T Yu, A Kumar, Y Chebotar, K Hausman, S Levine, C Finn
Advances in Neural Information Processing Systems 34, 11501-11516, 2021
752021
Conservative objective models for effective offline model-based optimization
B Trabucco, A Kumar, X Geng, S Levine
International Conference on Machine Learning, 10358-10368, 2021
812021
Conservative q-learning for offline reinforcement learning
A Kumar, A Zhou, G Tucker, S Levine
Advances in Neural Information Processing Systems 33, 1179-1191, 2020
16102020
Conservative safety critics for exploration
H Bharadhwaj, A Kumar, N Rhinehart, S Levine, F Shkurti, A Garg
arXiv preprint arXiv:2010.14497, 2020
1322020
D4rl: Datasets for deep data-driven reinforcement learning
J Fu, A Kumar, O Nachum, G Tucker, S Levine
arXiv preprint arXiv:2004.07219, 2020
9932020
Data-driven deep reinforcement learning
A Kumar
Berkeley Artificial Intelligence Research (BAIR), Tech. Rep, 2019
182019
Data-driven offline decision-making via invariant representation learning
H Qi, Y Su, A Kumar, S Levine
Advances in Neural Information Processing Systems 35, 13226-13237, 2022
142022
系统目前无法执行此操作,请稍后再试。
文章 1–20