关注
Haoran He
标题
引用次数
引用次数
年份
Diffusion Model is an Effective Planner and Data Synthesizer for Multi-Task Reinforcement Learning
H He, C Bai, K Xu, Z Yang, W Zhang, D Wang, B Zhao, X Li
Neural Information Processing Systems (NeurIPS), 2023
292023
Diffusion models for reinforcement learning: A survey
Z Zhu, H Zhao, H He, Y Zhong, S Zhang, Y Yu, W Zhang
arXiv preprint arXiv:2311.01223, 2023
122023
Robust Quadrupedal Locomotion via Risk-Averse Policy Learning
J Shi, C Bai, H He, L Han, D Wang, B Zhao, X Li, X Li
IEEE International Conference on Robotics and Automation (ICRA), 2024
62024
Large-scale actionless video pre-training via discrete diffusion for efficient policy learning
H He, C Bai, L Pan, W Zhang, B Zhao, X Li
arXiv preprint arXiv:2402.14407, 2024
42024
On the value of myopic behavior in policy reuse
K Xu, C Bai, S Qiu, H He, B Zhao, Z Wang, W Li, X Li
arXiv preprint arXiv:2305.17623, 2023
22023
Looking Backward: Retrospective Backward Synthesis for Goal-Conditioned GFlowNets
H He, C Chang, H Xu, L Pan
arXiv preprint arXiv:2406.01150, 2024
12024
Privileged Knowledge Distillation for Sim-to-Real Policy Generalization
H He, C Bai, H Lai, L Wang, W Zhang
arXiv preprint arXiv:2305.18464, 2023
12023
Rectifying Reinforcement Learning for Reward Matching
H He, E Bengio, Q Cai, L Pan
arXiv preprint arXiv:2406.02213, 2024
2024
SAM-E: Leveraging Visual Foundation Model with Sequence Imitation for Embodied Manipulation
J Zhang, C Bai, H He, W Xia, Z Wang, B Zhao, X Li, X Li
International Conference on Machine Learning (ICML), 2024
2024
Regularized Conditional Diffusion Model for Multi-Task Preference Alignment
X Yu, C Bai, H He, C Wang, X Li
arXiv preprint arXiv:2404.04920, 2024
2024
系统目前无法执行此操作,请稍后再试。
文章 1–10