关注
Hao Ge
Hao Ge
在 stu.pku.edu.cn 的电子邮件经过验证
标题
引用次数
引用次数
年份
Enabling Parallelism Hot Switching for Efficient Training of Large Language Models
H Ge, F Fu, H Li, X Wang, S Lin, Y Wang, X Nie, H Zhang, X Miao, B Cui
Proceedings of the ACM SIGOPS 30th Symposium on Operating Systems Principles …, 2024
12024
Demystifying Workload Imbalances in Large Transformer Model Training over Variable-length Sequences
H Li, F Fu, S Lin, H Ge, X Wang, J Niu, J Jiang, B Cui
arXiv preprint arXiv:2412.07894, 2024
2024
Malleus: Straggler-Resilient Hybrid Parallel Training of Large-scale Models via Malleable Data and Model Parallelization
H Li, F Fu, H Ge, S Lin, X Wang, J Niu, Y Wang, H Zhang, X Nie, B Cui
arXiv preprint arXiv:2410.13333, 2024
2024
系统目前无法执行此操作,请稍后再试。
文章 1–3