Internlm2 technical report Z Cai, M Cao, H Chen, K Chen, K Chen, X Chen, X Chen, Z Chen, Z Chen, ... arXiv preprint arXiv:2403.17297, 2024 | 192 | 2024 |
Internlm: A multilingual language model with progressively enhanced capabilities ILM Team 2023-01-06)[2023-09-27]. https://github. com/InternLM/InternLM, 2023 | 189 | 2023 |
Internlm-math: Open math large language models toward verifiable reasoning H Ying, S Zhang, L Li, Z Zhou, Y Shao, Z Fei, Y Ma, J Hong, K Liu, Z Wang, ... arXiv preprint arXiv:2402.06332, 2024 | 38 | 2024 |
Scaling laws of rope-based extrapolation X Liu, H Yan, S Zhang, C An, X Qiu, D Lin arXiv preprint arXiv:2310.05209, 2023 | 12 | 2023 |
Collie: Collaborative training of large language models in an efficient way K Lv, S Zhang, T Gu, S Xing, J Hong, K Chen, X Liu, Y Yang, H Guo, T Liu, ... Proceedings of the 2023 Conference on Empirical Methods in Natural Language …, 2023 | 6 | 2023 |
Efficient training of large language models on distributed infrastructures: A survey J Duan, S Zhang, Z Wang, L Jiang, W Qu, Q Hu, G Wang, Q Weng, H Yan, ... arXiv preprint arXiv:2407.20018, 2024 | 4 | 2024 |