L Zhao,
W Zeng, X Shi, H Zhou - arXiv preprint arXiv:2412.08946, 2024 - arxiv.org
Recently, LoRA has emerged as a crucial technique for fine-tuning large pre-trained models,
yet its performance in multi-task learning scenarios often falls short. In contrast, the MoE …