AR Junaid - Authorea Preprints, 2025 - techrxiv.org
Large language models (LLMs) have revolutionized the field of artificial intelligence, achieving unprecedented performance in tasks such as text generation, translation, and …
Despite the advanced intelligence abilities of large language models (LLMs) in various applications, they still face significant computational and storage demands. Knowledge …
C Yang, Y Zhu, W Lu, Y Wang, Q Chen, C Gao… - ACM Transactions on …, 2024 - dl.acm.org
Large Language Models (LLMs) have showcased exceptional capabilities in various domains, attracting significant interest from both academia and industry. Despite their …
X Zhu, J Li, C Ma, W Wang - arXiv preprint arXiv:2411.14698, 2024 - arxiv.org
Large Language Models (LLMs) demonstrate exceptional reasoning capabilities, often achieving state-of-the-art performance in various tasks. However, their substantial …
This survey presents an in-depth exploration of knowledge distillation (KD) techniques within the realm of Large Language Models (LLMs), spotlighting the pivotal role of KD in …
J Chen, T Wu, W Ji, F Wu - Frontiers of Digital Education, 2024 - Springer
Large language models (LLMs) have emerged as powerful tools in natural language processing (NLP), showing a promising future of artificial generated intelligence (AGI) …
H Xu, H Liu, W Gong, X Deng, H Wang - CCF International Conference on …, 2024 - Springer
Abstract Knowledge distillation is an effective method for reducing the computational overhead of large language models. However, recent optimization efforts in distilling large …
While large language models (LLMs) have demonstrated exceptional performance in recent natural language processing (NLP) tasks, their deployment poses substantial challenges …
S Tan, WL Tam, Y Wang, W Gong, Y Yang… - arXiv preprint arXiv …, 2023 - arxiv.org
Currently, the reduction in the parameter scale of large-scale pre-trained language models (PLMs) through knowledge distillation has greatly facilitated their widespread deployment …