A survey on model compression for large language models

X Zhu, J Li, Y Liu, C Ma, W Wang - Transactions of the Association for …, 2024 - direct.mit.edu
Abstract Large Language Models (LLMs) have transformed natural language processing
tasks successfully. Yet, their large size and high computational needs pose challenges for …

Llm inference unveiled: Survey and roofline model insights

Z Yuan, Y Shang, Y Zhou, Z Dong, Z Zhou… - arXiv preprint arXiv …, 2024 - arxiv.org
The field of efficient Large Language Model (LLM) inference is rapidly evolving, presenting a
unique blend of opportunities and challenges. Although the field has expanded and is …

A survey of low-bit large language models: Basics, systems, and algorithms

R Gong, Y Ding, Z Wang, C Lv, X Zheng, J Du… - arXiv preprint arXiv …, 2024 - arxiv.org
Large language models (LLMs) have achieved remarkable advancements in natural
language processing, showcasing exceptional performance across various tasks. However …

A survey on efficient inference for large language models

Z Zhou, X Ning, K Hong, T Fu, J Xu, S Li, Y Lou… - arXiv preprint arXiv …, 2024 - arxiv.org
Large Language Models (LLMs) have attracted extensive attention due to their remarkable
performance across various tasks. However, the substantial computational and memory …

Think: Thinner key cache by query-driven pruning

Y Xu, Z Jie, H Dong, L Wang, X Lu, A Zhou… - arXiv preprint arXiv …, 2024 - arxiv.org
Large Language Models (LLMs) have revolutionized the field of natural language
processing, achieving unprecedented performance across a variety of applications …

[HTML][HTML] The Future of Intelligent Healthcare: A Systematic Analysis and Discussion on the Integration and Impact of Robots Using Large Language Models for …

S Pashangpour, G Nejat - Robotics, 2024 - mdpi.com
The potential use of large language models (LLMs) in healthcare robotics can help address
the significant demand put on healthcare systems around the world with respect to an aging …

Dynamic memory compression: Retrofitting llms for accelerated inference

P Nawrot, A Łańcucki, M Chochowski, D Tarjan… - arXiv preprint arXiv …, 2024 - arxiv.org
Transformers have emerged as the backbone of large language models (LLMs). However,
generation remains inefficient due to the need to store in memory a cache of key-value …

Beyond kv caching: Shared attention for efficient llms

B Liao, DV Vargas - arXiv preprint arXiv:2407.12866, 2024 - arxiv.org
The efficiency of large language models (LLMs) remains a critical challenge, particularly in
contexts where computational resources are limited. Traditional attention mechanisms in …

LLMServingSim: A HW/SW Co-Simulation Infrastructure for LLM Inference Serving at Scale

J Cho, M Kim, H Choi, G Heo… - 2024 IEEE International …, 2024 - ieeexplore.ieee.org
Recently, there has been an extensive research effort in building efficient large language
model (LLM) inference serving systems. These efforts not only include innovations in the …

Unifying kv cache compression for large language models with leankv

Y Zhang, Y Hu, R Zhao, J Lui, H Chen - arXiv preprint arXiv:2412.03131, 2024 - arxiv.org
Large language models (LLMs) demonstrate exceptional performance but incur high serving
costs due to substantial memory demands, with the key-value (KV) cache being a primary …