Recent image generation models such as Stable Diffusion have exhibited an impressive ability to generate fairly realistic images starting from a simple text prompt. Could such …
The rapid progress in Large Language Models (LLMs) could transform many fields, but their fast development creates significant challenges for oversight, ethical creation, and building …
Z Zhang, X Zhang, Y Ren, S Shi, M Han… - Proceedings of the …, 2023 - aclanthology.org
Abstract Retrieval-Augmented Generation (RAG), by incorporating external knowledge with parametric memory of language models, has become the state-of-the-art architecture for …
G Kim, D Jang, E Yang - arXiv preprint arXiv:2402.12842, 2024 - arxiv.org
Recent advancements in large language models (LLMs) have raised concerns about inference costs, increasing the need for research into model compression. While knowledge …
Z Bai, X Liu, H Hu, T Guo, Q Zhang, Y Wang - arXiv preprint arXiv …, 2023 - arxiv.org
Data-Free Knowledge Distillation (DFKD) plays a vital role in compressing the model when original training data is unavailable. Previous works for DFKD in NLP mainly focus on …
P Vijayaraghavan, H Wang, L Shi, T Baldwin… - arXiv preprint arXiv …, 2024 - arxiv.org
Recently, there has been a growing availability of pre-trained text models on various model repositories. These models greatly reduce the cost of training new models from scratch as …