High occupational injury and fatality rate in the construction industry is a serious global concern. Recognizing AI as a solution to enhance safety performance, this study reviews …
Recent work on pruning large language models (LLMs) has shown that one can eliminate a large number of parameters without compromising performance, making pruning a …
Selecting suitable architecture parameters and training hyperparameters is essential for enhancing machine learning (ML) model performance. Several recent empirical studies …
Parameter-efficient fine-tuning methods, such as Low-Rank Adaptation (LoRA), are known to enhance training efficiency in Large Language Models (LLMs). Due to the limited …
Recent advances in foundation models have emphasized the need to align pre-trained models with specialized domains using small, curated datasets. Studies on these foundation …
The paper introduces LEMR, a framework that reduces annotation costs for model selection tasks. Our approach leverages ensemble methods to generate pseudo-labels, employs …
The now-globally recognized concerns of AI's environmental implications resulted in a growing awareness of the need to reduce AI carbon footprints, as well as to carry out AI …
P Li, L Yin, X Gao, S Liu - arXiv preprint arXiv:2405.18380, 2024 - arxiv.org
The rapid advancements in Large Language Models (LLMs) have revolutionized various natural language processing tasks. However, the substantial size of LLMs presents …
Modern training strategies of deep neural networks (NNs) tend to induce a heavy-tailed (HT) spectra of layer weights. Extensive efforts to study this phenomenon have found that NNs …