[PDF][PDF] Instruction Mining: Instruction Data Selection for Tuning Large Language Models

Y Cao, Y Kang, C Wang, L Sun - arXiv preprint …, 2023 - storage.prod.researchhub.com
Large language models typically undergo two training stages, pretraining and finetuning.
Despite that large-scale pretraining endows the model with strong capabilities to generate …

Cost-effective hyperparameter optimization for large language model generation inference

C Wang, X Liu, AH Awadallah - International Conference on …, 2023 - proceedings.mlr.press
Abstract Large Language Models (LLMs) have sparked significant interest in their
generative capabilities, leading to the development of various commercial applications. The …

Unleashing the power of data tsunami: A comprehensive survey on data assessment and selection for instruction tuning of language models

Y Qin, Y Yang, P Guo, G Li, H Shao, Y Shi, Z Xu… - arXiv preprint arXiv …, 2024 - arxiv.org
Instruction tuning plays a critical role in aligning large language models (LLMs) with human
preference. Despite the vast amount of open instruction datasets, naively training a LLM on …

Automl in the age of large language models: Current challenges, future opportunities and risks

A Tornede, D Deng, T Eimer, J Giovanelli… - arXiv preprint arXiv …, 2023 - arxiv.org
The fields of both Natural Language Processing (NLP) and Automated Machine Learning
(AutoML) have achieved remarkable results over the past years. In NLP, especially Large …

Data management for large language models: A survey

Z Wang, W Zhong, Y Wang, Q Zhu, F Mi, B Wang… - CoRR, 2023 - openreview.net
Data plays a fundamental role in the training of Large Language Models (LLMs). Effective
data management, particularly in the formulation of a well-suited training dataset, holds …

[HTML][HTML] Explainable ensemble learning predictive model for thermal conductivity of cement-based foam

C Cakiroglu, F Batool, K Islam, ML Nehdi - Construction and Building …, 2024 - Elsevier
Cement-based foam has emerged as a strong contender in sustainable construction owing
to its superior thermal and sound insulation properties, fire resistance, and cost …

An empirical study on hyperparameter optimization for fine-tuning pre-trained language models

X Liu, C Wang - arXiv preprint arXiv:2106.09204, 2021 - arxiv.org
The performance of fine-tuning pre-trained language models largely depends on the
hyperparameter configuration. In this paper, we investigate the performance of modern …

Geochemistry π: automated machine learning Python framework for tabular data

J ZhangZhou, C He, J Sun, J Zhao… - Geochemistry …, 2024 - Wiley Online Library
Although machine learning (ML) has brought new insights into geochemistry research, its
implementation is laborious and time‐consuming. Here, we announce Geochemistry π, an …

Adv3d: Generating safety-critical 3d objects through closed-loop simulation

J Sarva, J Wang, J Tu, Y Xiong, S Manivasagam… - arXiv preprint arXiv …, 2023 - arxiv.org
Self-driving vehicles (SDVs) must be rigorously tested on a wide range of scenarios to
ensure safe deployment. The industry typically relies on closed-loop simulation to evaluate …

AutoRAG-HP: Automatic Online Hyper-Parameter Tuning for Retrieval-Augmented Generation

J Fu, X Qin, F Yang, L Wang, J Zhang, Q Lin… - arXiv preprint arXiv …, 2024 - arxiv.org
Recent advancements in Large Language Models have transformed ML/AI development,
necessitating a reevaluation of AutoML principles for the Retrieval-Augmented Generation …