Evaluating large language models for radiology natural language processing

Z Liu, T Zhong, Y Li, Y Zhang, Y Pan, Z Zhao… - arXiv preprint arXiv …, 2023 - arxiv.org
The rise of large language models (LLMs) has marked a pivotal shift in the field of natural
language processing (NLP). LLMs have revolutionized a multitude of domains, and they …

Cif-bench: A chinese instruction-following benchmark for evaluating the generalizability of large language models

Y Li, G Zhang, X Qu, J Li, Z Li, Z Wang, H Li… - arXiv preprint arXiv …, 2024 - arxiv.org
The advancement of large language models (LLMs) has enhanced the ability to generalize
across a wide range of unseen natural language processing (NLP) tasks through instruction …

Nlebench+ norglm: A comprehensive empirical analysis and benchmark dataset for generative language models in norwegian

P Liu, L Zhang, T Farup, EW Lauvrak… - arXiv preprint arXiv …, 2023 - arxiv.org
Norwegian, spoken by only 5 million population, is under-representative within the most
impressive breakthroughs in NLP tasks. To the best of our knowledge, there has not yet …

Persianllama: Towards building first persian large language model

MA Abbasi, A Ghafouri, M Firouzmandi… - arXiv preprint arXiv …, 2023 - arxiv.org
Despite the widespread use of the Persian language by millions globally, limited efforts have
been made in natural language processing for this language. The use of large language …

LMTuner: An user-friendly and highly-integrable Training Framework for fine-tuning Large Language Models

Y Weng, Z Wang, H Liao, S He, S Liu, K Liu… - arXiv preprint arXiv …, 2023 - arxiv.org
With the burgeoning development in the realm of large language models (LLMs), the
demand for efficient incremental training tailored to specific industries and domains …

Create and find flatness: Building flat training spaces in advance for continual learning

W Shi, Y Chen, Z Zhao, W Lu, K Yan, X Du - ECAI 2023, 2023 - ebooks.iospress.nl
Catastrophic forgetting remains a critical challenge in the field of continual learning, where
neural networks struggle to retain prior knowledge while assimilating new information. Most …

Weight-inherited distillation for task-agnostic bert compression

T Wu, C Hou, S Lao, J Li, N Wong, Z Zhao… - arXiv preprint arXiv …, 2023 - arxiv.org
Knowledge Distillation (KD) is a predominant approach for BERT compression. Previous KD-
based methods focus on designing extra alignment losses for the student model to mimic the …

Enhanced ICD-10 code assignment of clinical texts: A summarization-based approach

Y Sun, L Sang, D Wu, S He, Y Chen, H Duan… - Artificial Intelligence in …, 2024 - Elsevier
Abstract Background Assigning International Classification of Diseases (ICD) codes to
clinical texts is a common and crucial practice in patient classification, hospital management …

FLIP-80M: 80 Million Visual-Linguistic Pairs for Facial Language-Image Pre-Training

Y Li, X Hou, Z Dezhi, L Shen, Z Zhao - Proceedings of the 32nd ACM …, 2024 - dl.acm.org
While significant progress has been made in multi-modal learning driven by large-scale
image-text datasets, there is still a noticeable gap in the availability of such datasets within …

Facilitating Multi-Role and Multi-Behavior Collaboration of Large Language Models for Online Job Seeking and Recruiting

H Sun, H Lin, H Yan, C Zhu, Y Song, X Gao… - arXiv preprint arXiv …, 2024 - arxiv.org
The emergence of online recruitment services has revolutionized the traditional landscape
of job seeking and recruitment, necessitating the development of high-quality industrial …