Pre-trained language models in biomedical domain: A systematic survey

B Wang, Q Xie, J Pei, Z Chen, P Tiwari, Z Li… - ACM Computing …, 2023 - dl.acm.org
Pre-trained language models (PLMs) have been the de facto paradigm for most natural
language processing tasks. This also benefits the biomedical domain: researchers from …

Recent advancements in emerging technologies for healthcare management systems: a survey

SB Junaid, AA Imam, AO Balogun, LC De Silva… - Healthcare, 2022 - mdpi.com
In recent times, the growth of the Internet of Things (IoT), artificial intelligence (AI), and
Blockchain technologies have quickly gained pace as a new study niche in numerous …

A large language model for electronic health records

X Yang, A Chen, N PourNejatian, HC Shin… - NPJ digital …, 2022 - nature.com
There is an increasing interest in developing artificial intelligence (AI) systems to process
and interpret electronic health records (EHRs). Natural language processing (NLP) powered …

Transfer learning in biomedical natural language processing: an evaluation of BERT and ELMo on ten benchmarking datasets

Y Peng, S Yan, Z Lu - arXiv preprint arXiv:1906.05474, 2019 - arxiv.org
Inspired by the success of the General Language Understanding Evaluation benchmark, we
introduce the Biomedical Language Understanding Evaluation (BLUE) benchmark to …

CharacterBERT: Reconciling ELMo and BERT for word-level open-vocabulary representations from characters

HE Boukkouri, O Ferret, T Lavergne, H Noji… - arXiv preprint arXiv …, 2020 - arxiv.org
Due to the compelling improvements brought by BERT, many recent representation models
adopted the Transformer architecture as their main building block, consequently inheriting …

MedCPT: Contrastive Pre-trained Transformers with large-scale PubMed search logs for zero-shot biomedical information retrieval

Q Jin, W Kim, Q Chen, DC Comeau, L Yeganova… - …, 2023 - academic.oup.com
Motivation Information retrieval (IR) is essential in biomedical knowledge acquisition and
clinical decision support. While recent progress has shown that language model encoders …

BioSentVec: creating sentence embeddings for biomedical texts

Q Chen, Y Peng, Z Lu - 2019 IEEE International Conference on …, 2019 - ieeexplore.ieee.org
Sentence embeddings have become an essential part of today's natural language
processing (NLP) systems, especially together advanced deep learning methods. Although …

Large language models are poor medical coders—benchmarking of medical code querying

A Soroush, BS Glicksberg, E Zimlichman, Y Barash… - NEJM AI, 2024 - ai.nejm.org
Abstract Background Large language models (LLMs) have attracted significant interest for
automated clinical coding. However, early data show that LLMs are highly error-prone when …

Gatortron: A large clinical language model to unlock patient information from unstructured electronic health records

X Yang, A Chen, N PourNejatian, HC Shin… - arXiv preprint arXiv …, 2022 - arxiv.org
There is an increasing interest in developing artificial intelligence (AI) systems to process
and interpret electronic health records (EHRs). Natural language processing (NLP) powered …

An empirical study of multi-task learning on BERT for biomedical text mining

Y Peng, Q Chen, Z Lu - arXiv preprint arXiv:2005.02799, 2020 - arxiv.org
Multi-task learning (MTL) has achieved remarkable success in natural language processing
applications. In this work, we study a multi-task learning model with multiple decoders on …