Pre-trained language models in biomedical domain: A systematic survey

B Wang, Q Xie, J Pei, Z Chen, P Tiwari, Z Li… - ACM Computing …, 2023 - dl.acm.org
Pre-trained language models (PLMs) have been the de facto paradigm for most natural
language processing tasks. This also benefits the biomedical domain: researchers from …

Med-unic: Unifying cross-lingual medical vision-language pre-training by diminishing bias

Z Wan, C Liu, M Zhang, J Fu, B Wang… - Advances in …, 2024 - proceedings.neurips.cc
The scarcity of data presents a critical obstacle to the efficacy of medical vision-language pre-
training (VLP). A potential solution lies in the combination of datasets from various language …

Clip in medical imaging: A comprehensive survey

Z Zhao, Y Liu, H Wu, Y Li, S Wang, L Teng… - arXiv preprint arXiv …, 2023 - arxiv.org
Contrastive Language-Image Pre-training (CLIP), a straightforward yet effective pre-training
paradigm, successfully introduces semantic-rich text supervision to vision models and has …

Prior: Prototype representation joint learning from medical images and reports

P Cheng, L Lin, J Lyu, Y Huang… - Proceedings of the …, 2023 - openaccess.thecvf.com
Contrastive learning based vision-language joint pre-training has emerged as a successful
representation learning strategy. In this paper, we present a prototype representation …

M-flag: Medical vision-language pre-training with frozen language models and latent space geometry optimization

C Liu, S Cheng, C Chen, M Qiao, W Zhang… - … Conference on Medical …, 2023 - Springer
Medical vision-language models enable co-learning and integrating features from medical
imaging and clinical text. However, these models are not easy to train and the latent …

Towards unifying medical vision-and-language pre-training via soft prompts

Z Chen, S Diao, B Wang, G Li… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Medical vision-and-language pre-training (Med-VLP) has shown promising improvements
on many downstream medical tasks owing to its applicability to extracting generic …

Imitate: Clinical prior guided hierarchical vision-language pre-training

C Liu, S Cheng, M Shi, A Shah, W Bai… - arXiv preprint arXiv …, 2023 - arxiv.org
In the field of medical Vision-Language Pre-training (VLP), significant efforts have been
devoted to deriving text and image features from both clinical reports and associated …

Carzero: Cross-attention alignment for radiology zero-shot classification

H Lai, Q Yao, Z Jiang, R Wang, Z He… - Proceedings of the …, 2024 - openaccess.thecvf.com
Abstract The advancement of Zero-Shot Learning in the medical domain has been driven
forward by using pre-trained models on large-scale image-text pairs focusing on image-text …

Learning multiscale consistency for self-supervised electron microscopy instance segmentation

Y Chen, W Huang, X Liu, S Deng… - ICASSP 2024-2024 …, 2024 - ieeexplore.ieee.org
Electron microscopy (EM) images are notoriously challenging to segment due to their
complex structures and lack of effective annotations. Fortunately, large-scale self-supervised …

A scoping review on multimodal deep learning in biomedical images and texts

Z Sun, M Lin, Q Zhu, Q Xie, F Wang, Z Lu… - Journal of Biomedical …, 2023 - Elsevier
Objective Computer-assisted diagnostic and prognostic systems of the future should be
capable of simultaneously processing multimodal data. Multimodal deep learning (MDL) …