Clip in medical imaging: A comprehensive survey

Z Zhao, Y Liu, H Wu, M Wang, Y Li, S Wang… - arXiv preprint arXiv …, 2023 - arxiv.org
Contrastive Language-Image Pre-training (CLIP), a simple yet effective pre-training
paradigm, successfully introduces text supervision to vision models. It has shown promising …

Imitate: Clinical prior guided hierarchical vision-language pre-training

C Liu, S Cheng, M Shi, A Shah, W Bai… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
In the field of medical Vision-Language Pretraining (VLP), significant efforts have been
devoted to deriving text and image features from both clinical reports and associated …

Foundation model for advancing healthcare: Challenges, opportunities, and future directions

Y He, F Huang, X Jiang, Y Nie, M Wang, J Wang… - arXiv preprint arXiv …, 2024 - arxiv.org
Foundation model, which is pre-trained on broad data and is able to adapt to a wide range
of tasks, is advancing healthcare. It promotes the development of healthcare artificial …

Enhancing representation in radiography-reports foundation model: A granular alignment algorithm using masked contrastive learning

W Huang, C Li, HY Zhou, H Yang, J Liu, Y Liang… - Nature …, 2024 - nature.com
Recently, multi-modal vision-language foundation models have gained significant attention
in the medical field. While these models offer great opportunities, they still face crucial …

Bimcv-r: A landmark dataset for 3d ct text-image retrieval

Y Chen, C Liu, X Liu, R Arcucci, Z Xiong - International Conference on …, 2024 - Springer
The burgeoning integration of 3D medical imaging into healthcare has led to a substantial
increase in the workload of medical professionals. To assist clinicians in their diagnostic …

Learning multiscale consistency for self-supervised electron microscopy instance segmentation

Y Chen, W Huang, X Liu, S Deng… - ICASSP 2024-2024 …, 2024 - ieeexplore.ieee.org
Electron microscopy (EM) images are notoriously challenging to segment due to their
complex structures and lack of effective annotations. Fortunately, large-scale self-supervised …

Etp: Learning transferable ecg representations via ecg-text pre-training

C Liu, Z Wan, S Cheng, M Zhang… - ICASSP 2024-2024 …, 2024 - ieeexplore.ieee.org
In the domain of cardiovascular healthcare, the Electrocardiogram (ECG) serves as a critical,
non-invasive diagnostic tool. Although recent strides in self-supervised learning (SSL) have …

T3d: Towards 3d medical image understanding through vision-language pre-training

C Liu, C Ouyang, Y Chen, CC Quilodrán-Casas… - arXiv preprint arXiv …, 2023 - arxiv.org
Expert annotation of 3D medical image for downstream analysis is resource-intensive,
posing challenges in clinical applications. Visual self-supervised learning (vSSL), though …

Mlip: Enhancing medical visual representation with divergence encoder and knowledge-guided contrastive learning

Z Li, LT Yang, B Ren, X Nie, Z Gao… - Proceedings of the …, 2024 - openaccess.thecvf.com
The scarcity of annotated data has sparked significant interest in unsupervised pre-training
methods that leverage medical reports as auxiliary signals for medical visual representation …

Continual self-supervised learning: Towards universal multi-modal medical data representation learning

Y Ye, Y Xie, J Zhang, Z Chen… - Proceedings of the …, 2024 - openaccess.thecvf.com
Self-supervised learning (SSL) is an efficient pre-training method for medical image
analysis. However current research is mostly confined to certain modalities consuming …