作者
Masoud Monajatipoor, Mozhdeh Rouhsedaghat, Liunian Harold Li, C-C Jay Kuo, Aichi Chien, Kai-Wei Chang
发表日期
2022/9/16
图书
International Conference on Medical Image Computing and Computer-Assisted Intervention
页码范围
725-734
出版商
Springer Nature Switzerland
简介
Vision-and-language () models take image and text as input and learn to capture the associations between them. These models can potentially deal with the tasks that involve understanding medical images along with their associated text. However, applying models in the medical domain is challenging due to the expensiveness of data annotations and the requirements of domain knowledge. In this paper, we identify that the visual representation in general models is not suitable for processing medical data. To overcome this limitation, we propose BERTHop, a transformer-based model based on PixelHop++ and VisualBERT for better capturing the associations between clinical notes and medical images.
Experiments on the OpenI dataset, a commonly used thoracic disease diagnosis benchmark, show that BERTHop achieves an average Area Under the Curve (AUC) of 98.12% which is 1.62% higher …
引用总数
学术搜索中的文章
M Monajatipoor, M Rouhsedaghat, LH Li, CC Jay Kuo… - International Conference on Medical Image Computing …, 2022