[HTML][HTML] Review of large vision models and visual prompt engineering

J Wang, Z Liu, L Zhao, Z Wu, C Ma, S Yu, H Dai… - Meta-Radiology, 2023 - Elsevier
Visual prompt engineering is a fundamental methodology in the field of visual and image
artificial general intelligence. As the development of large vision models progresses, the …

A comprehensive survey on segment anything model for vision and beyond

C Zhang, L Liu, Y Cui, G Huang, W Lin, Y Yang… - arXiv preprint arXiv …, 2023 - arxiv.org
Artificial intelligence (AI) is evolving towards artificial general intelligence, which refers to the
ability of an AI system to perform a wide range of tasks and exhibit a level of intelligence …

Multimodal foundation models: From specialists to general-purpose assistants

C Li, Z Gan, Z Yang, J Yang, L Li… - … and Trends® in …, 2024 - nowpublishers.com
Neural compression is the application of neural networks and other machine learning
methods to data compression. Recent advances in statistical machine learning have opened …

Desam: Decoupled segment anything model for generalizable medical image segmentation

Y Gao, W Xia, D Hu, W Wang, X Gao - International Conference on …, 2024 - Springer
Deep learning-based medical image segmentation models often suffer from domain shift,
where the models trained on a source domain do not generalize well to other unseen …

Sam-6d: Segment anything model meets zero-shot 6d object pose estimation

J Lin, L Liu, D Lu, K Jia - … of the IEEE/CVF Conference on …, 2024 - openaccess.thecvf.com
Zero-shot 6D object pose estimation involves the detection of novel objects with their 6D
poses in cluttered scenes presenting significant challenges for model generalizability …

Segment anything model (sam) for radiation oncology

L Zhang, Z Liu, L Zhang, Z Wu, X Yu, J Holmes… - arXiv preprint arXiv …, 2023 - arxiv.org
In this study, we evaluate the performance of the Segment Anything Model (SAM) model in
clinical radiotherapy. We collected real clinical cases from four regions at the Mayo Clinic …

VRP-SAM: SAM with visual reference prompt

Y Sun, J Chen, S Zhang, X Zhang… - Proceedings of the …, 2024 - openaccess.thecvf.com
In this paper we propose a novel Visual Reference Prompt (VRP) encoder that empowers
the Segment Anything Model (SAM) to utilize annotated reference images as prompts for …

On the robustness of segment anything

Y Huang, Y Cao, T Li, F Juefei-Xu, D Lin… - arXiv preprint arXiv …, 2023 - arxiv.org
Segment anything model (SAM) has presented impressive objectness identification
capability with the idea of prompt learning and a new collected large-scale dataset. Given a …

Distilling Semantic Priors from SAM to Efficient Image Restoration Models

Q Zhang, X Liu, W Li, H Chen, J Liu… - Proceedings of the …, 2024 - openaccess.thecvf.com
In image restoration (IR) leveraging semantic priors from segmentation models has been a
common approach to improve performance. The recent segment anything model (SAM) has …

Black-box targeted adversarial attack on segment anything (sam)

S Zheng, C Zhang, X Hao - IEEE Transactions on Multimedia, 2024 - ieeexplore.ieee.org
Deep recognition models are widely vulnerable to adversarial examples, which change the
model output by adding quasi-imperceptible perturbation to the image input. Recently …