Vip: A differentially private foundation model for computer vision

Y Yu, M Sanjabi, Y Ma, K Chaudhuri, C Guo - arXiv preprint arXiv …, 2023 - arxiv.org
Artificial intelligence (AI) has seen a tremendous surge in capabilities thanks to the use of
foundation models trained on internet-scale data. On the flip side, the uncurated nature of …

A Survey on Safe Multi-Modal Learning System

T Zhao, L Zhang, Y Ma, L Cheng - arXiv preprint arXiv:2402.05355, 2024 - arxiv.org
With the wide deployment of multimodal learning systems (MMLS) in real-world scenarios,
safety concerns have become increasingly prominent. The absence of systematic research …

Identity Inference from CLIP Models using Only Textual Data

S Li, R Cheng, X Jia - arXiv preprint arXiv:2405.14517, 2024 - arxiv.org
The widespread usage of large-scale multimodal models like CLIP has heightened
concerns about the leakage of personally identifiable information (PII). Existing methods for …

The Synergy between Data and Multi-Modal Large Language Models: A Survey from Co-Development Perspective

Z Qin, D Chen, W Zhang, L Yao, Y Huang… - arXiv preprint arXiv …, 2024 - arxiv.org
The rapid development of large language models (LLMs) has been witnessed in recent
years. Based on the powerful LLMs, multi-modal LLMs (MLLMs) extend the modality from …