How to dp-fy ml: A practical guide to machine learning with differential privacy

N Ponomareva, H Hazimeh, A Kurakin, Z Xu… - Journal of Artificial …, 2023 - jair.org
Abstract Machine Learning (ML) models are ubiquitous in real-world applications and are a
constant focus of research. Modern ML models have become more complex, deeper, and …

Detecting pretraining data from large language models

W Shi, A Ajith, M Xia, Y Huang, D Liu, T Blevins… - arXiv preprint arXiv …, 2023 - arxiv.org
Although large language models (LLMs) are widely deployed, the data used to train them is
rarely disclosed. Given the incredible scale of this data, up to trillions of tokens, it is all but …

Privacy implications of retrieval-based language models

Y Huang, S Gupta, Z Zhong, K Li, D Chen - arXiv preprint arXiv …, 2023 - arxiv.org
Retrieval-based language models (LMs) have demonstrated improved interpretability,
factuality, and adaptability compared to their parametric counterparts, by incorporating …

Posthoc privacy guarantees for collaborative inference with modified Propose-Test-Release

A Singh, P Vepakomma, V Sharma… - Advances in Neural …, 2023 - proceedings.neurips.cc
Cloud-based machine learning inference is an emerging paradigm where users query by
sending their data through a service provider who runs an ML model on that data and …

Cogenesis: A framework collaborating large and small language models for secure context-aware instruction following

K Zhang, J Wang, E Hua, B Qi, N Ding… - arXiv preprint arXiv …, 2024 - arxiv.org
With the advancement of language models (LMs), their exposure to private data is
increasingly inevitable, and their deployment (especially for smaller ones) on personal …

Private learning with public features

W Krichene, NE Mayoraz, S Rendle… - International …, 2024 - proceedings.mlr.press
We study a class of private learning problems in which the data is a join of private and public
features. This is often the case in private personalization tasks such as recommendation or …

Private matrix factorization with public item features

M Curmei, W Krichene, L Zhang… - Proceedings of the 17th …, 2023 - dl.acm.org
We consider the problem of training private recommendation models with access to public
item features. Training with Differential Privacy (DP) offers strong privacy guarantees, at the …

Gaussian membership inference privacy

T Leemann, M Pawelczyk… - Advances in Neural …, 2024 - proceedings.neurips.cc
We propose a novel and practical privacy notion called $ f $-Membership Inference Privacy
($ f $-MIP), which explicitly considers the capabilities of realistic adversaries under the …

Quantum local differential privacy and quantum statistical query model

A Angrisani, E Kashefi - arXiv preprint arXiv:2203.03591, 2022 - arxiv.org
Quantum statistical queries provide a theoretical framework for investigating the
computational power of a learner with limited quantum resources. This model is particularly …

Understanding how Differentially Private Generative Models Spend their Privacy Budget

G Ganev, K Xu, E De Cristofaro - arXiv preprint arXiv:2305.10994, 2023 - arxiv.org
Generative models trained with Differential Privacy (DP) are increasingly used to produce
synthetic data while reducing privacy risks. Navigating their specific privacy-utility tradeoffs …