How to dp-fy ml: A practical guide to machine learning with differential privacy

N Ponomareva, H Hazimeh, A Kurakin, Z Xu… - Journal of Artificial …, 2023 - jair.org
Abstract Machine Learning (ML) models are ubiquitous in real-world applications and are a
constant focus of research. Modern ML models have become more complex, deeper, and …

Large language models can be strong differentially private learners

X Li, F Tramer, P Liang, T Hashimoto - arXiv preprint arXiv:2110.05679, 2021 - arxiv.org
Differentially Private (DP) learning has seen limited success for building large deep learning
models of text, and straightforward attempts at applying Differentially Private Stochastic …

Differentially private fine-tuning of language models

D Yu, S Naik, A Backurs, S Gopi, HA Inan… - arXiv preprint arXiv …, 2021 - arxiv.org
We give simpler, sparser, and faster algorithms for differentially private fine-tuning of large-
scale pre-trained language models, which achieve the state-of-the-art privacy versus utility …

Flocks of stochastic parrots: Differentially private prompt learning for large language models

H Duan, A Dziedzic, N Papernot… - Advances in Neural …, 2024 - proceedings.neurips.cc
Large language models (LLMs) are excellent in-context learners. However, the sensitivity of
data contained in prompts raises privacy concerns. Our work first shows that these concerns …

What does it mean for a language model to preserve privacy?

H Brown, K Lee, F Mireshghallah, R Shokri… - Proceedings of the 2022 …, 2022 - dl.acm.org
Natural language reflects our private lives and identities, making its privacy concerns as
broad as those of real life. Language models lack the ability to understand the context and …

On protecting the data privacy of large language models (llms): A survey

B Yan, K Li, M Xu, Y Dong, Y Zhang, Z Ren… - arXiv preprint arXiv …, 2024 - arxiv.org
Large language models (LLMs) are complex artificial intelligence systems capable of
understanding, generating and translating human language. They learn language patterns …

Are large pre-trained language models leaking your personal information?

J Huang, H Shao, KCC Chang - arXiv preprint arXiv:2205.12628, 2022 - arxiv.org
Are Large Pre-Trained Language Models Leaking Your Personal Information? In this paper,
we analyze whether Pre-Trained Language Models (PLMs) are prone to leaking personal …

Differentially private natural language models: Recent advances and future directions

L Hu, I Habernal, L Shen, D Wang - arXiv preprint arXiv:2301.09112, 2023 - arxiv.org
Recent developments in deep learning have led to great success in various natural
language processing (NLP) tasks. However, these applications may involve data that …

Large-scale differentially private BERT

R Anil, B Ghazi, V Gupta, R Kumar… - arXiv preprint arXiv …, 2021 - arxiv.org
In this work, we study the large-scale pretraining of BERT-Large with differentially private
SGD (DP-SGD). We show that combined with a careful implementation, scaling up the batch …

Privacy-preserving prompt tuning for large language model services

Y Li, Z Tan, Y Liu - arXiv preprint arXiv:2305.06212, 2023 - arxiv.org
Prompt tuning provides an efficient way for users to customize Large Language Models
(LLMs) with their private data in the emerging LLM service scenario. However, the sensitive …