Privacy and security issues in deep learning: A survey

X Liu, L Xie, Y Wang, J Zou, J Xiong, Z Ying… - IEEE …, 2020 - ieeexplore.ieee.org
Deep Learning (DL) algorithms based on artificial neural networks have achieved
remarkable success and are being extensively applied in a variety of application domains …

Privacy side channels in machine learning systems

E Debenedetti, G Severi, N Carlini… - arXiv preprint arXiv …, 2023 - arxiv.org
Most current approaches for protecting privacy in machine learning (ML) assume that
models exist in a vacuum, when in reality, ML models are part of larger systems that include …

PAR-GAN: improving the generalization of generative adversarial networks against membership inference attacks

J Chen, WH Wang, H Gao, X Shi - Proceedings of the 27th ACM SIGKDD …, 2021 - dl.acm.org
Recent works have shown that Generative Adversarial Networks (GANs) may generalize
poorly and thus are vulnerable to privacy attacks. In this paper, we seek to improve the …

Privacy in deep learning: A survey

F Mireshghallah, M Taram, P Vepakomma… - arXiv preprint arXiv …, 2020 - arxiv.org
The ever-growing advances of deep learning in many areas including vision,
recommendation systems, natural language processing, etc., have led to the adoption of …

Membership inference attacks against adversarially robust deep learning models

L Song, R Shokri, P Mittal - 2019 IEEE Security and Privacy …, 2019 - ieeexplore.ieee.org
In recent years, the research community has increasingly focused on understanding the
security and privacy challenges posed by deep learning models. However, the security …

Model inversion attacks that exploit confidence information and basic countermeasures

M Fredrikson, S Jha, T Ristenpart - … of the 22nd ACM SIGSAC conference …, 2015 - dl.acm.org
Machine-learning (ML) algorithms are increasingly utilized in privacy-sensitive applications
such as predicting lifestyle choices, making medical diagnoses, and facial recognition. In a …

One parameter defense—defending against data inference attacks via differential privacy

D Ye, S Shen, T Zhu, B Liu… - IEEE Transactions on …, 2022 - ieeexplore.ieee.org
Machine learning models are vulnerable to data inference attacks, such as membership
inference and model inversion attacks. In these types of breaches, an adversary attempts to …

When does data augmentation help with membership inference attacks?

Y Kaya, T Dumitras - International conference on machine …, 2021 - proceedings.mlr.press
Deep learning models often raise privacy concerns as they leak information about their
training data. This leakage enables membership inference attacks (MIA) that can identify …

Comprehensive privacy analysis of deep learning: Passive and active white-box inference attacks against centralized and federated learning

M Nasr, R Shokri, A Houmansadr - 2019 IEEE symposium on …, 2019 - ieeexplore.ieee.org
Deep neural networks are susceptible to various inference attacks as they remember
information about their training data. We design white-box inference attacks to perform a …