Training data extraction from pre-trained language models: A survey

S Ishihara - arXiv preprint arXiv:2305.16157, 2023 - arxiv.org
As the deployment of pre-trained language models (PLMs) expands, pressing security
concerns have arisen regarding the potential for malicious extraction of training data, posing …

Identifying and mitigating privacy risks stemming from language models: A survey

V Smith, AS Shamsabadi, C Ashurst… - arXiv preprint arXiv …, 2023 - arxiv.org
Rapid advancements in language models (LMs) have led to their adoption across many
sectors. Alongside the potential benefits, such models present a range of risks, including …

Stealing the decoding algorithms of language models

A Naseh, K Krishna, M Iyyer… - Proceedings of the 2023 …, 2023 - dl.acm.org
A key component of generating text from modern language models (LM) is the selection and
tuning of decoding algorithms. These algorithms determine how to generate text from the …

Security challenges in natural language processing models

Q Xu, X He - Proceedings of the 2023 Conference on Empirical …, 2023 - aclanthology.org
Large-scale natural language processing models have been developed and integrated into
numerous applications, given the advantage of their remarkable performance. Nonetheless …

Anti-Backdoor Model: A Novel Algorithm To Remove Backdoors in a Non-invasive Way

C Chen, H Hong, T Xiang, M Xie - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Recent research findings suggest that machine learning models are highly susceptible to
backdoor poisoning attacks. Backdoor poisoning attacks can be easily executed and …

Large language models for conducting advanced text Analytics Information Systems Research

B Ampel, CH Yang, J Hu, H Chen - ACM Transactions on Management …, 2024 - dl.acm.org
The exponential growth of digital content has generated massive textual datasets,
necessitating the use of advanced analytical approaches. Large Language Models (LLMs) …

Divtheft: An ensemble model stealing attack by divide-and-conquer

Z Ma, X Liu, Y Liu, X Liu, Z Qin… - IEEE transactions on …, 2023 - ieeexplore.ieee.org
Recently, model stealing attacks are widely studied but most of them are focused on stealing
a single non-discrete model, eg, neural networks. For ensemble models, these attacks are …

REN-AI: A Video Game for AI Security Education Leveraging Episodic Memory

M Arai, K Tejima, Y Yamada, T Miura… - IEEE …, 2024 - ieeexplore.ieee.org
Education in cybersecurity is crucial in the current society, and it will be extended into the
artificial intelligence (AI) area, called AI security, in the near future. Although many video …

WARDEN: Multi-Directional Backdoor Watermarks for Embedding-as-a-Service Copyright Protection

A Shetty, Y Teng, K He, Q Xu - arXiv preprint arXiv:2403.01472, 2024 - arxiv.org
Embedding as a Service (EaaS) has become a widely adopted solution, which offers feature
extraction capabilities for addressing various downstream tasks in Natural Language …

Decepticon: Attacking secrets of transformers

M Al Rafi, Y Feng, F Yao, M Tang… - 2023 IEEE International …, 2023 - ieeexplore.ieee.org
With the growing burden of training deep learning models with huge datasets, transfer
learning has been widely adopted (eg, Transformers like BERT, GPT). Transfer learning …