Challenges and applications of large language models

J Kaddour, J Harris, M Mozes, H Bradley… - arXiv preprint arXiv …, 2023 - arxiv.org
Large Language Models (LLMs) went from non-existent to ubiquitous in the machine
learning discourse within a few years. Due to the fast pace of the field, it is difficult to identify …

Undetectable watermarks for language models

M Christ, S Gunn, O Zamir - The Thirty Seventh Annual …, 2024 - proceedings.mlr.press
Recent advances in the capabilities of large language models such as GPT-4 have spurred
increasing concern about our ability to detect AI-generated text. Prior works have suggested …

Unbiased watermark for large language models

Z Hu, L Chen, X Wu, Y Wu, H Zhang… - arXiv preprint arXiv …, 2023 - arxiv.org
The recent advancements in large language models (LLMs) have sparked a growing
apprehension regarding the potential misuse. One approach to mitigating this risk is to …

Who wrote this code? watermarking for code generation

T Lee, S Hong, J Ahn, I Hong, H Lee, S Yun… - arXiv preprint arXiv …, 2023 - arxiv.org
With the remarkable generation performance of large language models, ethical and legal
concerns about using them have been raised, such as plagiarism and copyright issues. For …

Publicly detectable watermarking for language models

J Fairoze, S Garg, S Jha, S Mahloujifar… - arXiv preprint arXiv …, 2023 - arxiv.org
We construct the first provable watermarking scheme for language models with public
detectability or verifiability: we use a private key for watermarking and a public key for …

Dipmark: A stealthy, efficient and resilient watermark for large language models

Y Wu, Z Hu, H Zhang, H Huang - arXiv preprint arXiv:2310.07710, 2023 - arxiv.org
Watermarking techniques offer a promising way to secure data via embedding covert
information into the data. A paramount challenge in the domain lies in preserving the …

Excuse me, sir? Your language model is leaking (information)

O Zamir - arXiv preprint arXiv:2401.10360, 2024 - arxiv.org
We introduce a cryptographic method to hide an arbitrary secret payload in the response of
a Large Language Model (LLM). A secret key is required to extract the payload from the …

Waterbench: Towards holistic evaluation of watermarks for large language models

S Tu, Y Sun, Y Bai, J Yu, L Hou, J Li - arXiv preprint arXiv:2311.07138, 2023 - arxiv.org
To mitigate the potential misuse of large language models (LLMs), recent research has
developed watermarking algorithms, which restrict the generation process to leave an …

Undetectable Steganography for Language Models

O Zamir - Transactions on Machine Learning Research, 2024 - openreview.net
We introduce a cryptographic method to hide an arbitrary secret payload in the response of
a Large Language Model (LLM). A secret key is required to extract the payload from the …

DALD: Improving Logits-based Detector without Logits from Black-box LLMs

C Zeng, S Tang, X Yang, Y Chen, Y Sun… - The Thirty-eighth …, 2024 - openreview.net
The advent of Large Language Models (LLMs) has revolutionized text generation, producing
outputs that closely mimic human writing. This blurring of lines between machine-and …