We study the problem of watermarking large language models (LLMs) generated text -- one of the most promising approaches for addressing the safety challenges of LLM usage. In this paper, we propose a rigorous theoretical framework to quantify the effectiveness and robustness of LLM watermarks. We propose a robust and high-quality watermark method, Unigram-Watermark, by extending an existing approach with a simplified fixed grouping strategy. We prove that our watermark method enjoys guaranteed generation quality, correctness in watermark detection, and is robust against text editing and paraphrasing. Experiments on three varying LLMs and two datasets verify that our Unigram-Watermark achieves superior detection accuracy and comparable gener...
Every day more and more content is published on the internet in social media sites, cloud sharing se...
Recent advances in large language models (LLMs) and the intensifying popularity of ChatGPT-like appl...
The prevalence and strong capability of large language models (LLMs) present significant safety and ...
We propose a methodology for planting watermarks in text from an autoregressive language model that ...
The recent advancements in large language models (LLMs) have sparked a growing apprehension regardin...
We present REMARK-LLM, a novel efficient, and robust watermarking framework designed for texts gener...
We construct the first provable watermarking scheme for language models with public detectability or...
Recently, text watermarking algorithms for large language models (LLMs) have been mitigating the pot...
Watermarking generative models consists of planting a statistical signal (watermark) in a model’s ou...
Large language models (LLMs) have show great ability in various natural language tasks. However, the...
Watermarking generative models consists of planting a statistical signal (watermark) in a model's ou...
Watermarking techniques are used to help identifying copies of publicly released information. They c...
We propose Easymark, a family of embarrassingly simple yet effective watermarks. Text watermarking i...
Watermarks should be introduced in the natural language outputs of AI systems in order to maintain t...
International audienceThis paper presents yet another attempt towards robust and secure watermarking...
Every day more and more content is published on the internet in social media sites, cloud sharing se...
Recent advances in large language models (LLMs) and the intensifying popularity of ChatGPT-like appl...
The prevalence and strong capability of large language models (LLMs) present significant safety and ...
We propose a methodology for planting watermarks in text from an autoregressive language model that ...
The recent advancements in large language models (LLMs) have sparked a growing apprehension regardin...
We present REMARK-LLM, a novel efficient, and robust watermarking framework designed for texts gener...
We construct the first provable watermarking scheme for language models with public detectability or...
Recently, text watermarking algorithms for large language models (LLMs) have been mitigating the pot...
Watermarking generative models consists of planting a statistical signal (watermark) in a model’s ou...
Large language models (LLMs) have show great ability in various natural language tasks. However, the...
Watermarking generative models consists of planting a statistical signal (watermark) in a model's ou...
Watermarking techniques are used to help identifying copies of publicly released information. They c...
We propose Easymark, a family of embarrassingly simple yet effective watermarks. Text watermarking i...
Watermarks should be introduced in the natural language outputs of AI systems in order to maintain t...
International audienceThis paper presents yet another attempt towards robust and secure watermarking...
Every day more and more content is published on the internet in social media sites, cloud sharing se...
Recent advances in large language models (LLMs) and the intensifying popularity of ChatGPT-like appl...
The prevalence and strong capability of large language models (LLMs) present significant safety and ...