all InfoSec news
On the Reliability of Watermarks for Large Language Models. (arXiv:2306.04634v1 [cs.LG])
cs.CR updates on arXiv.org arxiv.org
Large language models (LLMs) are now deployed to everyday use and positioned
to produce large quantities of text in the coming decade. Machine-generated
text may displace human-written text on the internet and has the potential to
be used for malicious purposes, such as spearphishing attacks and social media
bots. Watermarking is a simple and effective strategy for mitigating such harms
by enabling the detection and documentation of LLM-generated text. Yet, a
crucial question remains: How reliable is watermarking in realistic …
attacks bots coming generated human internet language language models large llms machine malicious may media reliability social social media spearphishing text written