all InfoSec news
Large Language Models are Advanced Anonymizers
Feb. 22, 2024, 5:11 a.m. | Robin Staab, Mark Vero, Mislav Balunovi\'c, Martin Vechev
cs.CR updates on arXiv.org arxiv.org
Abstract: Recent work in privacy research on large language models has shown that they achieve near human-level performance at inferring personal data from real-world online texts. With consistently increasing model capabilities, existing text anonymization methods are currently lacking behind regulatory requirements and adversarial threats. This raises the question of how individuals can effectively protect their personal data in sharing online texts. In this work, we take two steps to answer this question: We first present a …
advanced adversarial anonymization arxiv capabilities cs.ai cs.cl cs.cr cs.lg data human language language models large near performance personal personal data privacy question real regulatory requirements research text texts threats work world
More from arxiv.org / cs.CR updates on arXiv.org
Jobs in InfoSec / Cybersecurity
Technical Support Engineer - Cyber Security
@ Microsoft | Taipei, Taipei City, Taiwan
Senior Security Engineer
@ Workato | Barcelona, Spain
Regional Cybersecurity Specialist
@ Bayer | Singapore, Singapore, SG
Cyber Security Network Engineer
@ Nine | North Sydney, Australia
Professional, IAM Security
@ Ingram Micro | Manila Shared Services Center
Principal Windows Threat & Detection Security Researcher (Cortex)
@ Palo Alto Networks | Tel Aviv-Yafo, Israel