all InfoSec news
Ignore This Title and HackAPrompt: Exposing Systemic Vulnerabilities of LLMs through a Global Scale Prompt Hacking Competition
March 5, 2024, 3:12 p.m. | Sander Schulhoff, Jeremy Pinto, Anaum Khan, Louis-Fran\c{c}ois Bouchard, Chenglei Si, Svetlina Anati, Valen Tagliabue, Anson Liu Kost, Christopher Car
cs.CR updates on arXiv.org arxiv.org
Abstract: Large Language Models (LLMs) are deployed in interactive contexts with direct user engagement, such as chatbots and writing assistants. These deployments are vulnerable to prompt injection and jailbreaking (collectively, prompt hacking), in which models are manipulated to ignore their original instructions and follow potentially malicious ones. Although widely acknowledged as a significant security threat, there is a dearth of large-scale resources and quantitative studies on prompt hacking. To address this lacuna, we launch a global …
arxiv chatbots competition cs.ai cs.cl cs.cr engagement exposing global hacking injection jailbreaking language language models large llms prompt prompt injection scale vulnerabilities vulnerable writing
More from arxiv.org / cs.CR updates on arXiv.org
IDEA: Invariant Defense for Graph Adversarial Robustness
1 day, 1 hour ago |
arxiv.org
FairCMS: Cloud Media Sharing with Fair Copyright Protection
1 day, 1 hour ago |
arxiv.org
Jobs in InfoSec / Cybersecurity
SOC 2 Manager, Audit and Certification
@ Deloitte | US and CA Multiple Locations
Senior InfoSec Manager - Risk and Compliance
@ Federal Reserve System | Remote - Virginia
Security Analyst
@ Fortra | Mexico
Incident Responder
@ Babcock | Chester, GB, CH1 6ER
Vulnerability, Access & Inclusion Lead
@ Monzo | Cardiff, London or Remote (UK)
Information Security Analyst
@ Unissant | MD, USA