Nov. 5, 2023, 6:10 a.m. | Sam Toyer, Olivia Watkins, Ethan Adrian Mendes, Justin Svegliato, Luke Bailey, Tiffany Wang, Isaac Ong, Karim Elmaaroufi, Pieter Abbeel, Trevor Darrel

cs.CR updates on arXiv.org arxiv.org

While Large Language Models (LLMs) are increasingly being used in real-world
applications, they remain vulnerable to prompt injection attacks: malicious
third party prompts that subvert the intent of the system designer. To help
researchers study this problem, we present a dataset of over 126,000 prompt
injection attacks and 46,000 prompt-based "defenses" against prompt injection,
all created by players of an online game called Tensor Trust. To the best of
our knowledge, this is currently the largest dataset of human-generated
adversarial …

applications attacks dataset designer game injection injection attacks intent language language models large llms malicious party problem prompt prompt injection prompt injection attacks prompts real researchers study system tensor third trust vulnerable world

EY- GDS- Cybersecurity- Staff

@ EY | Miguel Hidalgo, MX, 11520

Staff Security Operations Engineer

@ Workiva | Ames

Public Relations Senior Account Executive (B2B Tech/Cybersecurity/Enterprise)

@ Highwire Public Relations | Los Angeles, CA

Airbus Canada - Responsable Cyber sécurité produit / Product Cyber Security Responsible

@ Airbus | Mirabel

Investigations (OSINT) Manager

@ Logically | India

Security Engineer I, Offensive Security Penetration Testing

@ Amazon.com | US, NY, Virtual Location - New York