April 12, 2023, 1:10 a.m. | Haoran Li, Dadi Guo, Wei Fan, Mingshi Xu, Yangqiu Song

cs.CR updates on arXiv.org arxiv.org

With the rapid progress of large language models (LLMs), many downstream NLP
tasks can be well solved given good prompts. Though model developers and
researchers work hard on dialog safety to avoid generating harmful content from
LLMs, it is still challenging to steer AI-generated content (AIGC) for the
human good. As powerful LLMs are devouring existing text data from various
domains (e.g., GPT-3 is trained on 45TB texts), it is natural to doubt whether
the private information is included in …

attacks chatgpt data developers dialog domains generated gpt gpt-3 hard human information jailbreaking language language models large llms nlp privacy progress prompts rapid researchers safety text threats training work

Red Team Operator

@ JPMorgan Chase & Co. | LONDON, United Kingdom

SOC Analyst

@ Resillion | Bengaluru, India

Director of Cyber Security

@ Revinate | San Francisco Bay Area

Jr. Security Incident Response Analyst

@ Kaseya | Miami, Florida, United States

Infrastructure Vulnerability Consultant - (Cloud Security , CSPM)

@ Blue Yonder | Hyderabad

Product Security Lead

@ Lely | Maassluis, Netherlands