all InfoSec news
Hijacking Large Language Models via Adversarial In-Context Learning
June 18, 2024, 4:20 a.m. | Yao Qiang, Xiangyu Zhou, Dongxiao Zhu
cs.CR updates on arXiv.org arxiv.org
Abstract: In-context learning (ICL) has emerged as a powerful paradigm leveraging LLMs for specific downstream tasks by utilizing labeled examples as demonstrations (demos) in the precondition prompts. Despite its promising performance, ICL suffers from instability with the choice and arrangement of examples. Additionally, crafted adversarial attacks pose a notable threat to the robustness of ICL. However, existing attacks are either easy to detect, rely on external models, or lack specificity towards ICL. This work introduces a …
adversarial adversarial attacks arxiv attacks context cs.cl cs.cr cs.lg examples hijacking instability language language models large llms paradigm performance prompts
More from arxiv.org / cs.CR updates on arXiv.org
Jobs in InfoSec / Cybersecurity
Information Technology Specialist I: Windows Engineer
@ Los Angeles County Employees Retirement Association (LACERA) | Pasadena, California
Information Technology Specialist I, LACERA: Information Security Engineer
@ Los Angeles County Employees Retirement Association (LACERA) | Pasadena, CA
Vice President, Controls Design & Development-7
@ State Street | Quincy, Massachusetts
Vice President, Controls Design & Development-5
@ State Street | Quincy, Massachusetts
Data Scientist & AI Prompt Engineer
@ Varonis | Israel
Contractor
@ Birlasoft | INDIA - MUMBAI - BIRLASOFT OFFICE, IN