all InfoSec news
BadEdit: Backdooring large language models by model editing
March 21, 2024, 4:10 a.m. | Yanzhou Li, Tianlin Li, Kangjie Chen, Jian Zhang, Shangqing Liu, Wenhan Wang, Tianwei Zhang, Yang Liu
cs.CR updates on arXiv.org arxiv.org
Abstract: Mainstream backdoor attack methods typically demand substantial tuning data for poisoning, limiting their practicality and potentially degrading the overall performance when applied to Large Language Models (LLMs). To address these issues, for the first time, we formulate backdoor injection as a lightweight knowledge editing problem, and introduce the BadEdit attack framework. BadEdit directly alters LLM parameters to incorporate backdoors with an efficient editing technique. It boasts superiority over existing backdoor injection techniques in several areas: …
address arxiv attack backdoor backdoor attack backdooring cs.ai cs.cr data demand editing injection knowledge language language models large llms mainstream performance poisoning problem
More from arxiv.org / cs.CR updates on arXiv.org
Jobs in InfoSec / Cybersecurity
Digital Security Infrastructure Manager
@ Wizz Air | Budapest, HU, H-1103
Sr. Solution Consultant
@ Highspot | Sydney
Cyber Security Analyst III
@ Love's Travel Stops | Oklahoma City, OK, US, 73120
Lead Security Engineer
@ JPMorgan Chase & Co. | Tampa, FL, United States
GTI Manager of Cybersecurity Operations
@ Grant Thornton | Tulsa, OK, United States
GCP Incident Response Engineer
@ Publicis Groupe | Dallas, Texas, United States