all InfoSec news
A Survey of Backdoor Attacks and Defenses on Large Language Models: Implications for Security Measures
June 12, 2024, 4:11 a.m. | Shuai Zhao, Meihuizi Jia, Zhongliang Guo, Leilei Gan, Jie Fu, Yichao Feng, Fengjun Pan, Luu Anh Tuan
cs.CR updates on arXiv.org arxiv.org
Abstract: The large language models (LLMs), which bridge the gap between human language understanding and complex problem-solving, achieve state-of-the-art performance on several NLP tasks, particularly in few-shot and zero-shot settings. Despite the demonstrable efficacy of LMMs, due to constraints on computational resources, users have to engage with open-source language models or outsource the entire training process to third-party platforms. However, research has demonstrated that language models are susceptible to potential security vulnerabilities, particularly in backdoor attacks. …
art arxiv attacks backdoor backdoor attacks bridge computational constraints cs.ai cs.cl cs.cr defenses gap human language language models large llms nlp performance problem problem-solving security security measures settings state survey understanding
More from arxiv.org / cs.CR updates on arXiv.org
Jobs in InfoSec / Cybersecurity
Senior Streaming Platform Engineer
@ Armis Security | Tel Aviv-Yafo, Tel Aviv District, Israel
Senior Streaming Platform Engineer
@ Armis Security | Tel Aviv-Yafo, Tel Aviv District, Israel
Deputy Chief Information Officer of Operations (Senior Public Service Administrator, Opt. 3)
@ State of Illinois | Springfield, IL, US, 62701-1222
Deputy Chief Information Officer of Operations (Senior Public Service Administrator, Opt. 3)
@ State of Illinois | Springfield, IL, US, 62701-1222
Analyst, Security
@ DailyPay | New York City
Analyst, Security
@ DailyPay | New York City