Feb. 19, 2024, 5:10 a.m. | R. Patrick Xian, Alex J. Lee, Vincent Wang, Qiming Cui, Russell Ro, Reza Abbasi-Asl

cs.CR updates on arXiv.org arxiv.org

arXiv:2402.10527v1 Announce Type: cross
Abstract: The increasing depth of parametric domain knowledge in large language models (LLMs) is fueling their rapid deployment in real-world applications. In high-stakes and knowledge-intensive tasks, understanding model vulnerabilities is essential for quantifying the trustworthiness of model predictions and regulating their use. The recent discovery of named entities as adversarial examples in natural language processing tasks raises questions about their potential guises in other settings. Here, we propose a powerscaled distance-weighted sampling scheme in embedding space …

adversarial applications arxiv biomedical cs.cl cs.cr deployment discovery domain entities high knowledge language language models large llms predictions question rapid real stat.ap trustworthiness understanding vulnerabilities world

Enterprise Security Architect

@ Proofpoint | Utah

Senior Incident Response and Digital Forensics Engineer

@ Danske Bank | Vilnius, Lithuania

SOC Analyst (Remote)

@ Bertelsmann | New York City, US, 10019

Risk Consulting - Protect Tech - Staff - IT Compliance - ISO-NIST-FISMA-PCI DSS and Privacy

@ EY | Bengaluru, KA, IN, 560016

Security Officer Warrenpoint Harbour

@ TSS | Newry, County Down, United Kingdom

Senior DevSecOps Engineer

@ Scientific Systems Company, Inc. | Burlington, Massachusetts, United States