Feb. 9, 2024, 5:10 a.m. | Jamie Hayes Ilia Shumailov Itay Yona

cs.CR updates on arXiv.org arxiv.org

Mixture of Experts (MoE) has become a key ingredient for scaling large foundation models while keeping inference costs steady. We show that expert routing strategies that have cross-batch dependencies are vulnerable to attacks. Malicious queries can be sent to a model and can affect a model's output on other benign queries if they are grouped in the same batch. We demonstrate this via a proof-of-concept attack in a toy experimental setting.

attacks batch buffer buffer overflow can cs.cr cs.lg dependencies expert experts foundation foundation models key large malicious overflow routing scaling strategies vulnerable

SOC 2 Manager, Audit and Certification

@ Deloitte | US and CA Multiple Locations

Network Security Engineer – Zscaler SME

@ Peraton | United States

Splunk Data Analytic Subject Matter Expert

@ Peraton | Woodlawn, MD, United States

Principal Consultant, Offensive Security, Proactive Services (Unit 42)- Remote

@ Palo Alto Networks | Santa Clara, CA, United States

Senior Engineer Software Product Security

@ Ford Motor Company | Mexico City, MEX, Mexico

Information System Security Engineer (Red Team)

@ Evolution | Riga, Latvia