May 31, 2023, 1:10 a.m. | Xuanqi Liu, Zhuotao Liu

cs.CR updates on arXiv.org arxiv.org

Prior works have attempted to build private inference frameworks for
transformer-based large language models (LLMs) in a server-client setting,
where the server holds the model parameters and the client inputs the private
data for inference. However, these frameworks impose significant overhead when
the private inputs are forward propagated through the original LLMs. In this
paper, we show that substituting the computation- and communication-heavy
operators in the transformer architecture with privacy-computing friendly
approximations can greatly reduce the private inference costs with …

build client computing data encrypted forward frameworks inputs language language models large llms privacy private private data server transformers understand

Cryptography Software Developer

@ Intel | USA - AZ - Chandler

Lead Consultant, Geology

@ WSP | Richmond, VA, United States

BISO Cybersecurity Director

@ ABM Industries | Alpharetta, GA, United States

TTECH Analista de ciberseguridad

@ Telefónica | LIMA, PE

TRANSCOM IGC - Cloud Security Engineer

@ IT Partners, Inc | St. Louis, Missouri, United States

Sr Cyber Threat Hunt Researcher

@ Peraton | Beltsville, MD, United States