April 27, 2023, 1:10 a.m. | Souvik Kundu, Yuke Zhang, Dake Chen, Peter A. Beerel

cs.CR updates on arXiv.org arxiv.org

Large number of ReLU and MAC operations of Deep neural networks make them
ill-suited for latency and compute-efficient private inference. In this paper,
we present a model optimization method that allows a model to learn to be
shallow. In particular, we leverage the ReLU sensitivity of a convolutional
block to remove a ReLU layer and merge its succeeding and preceding convolution
layers to a shallow block. Unlike existing ReLU reduction methods, our joint
reduction method can yield models with improved …

block compute large latency learn mac making merge networks neural networks non operations optimization private remove

SOC 2 Manager, Audit and Certification

@ Deloitte | US and CA Multiple Locations

Information Security Engineers

@ D. E. Shaw Research | New York City

IAM Engineer - SailPoint IIQ

@ IDMWORKS | Remote USA

Manager, Network Security

@ NFL | New York City, United States

Engineering Team Manager – Security Controls

@ H&M Group | Stockholm, Sweden

Senior Security Consultant

@ LRQA | USA, US