all InfoSec news
Unlocking High-Accuracy Differentially Private Image Classification through Scale. (arXiv:2204.13650v2 [cs.LG] UPDATED)
June 17, 2022, 1:20 a.m. | Soham De, Leonard Berrada, Jamie Hayes, Samuel L. Smith, Borja Balle
cs.CR updates on arXiv.org arxiv.org
Differential Privacy (DP) provides a formal privacy guarantee preventing
adversaries with access to a machine learning model from extracting information
about individual training points. Differentially Private Stochastic Gradient
Descent (DP-SGD), the most popular DP training method for deep learning,
realizes this protection by injecting noise during training. However previous
works have found that DP-SGD often leads to a significant degradation in
performance on standard image classification benchmarks. Furthermore, some
authors have postulated that DP-SGD inherently performs poorly on large models, …
More from arxiv.org / cs.CR updates on arXiv.org
Jobs in InfoSec / Cybersecurity
SOC 2 Manager, Audit and Certification
@ Deloitte | US and CA Multiple Locations
Senior Security Specialist
@ Lely | Maassluis, Netherlands
IT Security Manager (Corporate Security) (REF822R)
@ Deutsche Telekom IT Solutions | Budapest, Hungary
Senior Security Architect
@ Cassa Centrale Banca - Credito Cooperativo Italiano | Trento, IT, 38122
Senior DevSecOps Engineer
@ Raft | Las Vegas, NV (Remote)
Product Manager - Compliance
@ Arctic Wolf | Remote - Colorado