all InfoSec news
On the Privacy Effect of Data Enhancement via the Lens of Memorization. (arXiv:2208.08270v1 [cs.LG])
Aug. 18, 2022, 1:20 a.m. | Xiao Li, Qiongxiu Li, Zhanhao Hu, Xiaolin Hu
cs.CR updates on arXiv.org arxiv.org
Machine learning poses severe privacy concerns as it is shown that the
learned models can reveal sensitive information about their training data. Many
works have investigated the effect of widely-adopted data augmentation (DA) and
adversarial training (AT) techniques, termed data enhancement in the paper, on
the privacy leakage of machine learning models. Such privacy effects are often
measured by membership inference attacks (MIAs), which aim to identify whether
a particular example belongs to the training set or not. We propose …
More from arxiv.org / cs.CR updates on arXiv.org
Jobs in InfoSec / Cybersecurity
Social Engineer For Reverse Engineering Exploit Study
@ Independent study | Remote
Security Engineer II- Full stack Java with React
@ JPMorgan Chase & Co. | Hyderabad, Telangana, India
Cybersecurity SecOps
@ GFT Technologies | Mexico City, MX, 11850
Senior Information Security Advisor
@ Sun Life | Sun Life Toronto One York
Contract Special Security Officer (CSSO) - Top Secret Clearance
@ SpaceX | Hawthorne, CA
Early Career Cyber Security Operations Center (SOC) Analyst
@ State Street | Quincy, Massachusetts