all InfoSec news
Holding Secrets Accountable: Auditing Privacy-Preserving Machine Learning
Feb. 27, 2024, 5:11 a.m. | Hidde Lycklama, Alexander Viand, Nicolas K\"uchler, Christian Knabenhans, Anwar Hithnawi
cs.CR updates on arXiv.org arxiv.org
Abstract: Recent advancements in privacy-preserving machine learning are paving the way to extend the benefits of ML to highly sensitive data that, until now, have been hard to utilize due to privacy concerns and regulatory constraints. Simultaneously, there is a growing emphasis on enhancing the transparency and accountability of machine learning, including the ability to audit ML deployments. While ML auditing and PPML have both been the subjects of intensive research, they have predominately been examined …
accountability arxiv auditing benefits constraints cs.cr data hard machine machine learning privacy privacy concerns regulatory secrets sensitive sensitive data transparency transparency and accountability
More from arxiv.org / cs.CR updates on arXiv.org
Jobs in InfoSec / Cybersecurity
Information Security Engineers
@ D. E. Shaw Research | New York City
Technology Security Analyst
@ Halton Region | Oakville, Ontario, Canada
Senior Cyber Security Analyst
@ Valley Water | San Jose, CA
Sr. Staff Firmware Engineer – Networking & Firewall
@ Axiado | Bengaluru, India
Compliance Architect / Product Security Sr. Engineer/Expert (f/m/d)
@ SAP | Walldorf, DE, 69190
SAP Security Administrator
@ FARO Technologies | EMEA-Portugal