Feb. 8, 2024, 5:10 a.m. | Yvonne Zhou Mingyu Liang Ivan Brugere Dana Dachman-Soled Danial Dervovic Antigoni Polychroniadou Min W

cs.CR updates on arXiv.org arxiv.org

The growing use of machine learning (ML) has raised concerns that an ML model may reveal private information about an individual who has contributed to the training dataset. To prevent leakage of sensitive data, we consider using differentially-private (DP), synthetic training data instead of real training data to train an ML model. A key desirable property of synthetic data is its ability to preserve the low-order marginals of the original distribution. Our main contribution comprises novel upper and lower bounds …

contributed cs.cr cs.lg data dataset excess information linear machine machine learning may ml model private real reveal risk sensitive sensitive data synthetic synthetic data training training data

SOC 2 Manager, Audit and Certification

@ Deloitte | US and CA Multiple Locations

Open-Source Intelligence (OSINT) Policy Analyst (TS/SCI)

@ WWC Global | Reston, Virginia, United States

Security Architect (DevSecOps)

@ EUROPEAN DYNAMICS | Brussels, Brussels, Belgium

Infrastructure Security Architect

@ Ørsted | Kuala Lumpur, MY

Contract Penetration Tester

@ Evolve Security | United States - Remote

Senior Penetration Tester

@ DigitalOcean | Canada