May 21, 2024, 4:12 a.m. | Yuhao Zhang, Aws Albarghouthi, Loris D'Antoni

cs.CR updates on arXiv.org arxiv.org

arXiv:2301.11824v4 Announce Type: replace
Abstract: Neural networks are vulnerable to backdoor poisoning attacks, where the attackers maliciously poison the training set and insert triggers into the test input to change the prediction of the victim model. Existing defenses for backdoor attacks either provide no formal guarantees or come with expensive-to-compute and ineffective probabilistic guarantees. We present PECAN, an efficient and certified approach for defending against backdoor attacks. The key insight powering PECAN is to apply off-the-shelf test-time evasion certification techniques …

arxiv attackers attacks backdoor backdoor attacks certified change compute cs.cr cs.lg defense defenses input networks neural networks poisoning poisoning attacks prediction test training victim vulnerable

CyberSOC Technical Lead

@ Integrity360 | Sandyford, Dublin, Ireland

Cyber Security Strategy Consultant

@ Capco | New York City

Cyber Security Senior Consultant

@ Capco | Chicago, IL

Sr. Product Manager

@ MixMode | Remote, US

Sr. Manager, Information Security

@ Saviynt | Bengaluru

Director - Information Security

@ Saviynt | Bengaluru