Jan. 4, 2022, 2:20 a.m. | Harrison Foley, Liam Fowl, Tom Goldstein, Gavin Taylor

cs.CR updates on arXiv.org arxiv.org

Data poisoning for reinforcement learning has historically focused on general
performance degradation, and targeted attacks have been successful via
perturbations that involve control of the victim's policy and rewards. We
introduce an insidious poisoning attack for reinforcement learning which causes
agent misbehavior only at specific target states - all while minimally
modifying a small fraction of training observations without assuming any
control over policy or reward. We accomplish this by adapting a recent
technique, gradient alignment, to reinforcement learning. We …

data order poisoning

SOC 2 Manager, Audit and Certification

@ Deloitte | US and CA Multiple Locations

Information Security Engineers

@ D. E. Shaw Research | New York City

Dir-Information Security - Cyber Analytics

@ Marriott International | Bethesda, MD, United States

Security Engineer - Security Operations

@ TravelPerk | Barcelona, Barcelona, Spain

Information Security Mgmt- Risk Assessor

@ JPMorgan Chase & Co. | Bengaluru, Karnataka, India

SAP CO Consultant

@ Atos | Istanbul, TR