May 2, 2024, 4:11 a.m. | Xi Xin, Giles Hooker, Fei Huang

cs.CR updates on arXiv.org arxiv.org

arXiv:2404.18702v2 Announce Type: replace-cross
Abstract: The adoption of artificial intelligence (AI) across industries has led to the widespread use of complex black-box models and interpretation tools for decision making. This paper proposes an adversarial framework to uncover the vulnerability of permutation-based interpretation methods for machine learning tasks, with a particular focus on partial dependence (PD) plots. This adversarial framework modifies the original black box model to manipulate its predictions for instances in the extrapolation domain. As a result, it produces …

adoption adversarial adversarial attacks artificial artificial intelligence arxiv attacks box cs.cr cs.lg decision decision making framework industries intelligence led machine machine learning making partial stat.ap stat.ml tools trust uncover vulnerability

Information Security Engineers

@ D. E. Shaw Research | New York City

Technology Security Analyst

@ Halton Region | Oakville, Ontario, Canada

Senior Cyber Security Analyst

@ Valley Water | San Jose, CA

Senior Product Delivery Associate - Cybersecurity | CyberOps

@ JPMorgan Chase & Co. | NY, United States

Security Ops Infrastructure Engineer (Remote US):

@ RingCentral | Remote, USA

SOC Analyst-1

@ NTT DATA | Bengaluru, India