March 26, 2024, 4:11 a.m. | Md Abdul Kadir, GowthamKrishna Addluri, Daniel Sonntag

cs.CR updates on arXiv.org arxiv.org

arXiv:2403.16569v1 Announce Type: cross
Abstract: Explainable Artificial Intelligence (XAI) strategies play a crucial part in increasing the understanding and trustworthiness of neural networks. Nonetheless, these techniques could potentially generate misleading explanations. Blinding attacks can drastically alter a machine learning algorithm's prediction and explanation, providing misleading information by adding visually unnoticeable artifacts into the input, while maintaining the model's accuracy. It poses a serious challenge in ensuring the reliability of XAI methods. To ensure the reliability of XAI methods poses a …

algorithm artificial artificial intelligence arxiv attacks aware backdoors can cs.cr cs.cv cs.lg defense information intelligence machine machine learning networks neural networks parameter play prediction strategies techniques trustworthiness understanding vulnerabilities xai

Information Security Engineers

@ D. E. Shaw Research | New York City

Technology Security Analyst

@ Halton Region | Oakville, Ontario, Canada

Senior Cyber Security Analyst

@ Valley Water | San Jose, CA

Sr. Staff Firmware Engineer – Networking & Firewall

@ Axiado | Bengaluru, India

Compliance Architect / Product Security Sr. Engineer/Expert (f/m/d)

@ SAP | Walldorf, DE, 69190

SAP Security Administrator

@ FARO Technologies | EMEA-Portugal