all InfoSec news
XorSHAP: Privacy-Preserving Explainable AI for Decision Tree Models
Dec. 6, 2023, 3:36 p.m. |
IACR News www.iacr.org
ePrint Report: XorSHAP: Privacy-Preserving Explainable AI for Decision Tree Models
Dimitar Jetchev, Marius Vuille
Explainable AI (XAI) refers to the development of AI systems and machine learning models in a way that humans can understand, interpret and trust the predictions, decisions and outputs of these models. A common approach to explainability is feature importance, that is, determining which input features of the model have the most significant impact on the model prediction. Two major techniques for computing feature importance are …
decision development eprint report explainable ai humans machine machine learning machine learning models predictions privacy report systems trust understand xai
More from www.iacr.org / IACR News
$\mathsf{FRAST}$: TFHE-friendly Cipher Based on Random S-boxes
3 days, 12 hours ago |
www.iacr.org
A Deniability Analysis of Signal's Initial Handshake PQXDH
3 days, 12 hours ago |
www.iacr.org
BGJ15 Revisited: Sieving with Streamed Memory Access
3 days, 12 hours ago |
www.iacr.org
Quantum Key-Revocable Dual-Regev Encryption, Revisited
3 days, 12 hours ago |
www.iacr.org
Jobs in InfoSec / Cybersecurity
Information Security Engineers
@ D. E. Shaw Research | New York City
Technology Security Analyst
@ Halton Region | Oakville, Ontario, Canada
Senior Cyber Security Analyst
@ Valley Water | San Jose, CA
Computer and Forensics Investigator
@ ManTech | 221BQ - Cstmr Site,Springfield,VA
Senior Security Analyst
@ Oracle | United States
Associate Vulnerability Management Specialist
@ Diebold Nixdorf | Hyderabad, Telangana, India