all InfoSec news
An Adversarial Approach to Evaluating the Robustness of Event Identification Models
Feb. 20, 2024, 5:11 a.m. | Obai Bahwal, Oliver Kosut, Lalitha Sankar
cs.CR updates on arXiv.org arxiv.org
Abstract: Intelligent machine learning approaches are finding active use for event detection and identification that allow real-time situational awareness. Yet, such machine learning algorithms have been shown to be susceptible to adversarial attacks on the incoming telemetry data. This paper considers a physics-based modal decomposition method to extract features for event classification and focuses on interpretable classifiers including logistic regression and gradient boosting to distinguish two types of events: load loss and generation loss. The resulting …
adversarial adversarial attacks algorithms arxiv attacks awareness cs.cr cs.lg cs.sy data detection eess.sy event identification machine machine learning machine learning algorithms modal physics real robustness telemetry
More from arxiv.org / cs.CR updates on arXiv.org
IDEA: Invariant Defense for Graph Adversarial Robustness
1 day, 2 hours ago |
arxiv.org
FairCMS: Cloud Media Sharing with Fair Copyright Protection
1 day, 2 hours ago |
arxiv.org
Jobs in InfoSec / Cybersecurity
SOC 2 Manager, Audit and Certification
@ Deloitte | US and CA Multiple Locations
Senior InfoSec Manager - Risk and Compliance
@ Federal Reserve System | Remote - Virginia
Security Analyst
@ Fortra | Mexico
Incident Responder
@ Babcock | Chester, GB, CH1 6ER
Vulnerability, Access & Inclusion Lead
@ Monzo | Cardiff, London or Remote (UK)
Information Security Analyst
@ Unissant | MD, USA