all InfoSec news
An Adversarial Approach to Evaluating the Robustness of Event Identification Models
Feb. 20, 2024, 5:11 a.m. | Obai Bahwal, Oliver Kosut, Lalitha Sankar
cs.CR updates on arXiv.org arxiv.org
Abstract: Intelligent machine learning approaches are finding active use for event detection and identification that allow real-time situational awareness. Yet, such machine learning algorithms have been shown to be susceptible to adversarial attacks on the incoming telemetry data. This paper considers a physics-based modal decomposition method to extract features for event classification and focuses on interpretable classifiers including logistic regression and gradient boosting to distinguish two types of events: load loss and generation loss. The resulting …
adversarial adversarial attacks algorithms arxiv attacks awareness cs.cr cs.lg cs.sy data detection eess.sy event identification machine machine learning machine learning algorithms modal physics real robustness telemetry
More from arxiv.org / cs.CR updates on arXiv.org
Jobs in InfoSec / Cybersecurity
CyberSOC Technical Lead
@ Integrity360 | Sandyford, Dublin, Ireland
Cyber Security Strategy Consultant
@ Capco | New York City
Cyber Security Senior Consultant
@ Capco | Chicago, IL
Sr. Product Manager
@ MixMode | Remote, US
Security Compliance Strategist
@ Grab | Petaling Jaya, Malaysia
Cloud Security Architect, Lead
@ Booz Allen Hamilton | USA, VA, McLean (1500 Tysons McLean Dr)