all InfoSec news
MultiRobustBench: Benchmarking Robustness Against Multiple Attacks. (arXiv:2302.10980v1 [cs.LG])
cs.CR updates on arXiv.org arxiv.org
The bulk of existing research in defending against adversarial examples
focuses on defending against a single (typically bounded Lp-norm) attack, but
for a practical setting, machine learning (ML) models should be robust to a
wide variety of attacks. In this paper, we present the first unified framework
for considering multiple attacks against ML models. Our framework is able to
model different levels of learner's knowledge about the test-time adversary,
allowing us to model robustness against unforeseen attacks and robustness
against …
adversarial adversary attack attacks framework knowledge machine machine learning ml models research robustness single test