all AI news
Detecting AutoAttack Perturbations in the Frequency Domain
Feb. 21, 2024, 5:46 a.m. | Peter Lorenz, Paula Harder, Dominik Strassel, Margret Keuper, Janis Keuper
cs.CV updates on arXiv.org arxiv.org
Abstract: Recently, adversarial attacks on image classification networks by the AutoAttack (Croce and Hein, 2020b) framework have drawn a lot of attention. While AutoAttack has shown a very high attack success rate, most defense approaches are focusing on network hardening and robustness enhancements, like adversarial training. This way, the currently best-reported method can withstand about 66% of adversarial examples on CIFAR10. In this paper, we investigate the spatial and frequency domain properties of AutoAttack and propose …
abstract adversarial adversarial attacks adversarial training arxiv attacks attention classification cs.cr cs.cv defense domain framework image network networks rate robustness success training type
More from arxiv.org / cs.CV updates on arXiv.org
Jobs in AI, ML, Big Data
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US
Research Engineer
@ Allora Labs | Remote
Ecosystem Manager
@ Allora Labs | Remote
Founding AI Engineer, Agents
@ Occam AI | New York