Feb. 27, 2024, 5:46 a.m. | Yanqi Qiao, Dazhuang Liu, Rui Wang, Kaitai Liang

cs.CV updates on arXiv.org arxiv.org

arXiv:2402.15653v1 Announce Type: new
Abstract: While convolutional neural networks (CNNs) have achieved success in computer vision tasks, it is vulnerable to backdoor attacks. Such attacks could mislead the victim model to make attacker-chosen prediction with a specific trigger pattern. Until now, the trigger injection of existing attacks is mainly limited to spatial domain. Recent works take advantage of perceptual properties of planting specific patterns in the frequency domain, which only reflect indistinguishable pixel-wise perturbations in pixel domain. However, in the …

abstract algorithm arxiv attacks backdoor box cnns computer computer vision convolutional neural networks cs.cv domain low networks neural networks prediction spatial success tasks type via vision vulnerable

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Data Scientist (Database Development)

@ Nasdaq | Bengaluru-Affluence