all AI news
Do Counterfactual Examples Complicate Adversarial Training?
April 17, 2024, 4:42 a.m. | Eric Yeats, Cameron Darwin, Eduardo Ortega, Frank Liu, Hai Li
cs.LG updates on arXiv.org arxiv.org
Abstract: We leverage diffusion models to study the robustness-performance tradeoff of robust classifiers. Our approach introduces a simple, pretrained diffusion method to generate low-norm counterfactual examples (CEs): semantically altered data which results in different true class membership. We report that the confidence and accuracy of robust models on their clean training data are associated with the proximity of the data to their CEs. Moreover, robust models perform very poorly when evaluated on the CEs directly, as …
abstract accuracy adversarial adversarial training arxiv ces class classifiers confidence counterfactual cs.cv cs.lg data diffusion diffusion models examples generate low norm performance report results robust robust models robustness simple study training true type
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
#13721 - Data Engineer - AI Model Testing
@ Qualitest | Miami, Florida, United States
Elasticsearch Administrator
@ ManTech | 201BF - Customer Site, Chantilly, VA