all AI news
Provably Robust and Plausible Counterfactual Explanations for Neural Networks via Robust Optimisation
April 5, 2024, 4:42 a.m. | Junqi Jiang, Jianglin Lan, Francesco Leofante, Antonio Rago, Francesca Toni
cs.LG updates on arXiv.org arxiv.org
Abstract: Counterfactual Explanations (CEs) have received increasing interest as a major methodology for explaining neural network classifiers. Usually, CEs for an input-output pair are defined as data points with minimum distance to the input that are classified with a different label than the output. To tackle the established problem that CEs are easily invalidated when model parameters are updated (e.g. retrained), studies have proposed ways to certify the robustness of CEs under model parameter changes bounded …
abstract arxiv ces classifiers counterfactual cs.ai cs.lg data input-output major methodology network networks neural network neural networks optimisation robust type via
More from arxiv.org / cs.LG updates on arXiv.org
The Perception-Robustness Tradeoff in Deterministic Image Restoration
1 day, 12 hours ago |
arxiv.org
Jobs in AI, ML, Big Data
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne