March 13, 2024, 4:44 a.m. | Meir Yossef Levi, Guy Gilboa

cs.LG updates on arXiv.org arxiv.org

arXiv:2308.05525v3 Announce Type: replace-cross
Abstract: The ability to cope with out-of-distribution (OOD) corruptions and adversarial attacks is crucial in real-world safety-demanding applications. In this study, we develop a general mechanism to increase neural network robustness based on focus analysis.
Recent studies have revealed the phenomenon of \textit{Overfocusing}, which leads to a performance drop. When the network is primarily influenced by small input regions, it becomes less robust and prone to misclassify under noise and corruptions.
However, quantifying overfocusing is still …

abstract adversarial adversarial attacks analysis applications arxiv attacks cloud cs.cv cs.lg distribution focus general leads network networks neural network performance robustness safety studies study type world

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Global Data Architect, AVP - State Street Global Advisors

@ State Street | Boston, Massachusetts

Data Engineer

@ NTT DATA | Pune, MH, IN