all AI news
On the Detection of Anomalous or Out-Of-Distribution Data in Vision Models Using Statistical Techniques
March 26, 2024, 4:43 a.m. | Laura O'Mahony, David JP O'Sullivan, Nikola S. Nikolov
cs.LG updates on arXiv.org arxiv.org
Abstract: Out-of-distribution data and anomalous inputs are vulnerabilities of machine learning systems today, often causing systems to make incorrect predictions. The diverse range of data on which these models are used makes detecting atypical inputs a difficult and important task. We assess a tool, Benford's law, as a method used to quantify the difference between real and corrupted inputs. We believe that in many settings, it could function as a filter for anomalous data points and …
abstract arxiv cs.cv cs.lg data detection distribution diverse inputs learning systems machine machine learning predictions statistical systems type vision vision models vulnerabilities
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
#13721 - Data Engineer - AI Model Testing
@ Qualitest | Miami, Florida, United States
Elasticsearch Administrator
@ ManTech | 201BF - Customer Site, Chantilly, VA