all AI news
Undersampling is a Minimax Optimal Robustness Intervention in Nonparametric Classification. (arXiv:2205.13094v1 [cs.LG])
May 27, 2022, 1:11 a.m. | Niladri S. Chatterji, Saminul Haque, Tatsunori Hashimoto
stat.ML updates on arXiv.org arxiv.org
While a broad range of techniques have been proposed to tackle distribution
shift, the simple baseline of training on an $\textit{undersampled}$ dataset
often achieves close to state-of-the-art-accuracy across several popular
benchmarks. This is rather surprising, since undersampling algorithms discard
excess majority group data. To understand this phenomenon, we ask if learning
is fundamentally constrained by a lack of minority group samples. We prove that
this is indeed the case in the setting of nonparametric binary classification.
Our results show that …
More from arxiv.org / stat.ML updates on arXiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Data Management Assistant
@ World Vision | Amman Office, Jordan
Cloud Data Engineer, Global Services Delivery, Google Cloud
@ Google | Buenos Aires, Argentina