Feb. 13, 2024, 5:44 a.m. | Lukas Muttenthaler Robert A. Vandermeulen Qiuyi Zhang Thomas Unterthiner Klaus-Robert M\"uller

cs.LG updates on arXiv.org arxiv.org

Model overconfidence and poor calibration are common in machine learning and difficult to account for when applying standard empirical risk minimization. In this work, we propose a novel method to alleviate these problems that we call odd-$k$-out learning (OKO), which minimizes the cross-entropy error for sets rather than for single examples. This naturally allows the model to capture correlations across data examples and achieves both better accuracy and calibration, especially in limited training data and class-imbalanced regimes. Perhaps surprisingly, OKO …

call cross-entropy cs.cv cs.it cs.lg entropy error examples machine machine learning math.it novel risk set standard work

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne