April 15, 2024, 4:43 a.m. | Jiayi Huang, Sangwoo Park, Osvaldo Simeone

cs.LG updates on arXiv.org arxiv.org

arXiv:2305.07504v2 Announce Type: replace
Abstract: Deep learning models, including modern systems like large language models, are well known to offer unreliable estimates of the uncertainty of their decisions. In order to improve the quality of the confidence levels, also known as calibration, of a model, common approaches entail the addition of either data-dependent or data-independent regularization terms to the training loss. Data-dependent regularizers have been recently introduced in the context of conventional frequentist learning to penalize deviations between confidence and …

abstract arxiv bayesian confidence cs.lg data decisions deep learning eess.sp independent language language models large language large language models modern quality regularization systems terms type uncertainty

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Machine Learning Engineer

@ Apple | Sunnyvale, California, United States