all AI news
Optimal Learning via Moderate Deviations Theory
Feb. 15, 2024, 5:44 a.m. | Arnab Ganguly, Tobias Sutter
stat.ML updates on arXiv.org arxiv.org
Abstract: This paper proposes a statistically optimal approach for learning a function value using a confidence interval in a wide range of models, including general non-parametric estimation of an expected loss described as a stochastic programming problem or various SDE models. More precisely, we develop a systematic construction of highly accurate confidence intervals by using a moderate deviation principle-based approach. It is shown that the proposed confidence intervals are statistically optimal in the sense that they …
abstract arxiv confidence construction function general interval loss math.oc math.pr math.st non-parametric paper parametric programming stat.ml stat.th stochastic theory type value via
More from arxiv.org / stat.ML updates on arXiv.org
Inexact subgradient methods for semialgebraic functions
1 day, 12 hours ago |
arxiv.org
Online and Offline Robust Multivariate Linear Regression
1 day, 12 hours ago |
arxiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Senior Data Engineer
@ Quantexa | Sydney, New South Wales, Australia
Staff Analytics Engineer
@ Warner Bros. Discovery | NY New York 230 Park Avenue South