all AI news
Pessimism for Offline Linear Contextual Bandits using $\ell_p$ Confidence Sets. (arXiv:2205.10671v2 [cs.LG] UPDATED)
Oct. 6, 2022, 1:13 a.m. | Gene Li, Cong Ma, Nathan Srebro
stat.ML updates on arXiv.org arxiv.org
We present a family $\{\hat{\pi}\}_{p\ge 1}$ of pessimistic learning rules
for offline learning of linear contextual bandits, relying on confidence sets
with respect to different $\ell_p$ norms, where $\hat{\pi}_2$ corresponds to
Bellman-consistent pessimism (BCP), while $\hat{\pi}_\infty$ is a novel
generalization of lower confidence bound (LCB) to the linear setting. We show
that the novel $\hat{\pi}_\infty$ learning rule is, in a sense, adaptively
optimal, as it achieves the minimax performance (up to log factors) against all
$\ell_q$-constrained problems, and as such …
More from arxiv.org / stat.ML updates on arXiv.org
Jobs in AI, ML, Big Data
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Data Analyst
@ SEAKR Engineering | Englewood, CO, United States
Data Analyst II
@ Postman | Bengaluru, India
Data Architect
@ FORSEVEN | Warwick, GB
Director, Data Science
@ Visa | Washington, DC, United States
Senior Manager, Data Science - Emerging ML
@ Capital One | McLean, VA