all AI news
Off-policy evaluation for learning-to-rank via interpolating the item-position model and the position-based model. (arXiv:2210.09512v1 [cs.LG])
Oct. 19, 2022, 1:11 a.m. | Alexander Buchholz, Ben London, Giuseppe di Benedetto, Thorsten Joachims
cs.LG updates on arXiv.org arxiv.org
A critical need for industrial recommender systems is the ability to evaluate
recommendation policies offline, before deploying them to production.
Unfortunately, widely used off-policy evaluation methods either make strong
assumptions about how users behave that can lead to excessive bias, or they
make fewer assumptions and suffer from large variance. We tackle this problem
by developing a new estimator that mitigates the problems of the two most
popular off-policy estimators for rankings, namely the position-based model and
the item-position model. …
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
#13721 - Data Engineer - AI Model Testing
@ Qualitest | Miami, Florida, United States
Elasticsearch Administrator
@ ManTech | 201BF - Customer Site, Chantilly, VA