all AI news
Modeling Attrition in Recommender Systems with Departing Bandits
Feb. 19, 2024, 5:43 a.m. | Omer Ben-Porat, Lee Cohen, Liu Leqi, Zachary C. Lipton, Yishay Mansour
cs.LG updates on arXiv.org arxiv.org
Abstract: Traditionally, when recommender systems are formalized as multi-armed bandits, the policy of the recommender system influences the rewards accrued, but not the length of interaction. However, in real-world systems, dissatisfied users may depart (and never come back). In this work, we propose a novel multi-armed bandit setup that captures such policy-dependent horizons. Our setup consists of a finite set of user types, and multiple arms with Bernoulli payoffs. Each (user type, arm) tuple corresponds to …
abstract arxiv attrition cs.ir cs.lg modeling multi-armed bandits novel policy recommender systems setup stat.ml systems type work world
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Senior Machine Learning Engineer
@ GPTZero | Toronto, Canada
ML/AI Engineer / NLP Expert - Custom LLM Development (x/f/m)
@ HelloBetter | Remote
Doctoral Researcher (m/f/div) in Automated Processing of Bioimages
@ Leibniz Institute for Natural Product Research and Infection Biology (Leibniz-HKI) | Jena
Seeking Developers and Engineers for AI T-Shirt Generator Project
@ Chevon Hicks | Remote
Director, Venture Capital - Artificial Intelligence
@ Condé Nast | San Jose, CA
Senior Molecular Imaging Expert (Senior Principal Scientist)
@ University of Sydney | Cambridge (USA)