Feb. 24, 2022, 2:11 a.m. | Yikun Ban, Yunzhe Qi, Tianxin Wei, Jingrui He

cs.LG updates on arXiv.org arxiv.org

Contextual multi-armed bandits provide powerful tools to solve the
exploitation-exploration dilemma in decision making, with direct applications
in the personalized recommendation. In fact, collaborative effects among users
carry the significant potential to improve the recommendation. In this paper,
we introduce and study the problem by exploring `Neural Collaborative Filtering
Bandits', where the rewards can be non-linear functions and groups are formed
dynamically given different specific contents. To solve this problem, inspired
by meta-learning, we propose Meta-Ban (meta-bandits), where a meta-learner …

arxiv collaborative collaborative filtering learning meta

Senior Machine Learning Engineer

@ GPTZero | Toronto, Canada

ML/AI Engineer / NLP Expert - Custom LLM Development (x/f/m)

@ HelloBetter | Remote

Doctoral Researcher (m/f/div) in Automated Processing of Bioimages

@ Leibniz Institute for Natural Product Research and Infection Biology (Leibniz-HKI) | Jena

Seeking Developers and Engineers for AI T-Shirt Generator Project

@ Chevon Hicks | Remote

Senior Machine Learning Engineer

@ BlackStone eIT | Egypt - Remote

Machine Learning Engineer - 2

@ Parspec | Bengaluru, India