Feb. 24, 2022, 2:11 a.m. | Yikun Ban, Yunzhe Qi, Tianxin Wei, Jingrui He

cs.LG updates on arXiv.org arxiv.org

Contextual multi-armed bandits provide powerful tools to solve the
exploitation-exploration dilemma in decision making, with direct applications
in the personalized recommendation. In fact, collaborative effects among users
carry the significant potential to improve the recommendation. In this paper,
we introduce and study the problem by exploring `Neural Collaborative Filtering
Bandits', where the rewards can be non-linear functions and groups are formed
dynamically given different specific contents. To solve this problem, inspired
by meta-learning, we propose Meta-Ban (meta-bandits), where a meta-learner …

arxiv collaborative collaborative filtering learning meta

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US