Web: http://arxiv.org/abs/2102.07711

May 5, 2022, 1:12 a.m. | Anshuka Rangi, Long Tran-Thanh, Haifeng Xu, Massimo Franceschetti

cs.LG updates on arXiv.org arxiv.org

We study bandit algorithms under data poisoning attacks in a bounded reward
setting. We consider a strong attacker model in which the attacker can observe
both the selected actions and their corresponding rewards and can contaminate
the rewards with additive noise. We show that any bandit algorithm with regret
$O(\log T)$ can be forced to suffer a regret $\Omega(T)$ with an expected
amount of contamination $O(\log T)$. This amount of contamination is also
necessary, as we prove that there exists …

arxiv attacks data saving stochastic verification

More from arxiv.org / cs.LG updates on arXiv.org

Director, Applied Mathematics & Computational Research Division

@ Lawrence Berkeley National Lab | Berkeley, Ca

Business Data Analyst

@ MainStreet Family Care | Birmingham, AL

Assistant/Associate Professor of the Practice in Business Analytics

@ Georgetown University McDonough School of Business | Washington DC

Senior Data Science Writer

@ NannyML | Remote

Director of AI/ML Engineering

@ Armis Industries | Remote (US only), St. Louis, California

Digital Analytics Manager

@ Patagonia | Ventura, California