Web: http://arxiv.org/abs/2106.15419

May 4, 2022, 1:12 a.m. | Zhikang T. Wang, Masahito Ueda

cs.LG updates on arXiv.org arxiv.org

Despite the empirical success of the deep Q network (DQN) reinforcement
learning algorithm and its variants, DQN is still not well understood and it
does not guarantee convergence. In this work, we show that DQN can indeed
diverge and cease to operate in realistic settings. Although there exist
gradient-based convergent methods, we show that they actually have inherent
problems in learning dynamics which cause them to fail even in simple tasks. To
overcome these problems, we propose a convergent DQN …

algorithm arxiv deep network

More from arxiv.org / cs.LG updates on arXiv.org

Director, Applied Mathematics & Computational Research Division

@ Lawrence Berkeley National Lab | Berkeley, Ca

Business Data Analyst

@ MainStreet Family Care | Birmingham, AL

Assistant/Associate Professor of the Practice in Business Analytics

@ Georgetown University McDonough School of Business | Washington DC

Senior Data Science Writer

@ NannyML | Remote

Director of AI/ML Engineering

@ Armis Industries | Remote (US only), St. Louis, California

Digital Analytics Manager

@ Patagonia | Ventura, California