May 26, 2022, 1:11 a.m. | Yangfan Zhou, Kaizhu Huang, Cheng Cheng, Xuguang Wang, Amir Hussain, Xin Liu

stat.ML updates on arXiv.org arxiv.org

AdaBelief, one of the current best optimizers, demonstrates superior
generalization ability compared to the popular Adam algorithm by viewing the
exponential moving average of observed gradients. AdaBelief is theoretically
appealing in that it has a data-dependent $O(\sqrt{T})$ regret bound when
objective functions are convex, where $T$ is a time horizon. It remains however
an open problem whether the convergence rate can be further improved without
sacrificing its generalization ability. %on how to exploit strong convexity to
further improve the convergence …

arxiv belief convergence rate

Data Scientist (m/f/x/d)

@ Symanto Research GmbH & Co. KG | Spain, Germany

Senior Data Engineer x Analyst

@ QCP Capital | Singapore, Central Singapore, Singapore

Data Scientist Associate Sr - AI Core

@ JPMorgan Chase & Co. | Bengaluru, Karnataka, India

Mgr, Machine Learning Quality Engineering Mgmt

@ ServiceNow | Hyderabad, India

Software Engineer III - Data Analytics

@ JPMorgan Chase & Co. | Hyderabad, Telangana, India

Senior Data Analyst

@ Life360 | Remote, USA or Remote, Canada