Feb. 6, 2024, 5:46 a.m. | Sobihan SurendranLPSM Antoine Godichon-BaggioniLPSM Adeline FermanianLPSM Sylvain Le CorffLPSM

cs.LG updates on arXiv.org arxiv.org

Stochastic Gradient Descent (SGD) with adaptive steps is now widely used for training deep neural networks. Most theoretical results assume access to unbiased gradient estimators, which is not the case in several recent deep learning and reinforcement learning applications that use Monte Carlo methods. This paper provides a comprehensive non-asymptotic analysis of SGD with biased gradients and adaptive steps for convex and non-convex smooth functions. Our study incorporates time-dependent bias and emphasizes the importance of controlling the bias and Mean …

analysis applications approximation case cs.lg deep learning gradient networks neural networks paper reinforcement reinforcement learning stat.ml stochastic training unbiased

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Machine Learning Engineer - Sr. Consultant level

@ Visa | Bellevue, WA, United States