Oct. 14, 2022, 1:14 a.m. | Cedric Gerbelot, Emanuele Troiani, Francesca Mignacco, Florent Krzakala, Lenka Zdeborova

stat.ML updates on arXiv.org arxiv.org

We prove closed-form equations for the exact high-dimensional asymptotics of
a family of first order gradient-based methods, learning an estimator (e.g.
M-estimator, shallow neural network, ...) from observations on Gaussian data
with empirical risk minimization. This includes widely used algorithms such as
stochastic gradient descent (SGD) or Nesterov acceleration. The obtained
equations match those resulting from the discretization of dynamical mean-field
theory (DMFT) equations from statistical physics when applied to gradient flow.
Our proof method allows us to give an …

arxiv gradient math mean stochastic theory

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Data Analyst

@ SEAKR Engineering | Englewood, CO, United States

Data Analyst II

@ Postman | Bengaluru, India

Data Architect

@ FORSEVEN | Warwick, GB

Director, Data Science

@ Visa | Washington, DC, United States

Senior Manager, Data Science - Emerging ML

@ Capital One | McLean, VA