Oct. 19, 2022, 1:12 a.m. | Jingwei Zhang, Xunpeng Huang

cs.LG updates on arXiv.org arxiv.org

We consider optimizing two-layer neural networks in the mean-field regime
where the learning dynamics of network weights can be approximated by the
evolution in the space of probability measures over the weight parameters
associated with the neurons. The mean-field regime is a theoretically
attractive alternative to the NTK (lazy training) regime which is only
restricted locally in the so-called neural tangent kernel space around
specialized initializations. Several prior works (\cite{chizat2018global,
mei2018mean}) establish the asymptotic global optimality of the mean-field
regime, …

analysis arxiv convergence global linear mean networks neural networks

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Data Analyst (Commercial Excellence)

@ Allegro | Poznan, Warsaw, Poland

Senior Machine Learning Engineer

@ Motive | Pakistan - Remote

Summernaut Customer Facing Data Engineer

@ Celonis | Raleigh, US, North Carolina

Data Engineer Mumbai

@ Nielsen | Mumbai, India