April 13, 2022, 1:12 a.m. | Justin Sirignano, Konstantinos Spiliopoulos

cs.LG updates on arXiv.org arxiv.org

We analyze single-layer neural networks with the Xavier initialization in the
asymptotic regime of large numbers of hidden units and large numbers of
stochastic gradient descent training steps. The evolution of the neural network
during training can be viewed as a stochastic system and, using techniques from
stochastic analysis, we prove the neural network converges in distribution to a
random ODE with a Gaussian distribution. The limit is completely different than
in the typical mean-field results for neural networks due …

arxiv convergence global math networks neural networks pr scaling

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Senior Engineer - Data Science Operations

@ causaLens | London - Hybrid, England, United Kingdom

F0138 - LLM Developer (AI NLP)

@ Ubiquiti Inc. | Taipei

Staff Engineer, Database

@ Nagarro | Gurugram, India

Artificial Intelligence Assurance Analyst

@ Booz Allen Hamilton | USA, VA, McLean (8251 Greensboro Dr)