all AI news
Scaling Limit of Neural Networks with the Xavier Initialization and Convergence to a Global Minimum. (arXiv:1907.04108v3 [math.PR] UPDATED)
April 13, 2022, 1:12 a.m. | Justin Sirignano, Konstantinos Spiliopoulos
cs.LG updates on arXiv.org arxiv.org
We analyze single-layer neural networks with the Xavier initialization in the
asymptotic regime of large numbers of hidden units and large numbers of
stochastic gradient descent training steps. The evolution of the neural network
during training can be viewed as a stochastic system and, using techniques from
stochastic analysis, we prove the neural network converges in distribution to a
random ODE with a Gaussian distribution. The limit is completely different than
in the typical mean-field results for neural networks due …
arxiv convergence global math networks neural networks pr scaling
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Senior Engineer - Data Science Operations
@ causaLens | London - Hybrid, England, United Kingdom
F0138 - LLM Developer (AI NLP)
@ Ubiquiti Inc. | Taipei
Staff Engineer, Database
@ Nagarro | Gurugram, India
Artificial Intelligence Assurance Analyst
@ Booz Allen Hamilton | USA, VA, McLean (8251 Greensboro Dr)