Sept. 30, 2022, 1:12 a.m. | Lu Xia, Stefano Massei, Michiel E. Hochstenbach, Barry Koren

cs.LG updates on arXiv.org arxiv.org

When implementing the gradient descent method in low precision, the
employment of stochastic rounding schemes helps to prevent stagnation of
convergence caused by the vanishing gradient effect. Unbiased stochastic
rounding yields zero bias by preserving small updates with probabilities
proportional to their relative magnitudes. This study provides a theoretical
explanation for the stagnation of the gradient descent method in low-precision
computation. Additionally, we propose two new stochastic rounding schemes that
trade the zero bias property with a larger probability to …

arxiv computation convergence errors gradient influence low precision stochastic

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Business Intelligence Analyst

@ Rappi | COL-Bogotá

Applied Scientist II

@ Microsoft | Redmond, Washington, United States