May 27, 2022, 1:11 a.m. | Niladri S. Chatterji, Philip M. Long, Peter L. Bartlett

stat.ML updates on arXiv.org arxiv.org

The recent success of neural network models has shone light on a rather
surprising statistical phenomenon: statistical models that perfectly fit noisy
data can generalize well to unseen test data. Understanding this phenomenon of
$\textit{benign overfitting}$ has attracted intense theoretical and empirical
study. In this paper, we consider interpolating two-layer linear neural
networks trained with gradient flow on the squared loss and derive bounds on
the excess risk when the covariates satisfy sub-Gaussianity and
anti-concentration properties, and the noise is …

arxiv bias linear ml networks overfitting

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Healthcare Data Modeler/Data Architect - REMOTE

@ Perficient | United States

Data Analyst – Sustainability, Green IT

@ H&M Group | Stockholm, Sweden

RWE Data Analyst

@ Sanofi | Hyderabad

Machine Learning Engineer

@ JPMorgan Chase & Co. | Jersey City, NJ, United States