all AI news
Dropout Regularization Versus $\ell_2$-Penalization in the Linear Model
April 26, 2024, 4:44 a.m. | Gabriel Clara, Sophie Langer, Johannes Schmidt-Hieber
stat.ML updates on arXiv.org arxiv.org
Abstract: We investigate the statistical behavior of gradient descent iterates with dropout in the linear regression model. In particular, non-asymptotic bounds for the convergence of expectations and covariance matrices of the iterates are derived. The results shed more light on the widely cited connection between dropout and l2-regularization in the linear model. We indicate a more subtle relationship, owing to interactions between the gradient descent dynamics and the additional randomness induced by dropout. Further, we study …
abstract arxiv behavior convergence covariance dropout gradient light linear linear model linear regression math.st regression regularization results statistical stat.ml stat.th type
More from arxiv.org / stat.ML updates on arXiv.org
Learning linear dynamical systems under convex constraints
3 days, 15 hours ago |
arxiv.org
Inverse Unscented Kalman Filter
4 days, 15 hours ago |
arxiv.org
Jobs in AI, ML, Big Data
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Data Scientist (Database Development)
@ Nasdaq | Bengaluru-Affluence