all AI news
Scaling and renormalization in high-dimensional regression
May 2, 2024, 4:42 a.m. | Alexander B. Atanasov, Jacob A. Zavatone-Veth, Cengiz Pehlevan
cs.LG updates on arXiv.org arxiv.org
Abstract: This paper presents a succinct derivation of the training and generalization performance of a variety of high-dimensional ridge regression models using the basic tools of random matrix theory and free probability. We provide an introduction and review of recent results on these topics, aimed at readers with backgrounds in physics and deep learning. Analytic formulas for the training and generalization errors are obtained in a few lines of algebra directly from the properties of the …
abstract arxiv basic cond-mat.dis-nn cs.lg derivation free introduction matrix paper performance probability random readers regression results review ridge scaling stat.ml theory tools topics training type
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Software Engineer for AI Training Data (School Specific)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Python)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Tier 2)
@ G2i Inc | Remote
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US