all AI news
Asymptotics of Learning with Deep Structured (Random) Features
Feb. 22, 2024, 5:42 a.m. | Dominik Schr\"oder, Daniil Dmitriev, Hugo Cui, Bruno Loureiro
cs.LG updates on arXiv.org arxiv.org
Abstract: For a large class of feature maps we provide a tight asymptotic characterisation of the test error associated with learning the readout layer, in the high-dimensional limit where the input dimension, hidden layer widths, and number of training samples are proportionally large. This characterization is formulated in terms of the population covariance of the features. Our work is partially motivated by the problem of learning with Gaussian rainbow neural networks, namely deep non-linear fully-connected networks …
abstract arxiv class cond-mat.dis-nn cs.lg error feature features hidden layer maps math.st random samples stat.ml stat.th test training type
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US
Research Engineer
@ Allora Labs | Remote
Ecosystem Manager
@ Allora Labs | Remote
Founding AI Engineer, Agents
@ Occam AI | New York