Feb. 14, 2024, 5:43 a.m. | Joachim Bona-PellissierIMT Fran \c{c}ois MalgouyresIMT Fran \c{c}ois BachocIMT

cs.LG updates on arXiv.org arxiv.org

It is well known that neural networks with many more parameters than training examples do not overfit. Implicit regularization phenomena, which are still not well understood, occur during optimization and 'good' networks are favored. Thus the number of parameters is not an adequate measure of complexity if we do not consider all possible networks but only the 'good' ones. To better understand which networks are favored during optimization, we study the geometry of the output set as parameters vary. When …

complexity cs.ai cs.lg cs.ne examples geometry good math.oc math.st networks neural networks optimization parameters regularization relu stat.th training

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

AI Engineering Manager

@ M47 Labs | Barcelona, Catalunya [Cataluña], Spain