Nov. 23, 2022, 2:12 a.m. | Thomas George, Guillaume Lajoie, Aristide Baratin

cs.LG updates on arXiv.org arxiv.org

Among attempts at giving a theoretical account of the success of deep neural
networks, a recent line of work has identified a so-called lazy training regime
in which the network can be well approximated by its linearization around
initialization. Here we investigate the comparative effect of the lazy (linear)
and feature learning (non-linear) regimes on subgroups of examples based on
their difficulty. Specifically, we show that easier examples are given more
weight in feature learning mode, resulting in faster training …

arxiv example impacts lazy linearization networks

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Global Data Architect, AVP - State Street Global Advisors

@ State Street | Boston, Massachusetts

Data Engineer

@ NTT DATA | Pune, MH, IN