April 4, 2024, 4:41 a.m. | Guglielmo Bonifazi, Iason Chalas, Gian Hess, Jakub {\L}ucki

cs.LG updates on arXiv.org arxiv.org

arXiv:2404.02719v1 Announce Type: new
Abstract: This paper explores the connection between two recently identified phenomena in deep learning: plasticity loss and neural collapse. We analyze their correlation in different scenarios, revealing a significant association during the initial training phase on the first task. Additionally, we introduce a regularization approach to mitigate neural collapse, demonstrating its effectiveness in alleviating plasticity loss in this specific setting.

abstract analyze arxiv association correlation cs.ai cs.lg deep learning loss neural collapse paper regularization through training type

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Business Data Scientist, gTech Ads

@ Google | Mexico City, CDMX, Mexico

Lead, Data Analytics Operations

@ Zocdoc | Pune, Maharashtra, India