Feb. 7, 2024, 5:42 a.m. | Li Guo Keith Ross Zifan Zhao Andriopoulos George Shuyang Ling Yufeng Xu Zixuan Dong

cs.LG updates on arXiv.org arxiv.org

Label smoothing loss is a widely adopted technique to mitigate overfitting in deep neural networks. This paper studies label smoothing from the perspective of Neural Collapse (NC), a powerful empirical and theoretical framework which characterizes model behavior during the terminal phase of training. We first show empirically that models trained with label smoothing converge faster to neural collapse solutions and attain a stronger level of neural collapse. Additionally, we show that at the same level of NC1, models under label …

behavior cs.lg entropy framework loss model behavior networks neural collapse neural networks overfitting paper perspective show studies terminal training

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead Data Engineer

@ WorkMoney | New York City, United States - Remote