March 12, 2024, 4:45 a.m. | Joseph Shenouda, Rahul Parhi, Kangwook Lee, Robert D. Nowak

cs.LG updates on arXiv.org arxiv.org

arXiv:2305.16534v2 Announce Type: replace-cross
Abstract: This paper introduces a novel theoretical framework for the analysis of vector-valued neural networks through the development of vector-valued variation spaces, a new class of reproducing kernel Banach spaces. These spaces emerge from studying the regularization effect of weight decay in training networks with activations like the rectified linear unit (ReLU). This framework offers a deeper understanding of multi-output networks and their function-space characteristics. A key contribution of this work is the development of a …

abstract analysis arxiv class compression cs.lg development framework insights kernel multi-task learning network networks neural networks novel paper regularization spaces stat.ml studying through training type variation vector

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Data Science Analyst

@ Mayo Clinic | AZ, United States

Sr. Data Scientist (Network Engineering)

@ SpaceX | Redmond, WA