Feb. 12, 2024, 5:42 a.m. | Jascha Sohl-Dickstein

cs.LG updates on arXiv.org arxiv.org

Some fractals -- for instance those associated with the Mandelbrot and quadratic Julia sets -- are computed by iterating a function, and identifying the boundary between hyperparameters for which the resulting series diverges or remains bounded. Neural network training similarly involves iterating an update function (e.g. repeated steps of gradient descent), can result in convergent or divergent behavior, and can be extremely sensitive to small changes in hyperparameters. Motivated by these similarities, we experimentally examine the boundary between neural network …

cs.lg cs.ne fractal fractals function gradient instance julia network network training neural network nlin.cd series training update

Research Scholar (Technical Research)

@ Centre for the Governance of AI | Hybrid; Oxford, UK

HPC Engineer (x/f/m) - DACH

@ Meshcapade GmbH | Remote, Germany

Business Consultant-AI/ML

@ Bosch Group | Bengaluru, India

Senior Network Defense Analyst (AI/ML) - Hybrid

@ Noblis | Linthicum, MD, United States

Senior Data Analyst

@ Peloton | New York City

SC2024-003425 Data Scientist (NS) - WED 6 Mar

@ EMW, Inc. | Brussels, Brussels, Belgium