all AI news
Double Descent and Overfitting under Noisy Inputs and Distribution Shift for Linear Denoisers
March 18, 2024, 4:42 a.m. | Chinmaya Kausik, Kashvi Srivastava, Rishi Sonthalia
cs.LG updates on arXiv.org arxiv.org
Abstract: Despite the importance of denoising in modern machine learning and ample empirical work on supervised denoising, its theoretical understanding is still relatively scarce. One concern about studying supervised denoising is that one might not always have noiseless training data from the test distribution. It is more reasonable to have access to noiseless training data from a different dataset than the test dataset. Motivated by this, we study supervised denoising and noisy-input regression under distribution shift. …
abstract arxiv cs.lg data denoising distribution importance inputs linear machine machine learning math.st modern overfitting shift stat.ml stat.th studying test training training data type understanding work
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Senior Machine Learning Engineer
@ Samsara | Canada - Remote