all AI news
Minimum discrepancy principle strategy for choosing $k$ in $k$-NN regression
May 8, 2024, 4:43 a.m. | Yaroslav Averyanov, Alain Celisse
cs.LG updates on arXiv.org arxiv.org
Abstract: We present a novel data-driven strategy to choose the hyperparameter $k$ in the $k$-NN regression estimator without using any hold-out data. We treat the problem of choosing the hyperparameter as an iterative procedure (over $k$) and propose using an easily implemented in practice strategy based on the idea of early stopping and the minimum discrepancy principle. This model selection strategy is proven to be minimax-optimal, under the fixed-design assumption on covariates, over some smoothness function …
abstract arxiv cs.lg data data-driven estimator hyperparameter iterative math.st minimum novel practice regression stat.ml stat.th strategy type
More from arxiv.org / cs.LG updates on arXiv.org
Efficient Data-Driven MPC for Demand Response of Commercial Buildings
2 days, 18 hours ago |
arxiv.org
Testing the Segment Anything Model on radiology data
2 days, 18 hours ago |
arxiv.org
Calorimeter shower superresolution
2 days, 18 hours ago |
arxiv.org
Jobs in AI, ML, Big Data
Software Engineer for AI Training Data (School Specific)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Python)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Tier 2)
@ G2i Inc | Remote
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US