all AI news
Stereographic Markov Chain Monte Carlo
Feb. 22, 2024, 5:44 a.m. | Jun Yang, Krzysztof {\L}atuszy\'nski, Gareth O. Roberts
stat.ML updates on arXiv.org arxiv.org
Abstract: High-dimensional distributions, especially those with heavy tails, are notoriously difficult for off-the-shelf MCMC samplers: the combination of unbounded state spaces, diminishing gradient information, and local moves results in empirically observed ``stickiness'' and poor theoretical mixing properties -- lack of geometric ergodicity. In this paper, we introduce a new class of MCMC samplers that map the original high-dimensional problem in Euclidean space onto a sphere and remedy these notorious mixing problems. In particular, we develop random-walk …
abstract arxiv class combination gradient information markov mcmc paper spaces stat.co state stat.me stat.ml type
More from arxiv.org / stat.ML updates on arXiv.org
Mixture of partially linear experts
7 hours ago |
arxiv.org
Adaptive deep learning for nonlinear time series models
1 day, 7 hours ago |
arxiv.org
A Full Adagrad algorithm with O(Nd) operations
1 day, 7 hours ago |
arxiv.org
Minimax Regret Learning for Data with Heterogeneous Subgroups
1 day, 7 hours ago |
arxiv.org
Jobs in AI, ML, Big Data
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Machine Learning Engineer - Sr. Consultant level
@ Visa | Bellevue, WA, United States