all AI news
Markovian Flow Matching: Accelerating MCMC with Continuous Normalizing Flows
May 24, 2024, 4:46 a.m. | Alberto Cabezas, Louis Sharrock, Christopher Nemeth
cs.LG updates on arXiv.org arxiv.org
Abstract: Continuous normalizing flows (CNFs) learn the probability path between a reference and a target density by modeling the vector field generating said path using neural networks. Recently, Lipman et al. (2022) introduced a simple and inexpensive method for training CNFs in generative modeling, termed flow matching (FM). In this paper, we re-purpose this method for probabilistic inference by incorporating Markovian sampling methods in evaluating the FM objective and using the learned probability path to improve …
abstract arxiv continuous continuous normalizing flows cs.lg flow generative generative modeling learn mcmc modeling networks neural networks path probability reference said simple stat.me stat.ml training type vector
More from arxiv.org / cs.LG updates on arXiv.org
Machine-learned models for magnetic materials
2 days, 10 hours ago |
arxiv.org
Revisiting RIP guarantees for sketching operators on mixture models
2 days, 10 hours ago |
arxiv.org
Jobs in AI, ML, Big Data
Senior Data Engineer
@ Displate | Warsaw
Junior Data Analyst - ESG Data
@ Institutional Shareholder Services | Mumbai
Intern Data Driven Development in Sensor Fusion for Autonomous Driving (f/m/x)
@ BMW Group | Munich, DE
Senior MLOps Engineer, Machine Learning Platform
@ GetYourGuide | Berlin
Data Engineer, Analytics
@ Meta | Menlo Park, CA
Data Engineer
@ Meta | Menlo Park, CA