April 19, 2024, 4:41 a.m. | Masaki Adachi, Satoshi Hayakawa, Martin J{\o}rgensen, Saad Hamid, Harald Oberhauser, Michael A. Osborne

cs.LG updates on arXiv.org arxiv.org

arXiv:2404.12219v1 Announce Type: new
Abstract: Parallelisation in Bayesian optimisation is a common strategy but faces several challenges: the need for flexibility in acquisition functions and kernel choices, flexibility dealing with discrete and continuous variables simultaneously, model misspecification, and lastly fast massive parallelisation. To address these challenges, we introduce a versatile and modular framework for batch Bayesian optimisation via probabilistic lifting with kernel quadrature, called SOBER, which we present as a Python library based on GPyTorch/BoTorch. Our framework offers the following …

abstract acquisition arxiv bayesian challenges continuous cs.lg cs.na flexibility functions general kernel massive math.na optimisation optimization stat.ml strategy type variables via

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York