all AI news
Deep Gaussian Covariance Network with Trajectory Sampling for Data-Efficient Policy Search
March 26, 2024, 4:41 a.m. | Can Bogoclu, Robert Vosshall, Kevin Cremanns, Dirk Roos
cs.LG updates on arXiv.org arxiv.org
Abstract: Probabilistic world models increase data efficiency of model-based reinforcement learning (MBRL) by guiding the policy with their epistemic uncertainty to improve exploration and acquire new samples. Moreover, the uncertainty-aware learning procedures in probabilistic approaches lead to robust policies that are less sensitive to noisy observations compared to uncertainty unaware solutions. We propose to combine trajectory sampling and deep Gaussian covariance network (DGCN) for a data-efficient solution to MBRL problems in an optimal control setting. We …
abstract arxiv covariance cs.lg data efficiency exploration network policies policy reinforcement reinforcement learning robust samples sampling search stat.ml trajectory type uncertainty world world models
More from arxiv.org / cs.LG updates on arXiv.org
Efficient Data-Driven MPC for Demand Response of Commercial Buildings
2 days, 21 hours ago |
arxiv.org
Testing the Segment Anything Model on radiology data
2 days, 21 hours ago |
arxiv.org
Calorimeter shower superresolution
2 days, 21 hours ago |
arxiv.org
Jobs in AI, ML, Big Data
Software Engineer for AI Training Data (School Specific)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Python)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Tier 2)
@ G2i Inc | Remote
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US