all AI news
Offline Diversity Maximization Under Imitation Constraints
June 24, 2024, 4:46 a.m. | Marin Vlastelica, Jin Cheng, Georg Martius, Pavel Kolev
cs.LG updates on arXiv.org arxiv.org
Abstract: There has been significant recent progress in the area of unsupervised skill discovery, utilizing various information-theoretic objectives as measures of diversity. Despite these advances, challenges remain: current methods require significant online interaction, fail to leverage vast amounts of available task-agnostic data and typically lack a quantitative measure of skill utility. We address these challenges by proposing a principled offline algorithm for unsupervised skill discovery that, in addition to maximizing diversity, ensures that each learned skill …
abstract advances arxiv challenges constraints cs.ai cs.lg cs.ro current data discovery diversity fail information offline progress quantitative replace skill type unsupervised vast
More from arxiv.org / cs.LG updates on arXiv.org
MixerFlow: MLP-Mixer meets Normalising Flows
2 days, 11 hours ago |
arxiv.org
Machine Learning-Enabled Software and System Architecture Frameworks
2 days, 11 hours ago |
arxiv.org
Kernelised Normalising Flows
2 days, 11 hours ago |
arxiv.org
Jobs in AI, ML, Big Data
Data Scientist
@ Ford Motor Company | Chennai, Tamil Nadu, India
Systems Software Engineer, Graphics
@ Parallelz | Vancouver, British Columbia, Canada - Remote
Engineering Manager - Geo Engineering Team (F/H/X)
@ AVIV Group | Paris, France
Data Analyst
@ Microsoft | San Antonio, Texas, United States
Azure Data Engineer
@ TechVedika | Hyderabad, India
Senior Data & AI Threat Detection Researcher (Cortex)
@ Palo Alto Networks | Tel Aviv-Yafo, Israel