Oct. 13, 2022, 1:13 a.m. | Nur Muhammad Mahi Shafiullah, Zichen Jeff Cui, Ariuntuya Altanzaya, Lerrel Pinto

cs.LG updates on arXiv.org arxiv.org

While behavior learning has made impressive progress in recent times, it lags
behind computer vision and natural language processing due to its inability to
leverage large, human-generated datasets. Human behaviors have wide variance,
multiple modes, and human demonstrations typically do not come with reward
labels. These properties limit the applicability of current methods in Offline
RL and Behavioral Cloning to learn from large, pre-collected datasets. In this
work, we present Behavior Transformer (BeT), a new technique to model unlabeled
demonstration …

arxiv behavior cloning transformers

Data Scientist (m/f/x/d)

@ Symanto Research GmbH & Co. KG | Spain, Germany

Enterprise Data Quality, Senior Analyst

@ Toyota North America | Plano

Data Analyst & Audit Management Software (AMS) Coordinator

@ World Vision | Philippines - Home Working

Product Manager Power BI Platform Tech I&E Operational Insights

@ ING | HBP (Amsterdam - Haarlerbergpark)

Sr. Director, Software Engineering, Clinical Data Strategy

@ Moderna | USA-Washington-Seattle-1099 Stewart Street

Data Engineer (Data as a Service)

@ Xplor | Atlanta, GA, United States