April 8, 2022, 1:11 a.m. | Yafei Hu, Chen Wang, John Keller, Sebastian Scherer

cs.LG updates on arXiv.org arxiv.org

In traditional robot exploration methods, the robot usually does not have
prior biases about the environment it is exploring. Thus the robot assigns
equal importance to the goals which leads to insufficient exploration
efficiency. Alternative, often a hand-tuned policy is used to tweak the value
of goals. In this paper, we present a method to learn how "good" some states
are, measured by the state value function, to provide a hint for the robot to
make exploration decisions. We propose …

arxiv exploration learning robot value

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Staff Software Engineer, Generative AI, Google Cloud AI

@ Google | Mountain View, CA, USA; Sunnyvale, CA, USA

Expert Data Sciences

@ Gainwell Technologies | Any city, CO, US, 99999