Nov. 17, 2022, 2:12 a.m. | Yanqiu Wu, Xinyue Chen, Che Wang, Yiming Zhang, Keith W. Ross

cs.LG updates on arXiv.org arxiv.org

Recent advances in model-free deep reinforcement learning (DRL) show that
simple model-free methods can be highly effective in challenging
high-dimensional continuous control tasks. In particular, Truncated Quantile
Critics (TQC) achieves state-of-the-art asymptotic training performance on the
MuJoCo benchmark with a distributional representation of critics; and
Randomized Ensemble Double Q-Learning (REDQ) achieves high sample efficiency
that is competitive with state-of-the-art model-based methods using a high
update-to-data ratio and target randomization. In this paper, we propose a
novel model-free algorithm, Aggressive Q-Learning …

arxiv efficiency performance q-learning

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Senior Machine Learning Engineer

@ Samsara | Canada - Remote