April 11, 2022, 1:11 a.m. | Joosung Min, Lloyd T. Elliott

cs.LG updates on arXiv.org arxiv.org

$Q$-learning is the most fundamental model-free reinforcement learning
algorithm. Deployment of $Q$-learning requires approximation of the
state-action value function (also known as the $Q$-function). In this work, we
provide online random forests as $Q$-function approximators and propose a novel
method wherein the random forest is grown as learning proceeds (through
expanding forests). We demonstrate improved performance of our methods over
state-of-the-art Deep $Q$-Networks in two OpenAI gyms (`blackjack' and
`inverted pendulum') but not in the `lunar lander' gym. We suspect …

arxiv learning ml q-learning random random forests

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Data Analyst - Associate

@ JPMorgan Chase & Co. | Mumbai, Maharashtra, India

Staff Data Engineer (Data Platform)

@ Coupang | Seoul, South Korea

AI/ML Engineering Research Internship

@ Keysight Technologies | Santa Rosa, CA, United States

Sr. Director, Head of Data Management and Reporting Execution

@ Biogen | Cambridge, MA, United States

Manager, Marketing - Audience Intelligence (Senior Data Analyst)

@ Delivery Hero | Singapore, Singapore