Oct. 20, 2022, 1:12 a.m. | Byungchan Ko, Jungseul Ok

cs.LG updates on arXiv.org arxiv.org

In deep reinforcement learning (RL), data augmentation is widely considered
as a tool to induce a set of useful priors about semantic consistency and
improve sample efficiency and generalization performance. However, even when
the prior is useful for generalization, distilling it to RL agent often
interferes with RL training and degenerates sample efficiency. Meanwhile, the
agent is forgetful of the prior due to the non-stationary nature of RL. These
observations suggest two extreme schedules of distillation: (i) over the entire …

arxiv augmentation data reinforcement reinforcement learning scheduling

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Technology Consultant Master Data Management (w/m/d)

@ SAP | Walldorf, DE, 69190

Research Engineer, Computer Vision, Google Research

@ Google | Nairobi, Kenya