Oct. 21, 2022, 1:13 a.m. | Byungchan Ko, Jungseul Ok

cs.LG updates on arXiv.org arxiv.org

In deep reinforcement learning (RL), data augmentation is widely considered
as a tool to induce a set of useful priors about semantic consistency and
improve sample efficiency and generalization performance. However, even when
the prior is useful for generalization, distilling it to RL agent often
interferes with RL training and degenerates sample efficiency. Meanwhile, the
agent is forgetful of the prior due to the non-stationary nature of RL. These
observations suggest two extreme schedules of distillation: (i) over the entire …

arxiv augmentation data reinforcement reinforcement learning scheduling

Data Scientist (m/f/x/d)

@ Symanto Research GmbH & Co. KG | Spain, Germany

Data Operations Analyst

@ Workday | Poland, Warsaw

Reference Data Specialist - Operations Analyst

@ JPMorgan Chase & Co. | Bengaluru, Karnataka, India

Data Scientist (Redwood City)

@ Anomali | Redwood City, CA

Software Engineer, Database - Languages & Relational Technologies

@ YugabyteDB | United States (Remote) or Sunnyvale, CA

Data Analyst (m/f/d) Online Marketing

@ StepStone Group | Düsseldorf, Germany