Nov. 22, 2022, 2:14 a.m. | Marina Speranskaya, Sameh Methias, Benjamin Roth

cs.CL updates on arXiv.org arxiv.org

We propose to use reinforcement learning to inform transformer-based
contextualized link prediction models by providing paths that are most useful
for predicting the correct answer. This is in contrast to previous approaches,
that either used reinforcement learning (RL) to directly search for the answer,
or based their prediction on limited or randomly selected context. Our
experiments on WN18RR and FB15k-237 show that contextualized link prediction
models consistently outperform RL-based answer search, and that additional
improvements (of up to 13.5\% MRR) …

arxiv link prediction prediction reinforcement reinforcement learning

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote