May 16, 2022, 1:11 a.m. | Dominic Petrak, Nafise Sadat Moosavi, Iryna Gurevych

cs.CL updates on arXiv.org arxiv.org

State-of-the-art pretrained language models tend to perform below their
capabilities when applied out-of-the-box on tasks that require reasoning over
numbers. Recent work sees two main reasons for this: (1) popular tokenisation
algorithms are optimized for common words, and therefore have limited
expressiveness for numbers, and (2) common pretraining objectives do not target
numerical reasoning or understanding numbers at all. Recent approaches usually
address them separately and mostly by proposing architectural changes or
pretraining models from scratch. In this paper, we …

arxiv language language models numerical reasoning skills

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Healthcare Data Modeler/Data Architect - REMOTE

@ Perficient | United States

Data Analyst – Sustainability, Green IT

@ H&M Group | Stockholm, Sweden

RWE Data Analyst

@ Sanofi | Hyderabad

Machine Learning Engineer

@ JPMorgan Chase & Co. | Jersey City, NJ, United States