April 10, 2024, 4:47 a.m. | Victoria Basmov, Yoav Goldberg, Reut Tsarfaty

cs.CL updates on arXiv.org arxiv.org

arXiv:2404.06283v1 Announce Type: new
Abstract: The task of reading comprehension (RC), often implemented as context-based question answering (QA), provides a primary means to assess language models' natural language understanding (NLU) capabilities. Yet, when applied to large language models (LLMs) with extensive built-in world knowledge, this method can be deceptive. If the context aligns with the LLMs' internal knowledge, it is hard to discern whether the models' answers stem from context comprehension or from LLMs' internal information. Conversely, using data that …

abstract arxiv capabilities context cs.cl knowledge language language models language understanding large language large language models llms natural natural language nlu parametric question question answering reading type understanding world

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne