all AI news
Large Language Models Can Learn Temporal Reasoning
Feb. 21, 2024, 5:49 a.m. | Siheng Xiong, Ali Payani, Ramana Kompella, Faramarz Fekri
cs.CL updates on arXiv.org arxiv.org
Abstract: While large language models (LLMs) have demonstrated remarkable reasoning capabilities, they are not without their flaws and inaccuracies. Recent studies have introduced various methods to mitigate these limitations. Temporal reasoning (TR), in particular, presents a significant challenge for LLMs due to its reliance on diverse temporal expressions and intricate contextual details. In this paper, we propose TG-LLM, a new framework towards language-based TR. To be specific, we first teach LLM to translate the context into …
abstract arxiv capabilities challenge cs.cl diverse flaws language language models large language large language models learn limitations llms reasoning reliance studies temporal type
More from arxiv.org / cs.CL updates on arXiv.org
Benchmarking LLMs via Uncertainty Quantification
1 day, 14 hours ago |
arxiv.org
CARE: Extracting Experimental Findings From Clinical Literature
1 day, 14 hours ago |
arxiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Global Data Architect, AVP - State Street Global Advisors
@ State Street | Boston, Massachusetts
Data Engineer
@ NTT DATA | Pune, MH, IN