all AI news
Rule or Story, Which is a Better Commonsense Expression for Talking with Large Language Models?
Feb. 23, 2024, 5:48 a.m. | Ning Bian, Xianpei Han, Hongyu Lin, Yaojie Lu, Ben He, Le Sun
cs.CL updates on arXiv.org arxiv.org
Abstract: Building machines with commonsense has been a longstanding challenge in NLP due to the reporting bias of commonsense rules and the exposure bias of rule-based commonsense reasoning. In contrast, humans convey and pass down commonsense implicitly through stories. This paper investigates the inherent commonsense ability of large language models (LLMs) expressed through storytelling. We systematically investigate and compare stories and rules for retrieving and leveraging commonsense in LLMs. Experimental results on 28 commonsense QA datasets …
abstract arxiv bias building challenge contrast cs.cl humans language language models large language large language models machines nlp paper reasoning reporting rules stories story through type
More from arxiv.org / cs.CL updates on arXiv.org
Benchmarking LLMs via Uncertainty Quantification
1 day, 18 hours ago |
arxiv.org
CARE: Extracting Experimental Findings From Clinical Literature
1 day, 18 hours ago |
arxiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Director, Clinical Data Science
@ Aura | Remote USA
Research Scientist, AI (PhD)
@ Meta | Menlo Park, CA | New York City