April 2, 2024, 7:51 p.m. | Baolong Bi, Shenghua Liu, Yiwei Wang, Lingrui Mei, Xueqi Cheng

cs.CL updates on arXiv.org arxiv.org

arXiv:2404.00216v1 Announce Type: new
Abstract: The rapid development of large language models (LLMs) enables them to convey factual knowledge in a more human-like fashion. Extensive efforts have been made to reduce factual hallucinations by modifying LLMs with factuality decoding. However, they also pose risks of hindering knowledge updates, as they make models overly confident in known facts. In this work, we first revisite the current factuality decoding methods and verified their effectiveness in enhancing factual accuracy. Subsequently, we conduct further …

abstract arxiv benchmark cs.ai cs.cl decoding development editing evaluation fashion free hallucinations however human human-like knowledge language language models large language large language models llms reduce risks them type

Senior Machine Learning Engineer

@ GPTZero | Toronto, Canada

ML/AI Engineer / NLP Expert - Custom LLM Development (x/f/m)

@ HelloBetter | Remote

Doctoral Researcher (m/f/div) in Automated Processing of Bioimages

@ Leibniz Institute for Natural Product Research and Infection Biology (Leibniz-HKI) | Jena

Seeking Developers and Engineers for AI T-Shirt Generator Project

@ Chevon Hicks | Remote

Cloud Data Platform Engineer

@ First Central | Home Office (Remote)

Associate Director, Data Science

@ MSD | USA - New Jersey - Rahway