all AI news
Tailoring Self-Rationalizers with Multi-Reward Distillation
May 24, 2024, 4:55 a.m. | Sahana Ramnath, Brihi Joshi, Skyler Hallinan, Ximing Lu, Liunian Harold Li, Aaron Chan, Jack Hessel, Yejin Choi, Xiang Ren
cs.CL updates on arXiv.org arxiv.org
Abstract: Large language models (LMs) are capable of generating free-text rationales to aid question answering. However, prior work 1) suggests that useful self-rationalization is emergent only at significant scales (e.g., 175B parameter GPT-3); and 2) focuses largely on downstream performance, ignoring the semantics of the rationales themselves, e.g., are they faithful, true, and helpful for humans? In this work, we enable small-scale LMs (approx. 200x smaller than GPT-3) to generate rationales that not only improve downstream …
abstract arxiv cs.cl distillation free gpt gpt-3 however language language models large language large language models lms performance prior question question answering replace semantics text type work
More from arxiv.org / cs.CL updates on arXiv.org
Jobs in AI, ML, Big Data
AI Focused Biochemistry Postdoctoral Fellow
@ Lawrence Berkeley National Lab | Berkeley, CA
Senior Data Engineer
@ Displate | Warsaw
Associate Director, IT Business Partner, Cell Therapy Analytical Development
@ Bristol Myers Squibb | Warren - NJ
Solutions Architect
@ Lloyds Banking Group | London 125 London Wall
Senior Lead Cloud Engineer
@ S&P Global | IN - HYDERABAD ORION
Software Engineer
@ Applied Materials | Bengaluru,IND