all AI news
Tailoring Self-Rationalizers with Multi-Reward Distillation
May 24, 2024, 4:55 a.m. | Sahana Ramnath, Brihi Joshi, Skyler Hallinan, Ximing Lu, Liunian Harold Li, Aaron Chan, Jack Hessel, Yejin Choi, Xiang Ren
cs.CL updates on arXiv.org arxiv.org
Abstract: Large language models (LMs) are capable of generating free-text rationales to aid question answering. However, prior work 1) suggests that useful self-rationalization is emergent only at significant scales (e.g., 175B parameter GPT-3); and 2) focuses largely on downstream performance, ignoring the semantics of the rationales themselves, e.g., are they faithful, true, and helpful for humans? In this work, we enable small-scale LMs (approx. 200x smaller than GPT-3) to generate rationales that not only improve downstream …
abstract arxiv cs.cl distillation free gpt gpt-3 however language language models large language large language models lms performance prior question question answering replace semantics text type work
More from arxiv.org / cs.CL updates on arXiv.org
Dodo: Dynamic Contextual Compression for Decoder-only LMs
2 days, 14 hours ago |
arxiv.org
Active Learning for Multilingual Fingerspelling Corpora
2 days, 14 hours ago |
arxiv.org
Jobs in AI, ML, Big Data
Senior Data Engineer
@ Displate | Warsaw
Junior Data Analyst - ESG Data
@ Institutional Shareholder Services | Mumbai
Intern Data Driven Development in Sensor Fusion for Autonomous Driving (f/m/x)
@ BMW Group | Munich, DE
Senior MLOps Engineer, Machine Learning Platform
@ GetYourGuide | Berlin
Data Engineer, Analytics
@ Meta | Menlo Park, CA
Data Engineer
@ Meta | Menlo Park, CA