all AI news
Understanding Catastrophic Forgetting in Language Models via Implicit Inference
April 16, 2024, 4:45 a.m. | Suhas Kotha, Jacob Mitchell Springer, Aditi Raghunathan
cs.LG updates on arXiv.org arxiv.org
Abstract: We lack a systematic understanding of the effects of fine-tuning (via methods such as instruction-tuning or reinforcement learning from human feedback), particularly on tasks outside the narrow fine-tuning distribution. In a simplified scenario, we demonstrate that improving performance on tasks within the fine-tuning data distribution comes at the expense of capabilities on other tasks. We hypothesize that language models implicitly infer the task of the prompt and that fine-tuning skews this inference towards tasks in …
abstract arxiv catastrophic forgetting cs.cl cs.lg data distribution effects feedback fine-tuning human human feedback improving inference language language models narrow performance reinforcement reinforcement learning simplified tasks type understanding via
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
C003549 Data Analyst (NS) - MON 13 May
@ EMW, Inc. | Braine-l'Alleud, Wallonia, Belgium
Marketing Decision Scientist
@ Meta | Menlo Park, CA | New York City