April 24, 2024, 4:47 a.m. | Amir Saeidi, Shivanshu Verma, Chitta Baral

cs.CL updates on arXiv.org arxiv.org

arXiv:2404.14723v1 Announce Type: new
Abstract: Large Language Models (LLMs) have demonstrated remarkable performance across a spectrum of tasks. Recently, Direct Preference Optimization (DPO) has emerged as an RL-free approach to optimize the policy model on human preferences. However, several limitations hinder the widespread adoption of this method. To address these shortcomings, various versions of DPO have been introduced. Yet, a comprehensive evaluation of these variants across diverse tasks is still lacking. In this study, we aim to bridge this gap …

abstract adoption alignment arxiv cs.cl direct preference optimization dpo free hinder however human insights language language models large language large language models limitations llms multiple optimization performance policy spectrum tasks type variants

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne