Feb. 15, 2024, 5:46 a.m. | Feifan Song, Yuxuan Fan, Xin Zhang, Peiyi Wang, Houfeng Wang

cs.CL updates on arXiv.org arxiv.org

arXiv:2402.09320v1 Announce Type: new
Abstract: Large Language Models (LLMs) rely on Human Preference Alignment (HPA) to ensure the generation of safe content. Due to the heavy cost associated with fine-tuning, fine-tuning-free methods have emerged, typically modifying LLM decoding with external auxiliary methods. However, these methods do not essentially enhance the LLM itself. In this paper, we rethink the derivation procedures of DPO, based on which we conversely build an instant scorer using the states of the LLM before and after …

abstract alignment arxiv capability context cost cs.ai cs.cl decoding direct preference optimization fine-tuning free human language language models large language large language models llm llms optimization type via

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US