Feb. 14, 2024, 9:35 p.m. | /u/threevox

Machine Learning www.reddit.com

Let’s say I’m attempting to fine-tune a pretrained language model, and I’d like to alter its response format. Normally, I’d fine-tune on a bunch of examples of responses in the new format. But doing so would also change the model’s semantic behavior to more closely mimic the type of text present in the SFT examples. Is there a way to fine-tune on an example in the new format, then effectively *negatively* fine-tune on the same text in the finetuning example …

behavior change examples fine-tuning format language language model machinelearning negative normally pretrained language model prompting responses semantic type

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Sr. VBI Developer II

@ Atos | Texas, US, 75093

Wealth Management - Data Analytics Intern/Co-op Fall 2024

@ Scotiabank | Toronto, ON, CA