May 15, 2023, 1:13 p.m. | /u/igorsusmelj

Machine Learning www.reddit.com

I was wondering if anyone has looked into data sampling or active learning techniques to fine-tune LLMs. Using PEFT methods like LoRA we can use much fewer samples for fine-tuning. But the training data still requires some sort of labels or responses for questions. I found these two datasets that seem commonly used (Alpaca and OASST1). Both seem rather small.

[Alpaca](https://huggingface.co/datasets/tatsu-lab/alpaca) has 52k instructions.
[OpenAssistant Conversations Dataset (OASST1)](https://huggingface.co/datasets/OpenAssistant/oasst1) has 160k messages that result in "in over 10,000 fully annotated conversation …

active learning data fine-tuning labels llm llms lora machinelearning questions responses sampling training training data

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Vice President, AI Product Manager

@ JPMorgan Chase & Co. | New York City, United States

Binance Accelerator Program - Data Engineer

@ Binance | Asia