July 17, 2023, 1:13 p.m. | /u/krumb0y

Machine Learning www.reddit.com

Hello,

I've been curious as to how far we can take small(7B and less) models for production use cases with small amounts of training data for each task.

So far I've been able to fine-tune LoRAs for paraphrasing, changing the tone of a sentence, dialogue summarization and topic generation. The results look promising, especially the fact that all this can run on very modest hardware.

Finetuning was done in 4bit mode using bitsandbytes. Each task had \~1k training points.

I've …

cases data dialogue finetuning machinelearning paraphrasing production small summarization training training data use cases

Software Engineer for AI Training Data (School Specific)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Python)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Tier 2)

@ G2i Inc | Remote

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US