July 17, 2023, 2:07 p.m. | /u/krumb0y

Natural Language Processing www.reddit.com

Hello,

I've been curious as to how far we can take small(7B and less) models for production use cases with small amounts of training data for each task.

So far I've been able to fine-tune LoRAs for paraphrasing, changing the tone of a sentence, dialogue summarization and topic generation. The results look promising, especially the fact that all this can run on very modest hardware.

Finetuning was done in 4bit mode using bitsandbytes.

I've used a AMD Ryzen9 3900XT + …

cases data dialogue finetuning languagetechnology paraphrasing production small summarization training training data use cases

Software Engineer for AI Training Data (School Specific)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Python)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Tier 2)

@ G2i Inc | Remote

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US