Oct. 19, 2023, 6:13 a.m. | /u/Former_Goose_6894

Deep Learning www.reddit.com

I want to fine tune, models like llama-70b, code Alpaca or Starcoder for my use case. I actually have a dataset of 10,000 dockerfiles and want to finetune a model on this dataset.

What can be the expected time needed for this kind of training and what GPU should I use? Also, will the requirements differ for just using the model than training?

Ps: I am very new to the world of deep learning, LLMs and GPUs. Any help would …

alpaca case code dataset deeplearning finetuning gpu kind llama llm memory starcoder training understanding

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne