April 9, 2024, 1:56 p.m. | Maxim Saplin

DEV Community dev.to

I have been playing with Supervised Fine Tuning and LORA using my laptop with NVIDIA RTX 4060 8GB. The subject of SFT is vast, picking the correct training hyperparams is more magic than science, and there's a good deal of experimentation...


Yet let me share one small finding. GPU utilization and Shared Memory effect on training speed.


I used Stable LM 2 1.6B base model and turned it into a chat model using 4400 samples from OASTT2 dataset. Here is …

ai deal experimentation fine-tuning genai good gpu laptop llm lora machinelearning magic memory nvidia nvidia rtx performance playing programming rtx rtx 4060 science sft training vast

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Senior Data Scientist

@ ITE Management | New York City, United States