Feb. 22, 2024, 6:42 a.m. | Shaw Talebi

Towards Data Science - Medium towardsdatascience.com

QLoRA — How to Fine-Tune an LLM on a Single GPU

An introduction with Python example code (ft. Mistral-7b)

This article is part of a larger series on using large language models (LLMs) in practice. In the previous post, we saw how to fine-tune an LLM using OpenAI. The main limitation to this approach, however, is that OpenAI’s models are concealed behind their API, which limits what and how we can build with them. Here, I’ll discuss an alternative …

article code example fine-tuning gpu hands-on-tutorials introduction language language models large language large language models llm llms machine learning mistral openai part practice python qlora series

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Software Engineer, Machine Learning (Tel Aviv)

@ Meta | Tel Aviv, Israel

Senior Data Scientist- Digital Government

@ Oracle | CASABLANCA, Morocco