June 11, 2023, 12:15 p.m. | code_your_own_AI

code_your_own_AI www.youtube.com

QLoRA 4bit Quantization for memory efficient fine-tuning of LLMs explained in detailed. 4-bit quantization QLoRA for beginners, theory and code. PEFT - parameter efficient fine-tuning methods.

Based on my first videos on the theory of LoRA and other PEFT methods (https://youtu.be/YVU5wAA6Txo) and the detailed code implementation of LoRA in my video (https://youtu.be/A-a-l_sFtYM) now my third video on 4-bit quantization and QLoRA.

An additional Colab NB with code to fine-tune FALCON 7B with QLoRA 4-bit quantization and Transformer Reinforcement Learning (TLR). …

beginners code colab explained falcon fine-tuning huggingface llm llm models llms memory quantization reinforcement reinforcement learning theory transformer understanding

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne