all AI news
NEFTune: NEW LLM Fine-Tuning plus 25% Performance
Oct. 19, 2023, noon | code_your_own_AI
code_your_own_AI www.youtube.com
A new Instruction Fine-Tuning method increases LLM performance by up to 25%, with one line of code in HuggingFace TRL Transformer Library.
NEFTune explained in theory and on a practical example.
Arxiv pre-print available at (all rights with authors):
https://arxiv.org/pdf/2310.05914.pdf
HuggingFace TRL code implementation of NEFTune:
https://huggingface.co/docs/trl/main/en/sft_trainer#enhance-models-performances-using-neftune
#finetuning
#ai
#aieducation
arxiv authors code embeddings example explained fine-tuning finetuning huggingface implementation library line llm llm performance performance practical rights theory transformer
More from www.youtube.com / code_your_own_AI
Understand DSPy: Programming AI Pipelines
1 day, 10 hours ago |
www.youtube.com
Latest Insights in AI Performance Models
3 days, 10 hours ago |
www.youtube.com
New Discovery: Retrieval Heads for Long Context
5 days, 10 hours ago |
www.youtube.com
Multi-Token Prediction (forget next token LLM?)
6 days, 10 hours ago |
www.youtube.com
LLMs: Rewriting Our Tomorrow (plus code) #ai
1 week, 1 day ago |
www.youtube.com
Autonomous AI Agents: 14 % MAX Performance
1 week, 3 days ago |
www.youtube.com
480B LLM as 128x4B MoE? WHY?
1 week, 5 days ago |
www.youtube.com
No more Fine-Tuning: Unsupervised ICL+
1 week, 6 days ago |
www.youtube.com
Jobs in AI, ML, Big Data
Lead Developer (AI)
@ Cere Network | San Francisco, US
Research Engineer
@ Allora Labs | Remote
Ecosystem Manager
@ Allora Labs | Remote
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US
AI Research Scientist
@ Vara | Berlin, Germany and Remote