all AI news
[R] Why can Llama-3 work with 32K context if it only had 8K context length?
May 6, 2024, 5:43 a.m. | /u/sunchipsster
Machine Learning www.reddit.com
I didn't understand what he meant by "with zero-training (actually just a simple 2 line config) you can get 32k context out of llama-3 models"
Does someone know what this **dynamic scaling trick** is? Much appreciated! :)
32k context config context dynamic line llama machinelearning scaling simple training trick
More from www.reddit.com / Machine Learning
Jobs in AI, ML, Big Data
Software Engineer for AI Training Data (School Specific)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Python)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Tier 2)
@ G2i Inc | Remote
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US