all AI news
[D] Removed 50% of the weights from a top leaderboard LLM without negatively impacting the evals
Dec. 21, 2023, 1:37 p.m. | /u/mwitiderrick
Machine Learning www.reddit.com
Using SparseML I was able to zero out 50% of the
SOLAR-10.7B-Instruct-v1.0 weights.
I then quantized the remaining weights to INT8.
The results are amazing!
https://preview.redd.it/uefy5u1hin7c1.png?width=927&format=png&auto=webp&s=35f9c3a07ab3e7f3a0e22a7528adeafc71c4e8e5
Even after pruning and quantizing the model to 50% I still got stellar zero-shot evaluation results.
Try the model:
https://preview.redd.it/r5tmixshin7c1.png?width=1999&format=png&auto=webp&s=61370090bb0083fecde7b00310bda71527e2eb61
Interestingly, the model is pruned and quantized in one shot. This means that no retraining …
More from www.reddit.com / Machine Learning
Jobs in AI, ML, Big Data
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne