all AI news
[N] OpenLLaMA: An Open Reproduction of LLaMA
May 3, 2023, 8:51 a.m. | /u/Philpax
Machine Learning www.reddit.com
> We train our models on the RedPajama dataset released by Together, which is a reproduction of the LLaMA training dataset containing over 1.2 trillion tokens. We follow the exactly same preprocessing steps and training hyperparameters as the original LLaMA paper, including model architecture, context length, training steps, learning rate schedule, and optimizer. The only difference between our setting and the original one is the dataset used: OpenLLaMA employs the RedPajama dataset rather than the one utilized by the …
architecture context dataset difference llama machinelearning paper rate redpajama together tokens training
More from www.reddit.com / Machine Learning
[D] ICLR Outstanding Paper Awards. Congratulations!
1 day, 1 hour ago |
www.reddit.com
[D] Where does the term "feature" come from?
1 day, 3 hours ago |
www.reddit.com
Jobs in AI, ML, Big Data
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US
Research Engineer
@ Allora Labs | Remote
Ecosystem Manager
@ Allora Labs | Remote
Founding AI Engineer, Agents
@ Occam AI | New York