all AI news
[N] Huggingface/nvidia release open source GPT-2B trained on 1.1T tokens
May 1, 2023, 3:46 p.m. | /u/norcalnatv
Machine Learning www.reddit.com
## Model Description
GPT-2B-001 is a transformer-based language model. GPT refers to a class of transformer decoder-only models similar to GPT-2 and 3 while 2B refers to the total trainable parameter count (2 Billion) \[1, 2\].
This model was trained on 1.1T tokens with [NeMo](https://docs.nvidia.com/deeplearning/nemo/user-guide/docs/en/stable/nlp/nemo_megatron/intro.html).
Requires Ampere or Hopper devices.
ampere count decoder devices gpt gpt-2 hopper language language model machinelearning transformer
More from www.reddit.com / Machine Learning
[D] software to design figures
15 hours ago |
www.reddit.com
[Discussion] Should I go to ICML and present my paper?
1 day, 9 hours ago |
www.reddit.com
Jobs in AI, ML, Big Data
AI Engineer Intern, Agents
@ Occam AI | US
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Data Engineer - Takealot Group (Takealot.com | Superbalist.com | Mr D Food)
@ takealot.com | Cape Town