all AI news
[N] Huggingface/nvidia release open source GPT-2B trained on 1.1T tokens
May 1, 2023, 3:46 p.m. | /u/norcalnatv
Machine Learning www.reddit.com
## Model Description
GPT-2B-001 is a transformer-based language model. GPT refers to a class of transformer decoder-only models similar to GPT-2 and 3 while 2B refers to the total trainable parameter count (2 Billion) \[1, 2\].
This model was trained on 1.1T tokens with [NeMo](https://docs.nvidia.com/deeplearning/nemo/user-guide/docs/en/stable/nlp/nemo_megatron/intro.html).
Requires Ampere or Hopper devices.
ampere count decoder devices gpt gpt-2 hopper language language model machinelearning transformer
More from www.reddit.com / Machine Learning
Jobs in AI, ML, Big Data
Seeking Developers and Engineers for AI T-Shirt Generator Project
@ Chevon Hicks | Remote
Software Engineer for AI Training Data (School Specific)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Python)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Tier 2)
@ G2i Inc | Remote
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Stage - Product Owner Assistant - Data Platform / Business Intelligence (M/F)
@ Pernod Ricard | FR - Paris - The Island