all AI news
Next-Gen AI: RecurrentGemma (Long Context Length)
April 21, 2024, 2 p.m. | code_your_own_AI
code_your_own_AI www.youtube.com
Google developed RecurrentGemma-2B and compares this new LM architecture (!) with the classical transformer based, quadratic complexity of a self-attention Gemma 2B. And the new throughput is: about 6000 tokens per second.
Introduction and Model Architecture:
The original paper by Google introduces "RecurrentGemma-2B," leveraging the Griffin architecture, which moves away from traditional global attention mechanisms in favor of a combination of linear recurrences and local attention. This design enables the …
architecture attention brand complexity context gemma gen gen ai google introduction language language model moving next next-gen paper per self-attention tokens transformer transformers
More from www.youtube.com / code_your_own_AI
Multi-Token Prediction (forget next token LLM?)
1 day, 14 hours ago |
www.youtube.com
NEW LLM Test: Reasoning & gpt2-chatbot
2 days, 19 hours ago |
www.youtube.com
LLMs: Rewriting Our Tomorrow (plus code) #ai
4 days, 2 hours ago |
www.youtube.com
Autonomous AI Agents: 14 % MAX Performance
5 days, 14 hours ago |
www.youtube.com
No more Fine-Tuning: Unsupervised ICL+
1 week, 2 days ago |
www.youtube.com
NEW Phi-3 mini 3.8B LLM for Your PHONE: 1st TEST
1 week, 2 days ago |
www.youtube.com
BEST LLMs for Coding, Long Context, Overall Perform
1 week, 3 days ago |
www.youtube.com
Jobs in AI, ML, Big Data
AI Engineer Intern, Agents
@ Occam AI | US
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Lead Data Modeler
@ Sherwin-Williams | Cleveland, OH, United States