all AI news
JAMBA MoE: Open Source MAMBA w/ Transformer: CODE
March 29, 2024, 4 p.m. | code_your_own_AI
code_your_own_AI www.youtube.com
Databricks DBRX compared to AI21Labs JAMBA (architecture, size, free trainable parameters)
Video includes:
1. JAMBA inference code, plus 8-bit quantization code
2. JAMBA Fine-tuning Python Code w/ SFT trainer from HuggingFace.
3. Performance data of JAMBA vs MIXTRAL in three categories.
#airesearch
#ai
#newtech
architecture attention code databricks dbrx fine-tuning free huggingface inference jamba llm mamba moe open source parameters python quantization self-attention sft space state state space model trainer transformer video
More from www.youtube.com / code_your_own_AI
New xLSTM explained: Better than Transformer LLMs?
2 days, 7 hours ago |
www.youtube.com
Stealth LLM: im-a-good-gpt2-chatbot
4 days, 7 hours ago |
www.youtube.com
Understand DSPy: Programming AI Pipelines
6 days, 7 hours ago |
www.youtube.com
Latest Insights in AI Performance Models
1 week, 1 day ago |
www.youtube.com
New Discovery: Retrieval Heads for Long Context
1 week, 3 days ago |
www.youtube.com
Multi-Token Prediction (forget next token LLM?)
1 week, 4 days ago |
www.youtube.com
NEW LLM Test: Reasoning & gpt2-chatbot
1 week, 5 days ago |
www.youtube.com
LLMs: Rewriting Our Tomorrow (plus code) #ai
1 week, 6 days ago |
www.youtube.com
Jobs in AI, ML, Big Data
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US
Research Engineer
@ Allora Labs | Remote
Ecosystem Manager
@ Allora Labs | Remote
Founding AI Engineer, Agents
@ Occam AI | New York