all AI news
JAMBA MoE: Open Source MAMBA w/ Transformer: CODE
March 29, 2024, 4 p.m. | code_your_own_AI
code_your_own_AI www.youtube.com
Databricks DBRX compared to AI21Labs JAMBA (architecture, size, free trainable parameters)
Video includes:
1. JAMBA inference code, plus 8-bit quantization code
2. JAMBA Fine-tuning Python Code w/ SFT trainer from HuggingFace.
3. Performance data of JAMBA vs MIXTRAL in three categories.
#airesearch
#ai
#newtech
architecture attention code databricks dbrx fine-tuning free huggingface inference jamba llm mamba moe open source parameters python quantization self-attention sft space state state space model trainer transformer video
More from www.youtube.com / code_your_own_AI
Multi-Token Prediction (forget next token LLM?)
1 day, 10 hours ago |
www.youtube.com
NEW LLM Test: Reasoning & gpt2-chatbot
2 days, 16 hours ago |
www.youtube.com
LLMs: Rewriting Our Tomorrow (plus code) #ai
3 days, 22 hours ago |
www.youtube.com
Autonomous AI Agents: 14 % MAX Performance
5 days, 10 hours ago |
www.youtube.com
No more Fine-Tuning: Unsupervised ICL+
1 week, 1 day ago |
www.youtube.com
NEW Phi-3 mini 3.8B LLM for Your PHONE: 1st TEST
1 week, 2 days ago |
www.youtube.com
BEST LLMs for Coding, Long Context, Overall Perform
1 week, 3 days ago |
www.youtube.com
Jobs in AI, ML, Big Data
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Data Science Analyst
@ Mayo Clinic | AZ, United States
Sr. Data Scientist (Network Engineering)
@ SpaceX | Redmond, WA