all AI news
Self-Extend LLM: Upgrade your context length
Jan. 16, 2024, 1 p.m. | code_your_own_AI
code_your_own_AI www.youtube.com
Neural networks (NNs) and in particular LLMs are susceptible to unpredictable behaviors when dealing with O.O.D inputs. We analyse a new solution, to increase the context length of LLM during inference!
Introducing grouped self-attention, that extends the classical self-attention of transformers outside of their pre-trained context length!
All rights w/ authors: …
context context window distribution encoding inference inputs llm llms networks neural networks nns positional encoding pre-training solution text training upgrade
More from www.youtube.com / code_your_own_AI
NEW LLM Test: Reasoning & gpt2-chatbot
1 day, 20 hours ago |
www.youtube.com
LLMs: Rewriting Our Tomorrow (plus code) #ai
3 days, 3 hours ago |
www.youtube.com
Autonomous AI Agents: 14 % MAX Performance
4 days, 15 hours ago |
www.youtube.com
480B LLM as 128x4B MoE? WHY?
6 days, 15 hours ago |
www.youtube.com
No more Fine-Tuning: Unsupervised ICL+
1 week, 1 day ago |
www.youtube.com
NEW Phi-3 mini 3.8B LLM for Your PHONE: 1st TEST
1 week, 1 day ago |
www.youtube.com
BEST LLMs for Coding, Long Context, Overall Perform
1 week, 2 days ago |
www.youtube.com
Next-Gen AI: RecurrentGemma (Long Context Length)
1 week, 4 days ago |
www.youtube.com
Jobs in AI, ML, Big Data
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Data Analyst (Digital Business Analyst)
@ Activate Interactive Pte Ltd | Singapore, Central Singapore, Singapore