Jan. 12, 2024, 1 p.m. | code_your_own_AI

code_your_own_AI www.youtube.com

Context length is a limiting factor for Large Language Models, especially for RAG systems. But 2/3 of current LLMs fail at 2K token length, in Jan 2024! Is RAG-induced multi-placement a solution?

My test prompt is with 5272 characters too long to upload to Youtube.
I use a prompt of 741 words, corresponding to 1254 tokens (of my adapted tokenizer).
Every tested models applies its specific tokenization.

It is amazing, that open source LLM outperform even huge commercial LLMs.

By …

characters context current jan language language models large language large language models llms placement prompt rag solution systems test token youtube

Software Engineer for AI Training Data (School Specific)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Python)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Tier 2)

@ G2i Inc | Remote

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US