Jan. 13, 2024, 1 p.m. | code_your_own_AI

code_your_own_AI www.youtube.com

A LLama-2 7B LLM with 400K context length has been build with a new method. Based on activation compression and activation beacons for context intervals. With a sliding window methodology a context length of 400K for a LLama-2 7B LLM has been tested. Here are the results.

A new competitive method to extend context lengths of LLM not just by fine-tuning? Is it useful to extend your LLM from 4K to maybe just 32K context length? What compute infrastructure do …

beacons beyond build compression context llama llm methodology

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne